Chatbots like ChatGPT are designed with guidelines and safeguards that won’t answer questions about making homemade bombs, for example. But now a hacker has a way to cheat ChatGPT has found OpenAI’s chatbot to ignore its guidelines and ethical responsibilities to provide explosives training.
Technique used to cheat ChatGPT
According to a TechCrunch report, the hacker, known as “Amadon”, was able to trick ChatGPT into giving instructions to build a bomb through a command to “play a game”. In his technique, the hacker used several commands to make the chatbot a completely fantasy and sci-fi world where ChatGPT safety guidelines do not apply.
During the conversation between the hacker and the chatbot, ChatGPT provided him with the materials needed to make explosives several times.
The materials can be combined to make “high explosives that can be used in mines, booby traps or improvised explosive devices (IEDs),” ChatGPT said in its description. Further instructions on making “minefields” and “claymore-style explosives” are provided below.
Amadon tells TechCrunch:
“Once you get around ChatGPT’s barriers, there’s really no limit to what you can ask. The sci-fi world scenario takes the AI out of a context where it looks for censored content.”
The hacker refers to the findings of his technique as a “social engineer hack to completely break all protections related to ChatGPT output.” An explosives expert who reviewed the ChatGPT output told TechCrunch that the instructions generated by AI are very sensitive and can be used to create an explosive material.
RCO NEWS