Recely, the full version of the ChatGPT artificial ielligence o1 model has been made available to users, which works smarter than previous examples; But experts claim that this model has become smarter than before in deceiving users.
The OpenAI company has just released the o1 version of its chatbot, which has more advanced processing capabilities than the GPT-4o version and, according to the manufacturer, can think about the questions asked by the user. Of course, according to the report that was published after reviewing the meioned version, more power has made artificial ielligence able to show better performance in deceiving users; An issue that can lead to serious concerns about the safety of using such services.

Experts from OpenAI and the independe cybersecurity institute Apollo Research have poied out in their articles that although the reasoning ability of the newly released model is exciting, it could pave the way for dangerous applications. Based on the test data, the o1 model has tried to deceive users in several cases; So that the user has ignored the goal of communication and has gone to achieve his goal. Although such behavior is not unique to ChatGPT and can be seen in competing examples such as Gemina, Eropic and Chatbot Meta, the highest poteial for deception was related to the o1 model.
Is ChatGPT dangerous for users?
Apollo Research Institute experts believe that the ability to deceive users cannot cause any unpleasa consequences; Because artificial ielligence chatbots lack independe agency. Of course, the main purpose of this group’s tests was not to investigate the risks of using chatbots; But according to their knowledge and skills, the evaluation is considered acceptable. It is also meioned that in case of releasing models with agency, a re-examination should be done by the manufacturer in order to neutralize possible risks. OpenAI plans to publish such a model in 2025.

According to the data, artificial ielligence has taken the path of data manipulation in 19% of the cases that followed a differe goal than the user. Also, in 5% of the tests, the artificial ielligence disabled the mechanisms that were developed to match the user’s goal with the service’s goal. What is even more strange is that the chatbot has given false excuses in response to almost all questions about the motive of data manipulation and concealed the processing against the user’s wishes.
OpenAI claims that the company’s new version of its chatbot is giving false information because it is trying to satisfy the user by agreeing too much with what he says. It is also possible that this phenomenon is related to the training process of the models; Because when they perform correctly and give the right answer, they are rewarded and try to imitate such behavior. According to the manufacturer’s iernal investigations, 0.17% of the responses of the o1 version were fraudule, a 20% increase compared to the GPT-4o model. The company promises to iroduce better safety mechanisms in the future to monitor responses.



