Google presented its latest artificial intelligence model; This model is called PaliGemma 2 and has the ability to analyze images and provide textual output based on their content.
Google has claimed that the new artificial intelligence model is capable of recognizing people’s emotions; Experts are worried about the consequences of such a capability.
Not long ago, Google started introducing a new family of artificial intelligence models called PaliGemma 2 that can analyze photos and identify their content. Other features of these models include automatically writing descriptions after viewing photos and answering users’ questions about the images.
Google has also claimed that this tool can not only recognize objects or actions in photos; It also has the ability to identify people’s feelings. However, it should be noted that according to Google, the feature of recognizing people’s feelings is not active by default in PaliGemma 2, and detailed settings must be made to benefit from it.
On the other hand, the novelty of this feature has brought doubts about its exact performance; However, one of the important applications of such technology is to detect the emotions of customers and it can create a revolution in marketing. This issue has caused the concern of experts in this field. Google’s extensive activity in the field of internet advertising and access to a huge amount of user data also add to the concerns.
According to the studies conducted in recent years, most of the emotion recognition models available in the market do not have the necessary accuracy, and even some of these models have a background in attributing emotions to people; For example, negative emotions in one of them were detected in most cases for black people. However, Google has announced that it has conducted extensive investigations to ensure the absence of similar problems in its model.
Google’s new artificial intelligence model has not yet gone through the trial and error process with users; But it has been mentioned that in the FairFace benchmark, it was able to get a good score. The aforementioned benchmark is a set of images of people’s faces that are used to evaluate the performance of such services; But critics believe that the figures in it are not very diverse and its rating cannot be trusted.
It should be remembered that the comprehensive law of artificial intelligence indicates that the use of emotion recognition tools is prohibited for institutions such as schools and companies; But regulatory bodies like the police can take advantage of them.
RCO NEWS