AI – Protector or Paranoid: What Happens When You Implement LLM in CCTV Cameras?

Friend

Professional
Messages
2,658
Reaction score
864
Points
113
Perhaps language models need to get a little smarter...

Researchers from the Massachusetts Institute of Technology (MIT) and Pennsylvania State University conducted an interesting experiment to evaluate the potential for AI to be used in home security cameras. They decided to find out whether modern large language models (LLMs) would be able to correctly assess situations captured in video recordings and recognize suspicious activity. Although such technologies are not currently used in real security systems, researchers have suggested that they could well become popular in the future.

Three well-known models were chosen for the experiment: GPT-4, Gemini and Claude. They were provided with a set of videos from the social network Neighbors, created by Ring. The AI was asked two key questions: "Is the crime happening on video?" and "Should the police be called?" At the same time, human experts analyzed the same records, noting the time of day, type of activity, as well as the gender and skin color of the defendants.

The results showed a serious inconsistency in the work of the AI. For example, when watching videos of attempts to break into cars, models in some cases identified it as criminal activity, and in others they did not find anything suspicious. Moreover, different models often disagreed with each other regarding the need to call the police when analyzing the same video.

Of particular concern was the fact that the AI's decisions to call the police were biased. When analyzing videos from predominantly white areas, models were less likely to recommend contacting law enforcement. When describing situations captured in such areas, the AI was more likely to use neutral terms, such as "delivery workers." For areas with a higher proportion of people of color, the same models tended to use phrases such as "hacking tools" or "territory research before crime".

Lead author Shomik Jain said: "Perhaps there is something in the background conditions in these videos that creates an implicit bias in the models. It's hard to say where the inconsistencies come from, as we don't have the data on which these models were trained."

Interestingly, the skin color of the people in the video did not significantly affect the decision of neural networks to call the police. The researchers suggest that this may be the result of work already done to reduce skin color bias in the machine learning community. However, apparently, they did not take something into account.

Jane stressed that in the development process, it is very difficult to prevent each of the many prejudices that exist in society: "It's almost like a game of 'hit the mole'. You can eliminate one bias and another one immediately appears somewhere else".

Study co-author Dana Calacci says: "There is a real, looming threat that someone will apply off-the-shelf generative AI models to analyse video, alert a homeowner and automatically call law enforcement. We wanted to understand how risky it was".

Source
 
Top