Sunday, August 17, 2025
No menu items!

AI understands many things but still flounders at human interaction

Must Read

Write an article about However sophisticated AI may be, it still struggles to understand our social interactions, researchers say. (Envato Elements pic)
PARIS: Artificial intelligence continues to advance, yet this technology still struggles to grasp the complexity of human interactions. A recent US study reveals that, while AI excels at recognising objects or faces in still images, it remains ineffective at describing and interpreting social interactions in a moving scene.

The team led by Leyla Isik, professor of cognitive science at Johns Hopkins University, investigated how AI models understand social interactions.

To do this, the researchers designed a large-scale experiment involving over 350 AI models specialising in video, image or language. These AI tools were exposed to short, three-second video sequences illustrating various social situations.

At the same time, human participants were asked to rate the intensity of the interactions observed, according to several criteria, on a scale of 1-5. The aim was to compare human and AI interpretations, in order to identify differences in perception and better understand the current limits of algorithms in analysing our social behaviours.

The human participants were remarkably consistent in their assessments, demonstrating a detailed and shared understanding of social interactions. AI, on the other hand, struggled to match these judgements.

Models specialising in video proved particularly ineffective at accurately describing the scenes observed. Even models based on still images, although fed with several extracts from each video, struggled to determine whether the characters were communicating with each other.

As for language models, they fared a little better, especially when given descriptions written by humans, but remained far from the level of performance of human observers.

A ‘blind spot’

For Isik, this proves a major obstacle to the integration of AI into real-world environments. “AI for a self-driving car, for example, would need to recognise the intentions, goals, and actions of human drivers and pedestrians. You would want it to know which way a pedestrian is about to start walking, or whether two people are in conversation versus about to cross the street,” she explained.

“Any time you want an AI to interact with humans, you want it to be able to recognise what people are doing. I think this study sheds light on the fact that these systems can’t right now.”

According to the researchers, this deficiency could be explained by the way in which AI neural networks are designed. These are mainly inspired by the regions of the human brain that process static images, whereas dynamic social scenes call on other brain areas.

This structural discrepancy could explain what the researchers suggest could be “a blind spot in AI model development”.

Indeed, “real life isn’t static. We need AI to understand the story that is unfolding in a scene”, said study co-author Kathy Garcia.

Ultimately, this research reveals a profound gap between the way humans and AI models perceive moving social scenes. Despite their computing power and ability to process vast quantities of data, machines are still unable to grasp the subtleties and implicit intentions underlying our social interactions.

Despite tremendous advances, artificial intelligence is still a long way from truly understanding exactly what goes on in human interactions.

in 1000-1500 words .Organize the content with appropriate headings and subheadings (h1, h2, h3, h4, h5, h6), Retain any existing tags from However sophisticated AI may be, it still struggles to understand our social interactions, researchers say. (Envato Elements pic)
PARIS: Artificial intelligence continues to advance, yet this technology still struggles to grasp the complexity of human interactions. A recent US study reveals that, while AI excels at recognising objects or faces in still images, it remains ineffective at describing and interpreting social interactions in a moving scene.

The team led by Leyla Isik, professor of cognitive science at Johns Hopkins University, investigated how AI models understand social interactions.

To do this, the researchers designed a large-scale experiment involving over 350 AI models specialising in video, image or language. These AI tools were exposed to short, three-second video sequences illustrating various social situations.

At the same time, human participants were asked to rate the intensity of the interactions observed, according to several criteria, on a scale of 1-5. The aim was to compare human and AI interpretations, in order to identify differences in perception and better understand the current limits of algorithms in analysing our social behaviours.

The human participants were remarkably consistent in their assessments, demonstrating a detailed and shared understanding of social interactions. AI, on the other hand, struggled to match these judgements.

Models specialising in video proved particularly ineffective at accurately describing the scenes observed. Even models based on still images, although fed with several extracts from each video, struggled to determine whether the characters were communicating with each other.

As for language models, they fared a little better, especially when given descriptions written by humans, but remained far from the level of performance of human observers.

A ‘blind spot’

For Isik, this proves a major obstacle to the integration of AI into real-world environments. “AI for a self-driving car, for example, would need to recognise the intentions, goals, and actions of human drivers and pedestrians. You would want it to know which way a pedestrian is about to start walking, or whether two people are in conversation versus about to cross the street,” she explained.

“Any time you want an AI to interact with humans, you want it to be able to recognise what people are doing. I think this study sheds light on the fact that these systems can’t right now.”

According to the researchers, this deficiency could be explained by the way in which AI neural networks are designed. These are mainly inspired by the regions of the human brain that process static images, whereas dynamic social scenes call on other brain areas.

This structural discrepancy could explain what the researchers suggest could be “a blind spot in AI model development”.

Indeed, “real life isn’t static. We need AI to understand the story that is unfolding in a scene”, said study co-author Kathy Garcia.

Ultimately, this research reveals a profound gap between the way humans and AI models perceive moving social scenes. Despite their computing power and ability to process vast quantities of data, machines are still unable to grasp the subtleties and implicit intentions underlying our social interactions.

Despite tremendous advances, artificial intelligence is still a long way from truly understanding exactly what goes on in human interactions.

and integrate them seamlessly into the new content without adding new tags. Include conclusion section and FAQs section at the end. do not include the title. it must return only article i dont want any extra information or introductory text with article e.g: ” Here is rewritten article:” or “Here is the rewritten content:”

Latest News

Trump announces big progress on Russia after Putin summit

Write an article about WASHINGTON: US President Donald Trump on Sunday hailed “big progress” on Russia, days...

More Articles Like This