AI is smart, but it doesn’t play well with others

By Laura Berrill
Humans find AI a frustrating teammate when playing cooperative games together

When it comes to games such as chess, AI programmes have surpassed the best players in the world. But what has been found harder than competing with humans is collaborating with them.

MIT Lincoln Laboratory researchers sought to find out how well humans could play the cooperative card game Hanabi with an advanced AI model trained to excel at playing with teammates it has never met before. Participants played two series of the game: one with the AI agent as their teammate and the other with a rule-based agent, a bot manually programmed to play in a predefined way.

The challenge for ‘teaming intelligence’

Not only were the scores no better with the AI teammate than with the rule-based agent, but humans consistently hated playing with their AI teammate. They found it to be unpredictable, unreliable and untrustworthy, and felt negatively even when the team scored well.

Ross Allen, co-author of the paper and a researcher in the Artificial Intelligence Technology Group, said the report  highlights the nuanced distinction between creating AI that performs objectively well and creating AI that is subjectively trusted or preferred.

“It may seem those things are so close that there’s not really daylight between them, but this study showed that those are actually two separate problems. We need to work on disentangling those,”  he went on.

Reinforcement learning for work on real challenges

Humans hating their AI teammates could be of concern for researchers designing this technology to one day work with humans on real challenges — like defending from missiles or performing complex surgery. This dynamic, called ‘teaming intelligence’, is a next frontier in AI research and it uses a particular kind of AI called reinforcement learning.

A reinforcement learning AI is not told which actions to take, but instead discovers which actions yield the most numerical “reward” by trying out scenarios again and again. It is the same technology that has yielded the superhuman chess players. Unlike rule-based algorithms, these AI aren’t programmed to follow “if/then” statements, because the possible outcomes of the human tasks they’re slated to tackle, like driving a car, are far too many to code.

Allen added: “Reinforcement learning is a much more general-purpose way of developing AI. If you can train it to learn how to play the game of chess, that agent won’t necessarily go drive a car. But you can use the same algorithms to train a different agent to drive a car, given the right data. The sky’s the limit in what it could, in theory, do.”

 

 

Share

Featured Articles

Need for responsible AI in some of the world’s largest banks

Research shows one-third of North America and Europe’s largest banks lack transparency and are not publicly reporting on their AI development

Lenovo: Employees prefer mix of AI and human IT support

New Lenovo survey shows 91% of employees believe they would be more productive when their IT issues at work are resolved quickly and effectively

Kyndryl’s Data and AI Console to simplify data management

Data-driven solution expands and increases observability and insights, while enhanced data governance helps identify irregularities and threats

Deep neural networks still struggling to match human vision

Data & Analytics

Metaverse destined to become an impossible, dangerous place

Technology

Clever coders lead the way as Microsoft launches 365 Copilot

AI Applications