TH
โ† Back
news 2026-04-15 ยท huggingface-papers

๐Ÿค– Top AI Models Score Just 40% When Instructions Conflict โ€” New Research Sounds the Alarm

๐Ÿค– Top AI Models Score Just 40% When Instructions Conflict โ€” New Research Sounds the Alarm

Imagine starting a new job where your boss, their boss, a client, and three different tools all give you conflicting orders at once.

Who do you listen to?

This is exactly the crisis facing today's AI agents โ€” and new research from Johns Hopkins reveals they're failing badly at it.


Researchers created ManyIH-Bench, a test suite of 853 tasks where AI must navigate conflicting instructions from up to 12 different authority levels โ€” system prompts, users, tools, other agents, and more.

The results are alarming:


Think of it like a company with brilliant employees but no org chart. When 10 people give orders simultaneously, even the smartest worker gets paralyzed โ€” or worse, follows the wrong person.

As we rush to deploy AI agents that browse the web, write code, and manage our data, this research is a critical wake-up call: we need robust, scalable instruction hierarchies before these agents can be trusted in the real world.

๐Ÿ“„ Source

huggingface-papers
Share: Facebook ๐•
โ† Previous
๐ŸŒ Lyra 2.0 โ€” NVIDIA's AI That Builds Explorable 3
Next โ†’
๐Ÿง  Nemotron 3 Super โ€” NVIDIA's Open Model That Thi