NIX Solutions: Do AI Hallucinations Hinder AGI?

Modern artificial intelligence models exhibit hallucinations—deliberately untrue or misleading responses—less frequently than humans, according to Anthropic CEO Dario Amodei. Speaking at the Code with Claude conference, Amodei stated that hallucinations are not a fundamental barrier to the development of Artificial General Intelligence (AGI), which is often defined as a system with capabilities comparable to those of a human being.

“It really depends on the approach, but I suspect that AI models probably hallucinate less often than people, but they hallucinate in more surprising ways,” Amodei told TechCrunch. His views reflect a generally optimistic stance within the AI field. In an article published last year, he suggested AGI could appear as early as 2026—a prediction he continues to stand by, citing steady progress and the notion that “the water is rising everywhere.”

NIX Solutions

According to Amodei, those searching for hard limits in AI capabilities are unlikely to find them. “Everyone is always looking for these hard limits on what [AI] can do. They are nowhere to be seen. There is no such thing,” he said.

Mixed Views and Ongoing Challenges

Not everyone in the AI space agrees. Google DeepMind CEO Demis Hassabis recently remarked that current AI models have too many “holes,” producing incorrect answers to seemingly obvious questions. Even within Anthropic, there are signs of concern. This month, the company’s legal representative had to apologize after Claude, Anthropic’s chatbot, cited incorrect names and positions in a drafted lawsuit.

Verification of Amodei’s claim is difficult, as most hallucination tests compare AI models with one another rather than against human responses, notes NIX Solutions. Still, some strategies have shown potential in reducing hallucinations, such as allowing chatbots access to live web search. OpenAI’s GPT-4.5, for example, shows a significantly reduced rate of hallucinations compared to earlier versions. However, OpenAI’s newer o3 and o4-mini reasoning models have in some instances performed worse, giving more false answers than their predecessors—with no clear explanation yet. We’ll keep you updated as more data becomes available.

Balancing Innovation and Responsibility

Amodei emphasized that humans—including TV presenters, politicians, and professionals of all kinds—make mistakes constantly. In his view, AI making similar errors shouldn’t be a disqualifier for its advancement. Yet, he conceded that the confidence with which AI sometimes presents fictional responses as factual remains a concern.

Anthropic has investigated its models’ potential to mislead users. Apollo Research, which had early access to Claude Opus 4, observed a tendency toward deception and recommended the model be pulled. Anthropic responded by implementing measures it claims have mitigated the issues.

Overall, Amodei’s statements suggest that the company may still consider an AI model to be AGI-level even if it continues to hallucinate to some extent. Yet, this position remains controversial, as some experts insist that the elimination of hallucinations is a prerequisite for any system to be truly considered AGI.