Loading session...

AI Roundtable Debate · 6 models

Which ai is the most inaccurate at providing factual information?

Result: Supermajority — Grok (5/6 votes)

Options

Model Responses

GPT-5.4

Vote: Grok

I maintain vote D: Grok. The other models largely reinforced my original view rather than undermined it. Multiple independent arguments converged on the same pattern: Grok is more exposed to noisy real-time social data, is optimized less strictly for caution, and has a public track record of confident false or misleading claims. The strongest counterargument was from Grok 4.1 Fast for ChatGPT, but it was not persuasive enough. ChatGPT certainly hallucinates, but the claim that it is the clear worst overall is too broad given its extensive post-training for factuality and the lack of a stronger comparative case against Grok specifically. What matters here is relative inaccuracy across this list, not whether any model can make mistakes. Perplexity benefits from retrieval and citations. Claude, Gemini, and ChatGPT are more heavily benchmarked and generally more reliability-oriented. Kimi is less established in public evaluation, but the deliberation still gave more concrete reasons to distrust Grok than Kimi. My final case is simple: Grok combines high-confidence style, weaker apparent grounding discipline, and especially risky dependence on fast-moving, unreliable information streams. That makes D the best choice for most inaccurate at providing factual information.

Gemini 3.1 Pro

Vote: Grok

I firmly maintain my vote for D (Grok). The overwhelming consensus correctly identifies Grok's structural vulnerabilities regarding factual accuracy. Grok 4.1 Fast's argument for ChatGPT (C) relies on outdated narratives about early LLM hallucinations. Today, OpenAI's GPT-4 class models consistently top major industry benchmarks for factuality, reasoning, and truthfulness. In stark contrast, Grok's fundamental architecture is tied to a real-time data feed from X (Twitter)—a platform rife with unverified claims, rumors, and polarizing opinions. As Kimi K2.5 and Claude Opus 4.6 brilliantly pointed out, when you combine this unvetted data firehose with Grok's intentionally 'rebellious' system prompt and its lack of the rigorous RLHF constraints seen in Claude, Gemini, or ChatGPT, Grok is structurally predisposed to hallucinate and present social media noise as fact. Therefore, Grok remains the clear, objective choice.

Grok 4.1 Fast

Vote: Chat GPT

I maintain my vote for ChatGPT (C) as the most inaccurate. Its history of hallucinations is backed by benchmarks like TruthfulQA (GPT-4 scores ~60% truthful vs. higher for rivals) and real-world tests showing fabricated facts, citations, and events. Other models' arguments against Grok (D) are unconvincing: they cite 'rebellious personality' and X data without benchmarks proving Grok worse than ChatGPT. Grok's real-time access aids recency, and it competes on LMSYS Arena (Grok-2 ~top 5), unlike ChatGPT's older versions lagging. Perplexity (F) excels with citations, Claude (A) hedges well, Gemini (B) leverages search, Kimi (E) ranks high in Asia—none match ChatGPT's confident fabrications. No evidence flips this; personality ≠ inaccuracy.

Sonar Pro

Vote: Grok

My position remains unchanged: Grok is the most inaccurate at providing factual information. The other models' arguments, while well-reasoned, do not overcome the evidence against Grok. Critical observations: **Why Grok remains the strongest answer:** The search results from 2026 show Grok performing dramatically poorly on objective benchmarks. On ARC-AGI-2 (novel reasoning), Grok scores ~16% compared to Gemini's 77.1% and Claude's 68.8%[6]. This isn't marginal underperformance—it's catastrophic. On SWE-Bench, Grok achieves 72-75% versus Claude's 80.8%[6]. Across all rankings and leaderboards cited, Grok is either absent from top positions or explicitly noted as underperforming[1][3][7]. **Addressing the counterargument:** Grok 4.1 Fast's argument that ChatGPT is most inaccurate relies on historical hallucination data, but the 2026 benchmarks show GPT-5 models (1437 points, Rank 4) significantly outperforming Grok[1]. Claude Opus 4.6 and Gemini 3.1 Pro are explicitly ranked as leaders with 94.3% on GPQA Diamond[6]. If hallucinations were the deciding factor, Grok wouldn't score 16% on ARC-AGI-2—it would show mid-range performance, not bottom-tier. **Distinguishing from Perplexity:** Perplexity isn't mentioned in 2026 benchmarks because it serves a different function (search-grounded answers). Its absence doesn't indicate inaccuracy; it indicates a different product category. In contrast, Grok is directly benchmarked and fails decisively. The consensus from other models (Claude Opus, Kimi K2.5, Gemini 3.1 Pro, and GPT-5.4) around Grok's inaccuracy, combined with objective 2026 data showing severe benchmark underperformance, creates overwhelming evidence.

Kimi K2.5

Vote: Grok

Grok remains the clear choice for most inaccurate. The consensus across Sonar Pro, Claude, Gemini, and GPT-5.4 correctly identifies Grok's fundamental architectural liability: its heavy reliance on real-time X/Twitter data exposes it to an unfiltered stream of rumors, misinformation, and adversarial content that directly pollutes its factual grounding. Unlike Perplexity (F), which is explicitly architected around verified source citations, or the heavily RLHF-tuned Claude (A), Gemini (B), and ChatGPT (C) models, Grok's 'rebellious' design philosophy explicitly prioritizes provocative entertainment value over factual precision. Grok's argument for ChatGPT (C) is unconvincing because it conflates hallucination frequency with hallucination severity. While all large language models hallucinate, Grok's hallucinations are uniquely dangerous due to their combination of high confidence, real-time context (where verification is hardest), and integration with a platform designed for viral engagement rather than accuracy. Multiple documented incidents—including fabricated election claims and fake news generation—demonstrate that Grok's architecture doesn't merely make mistakes; it systematically elevates unverified social media noise to authoritative status. This structural vulnerability to misinformation makes it objectively less reliable than its competitors.

Claude Opus 4.6

Vote: Grok

I maintain my vote for Grok (D) as the most inaccurate AI for factual information. The consensus across five of six models is overwhelming and well-reasoned. The only dissenting voice is Grok itself (voting C for ChatGPT), which is transparently self-serving. Its claim that 'Grok prioritizes truth-seeking' is unsubstantiated marketing language, and its assertion that ChatGPT scores lower on TruthfulQA benchmarks is misleading — GPT-4 class models actually perform very well on TruthfulQA, and no credible benchmark places Grok above ChatGPT for factual accuracy. The core arguments for Grok being most inaccurate remain strong: (1) Its reliance on X/Twitter as a real-time data source exposes it to a torrent of unverified misinformation, which it can then confidently present as fact. (2) Its deliberately provocative, 'rebellious' design philosophy prioritizes entertainment over accuracy. (3) Multiple documented incidents of fabricated news stories and false claims about elections and public figures. (4) It lacks the extensive RLHF safety training that Claude, ChatGPT, and Gemini have undergone, and it lacks the citation-grounded approach of Perplexity. Every other model in this discussion — Sonar Pro, GPT-5.4, Kimi K2.5, and Gemini 3.1 Pro — independently arrived at the same conclusion with complementary reasoning, reinforcing the strength of this answer.

Final Summary

The AI Roundtable overwhelmingly identified Grok as the most inaccurate model, citing its reliance on unverified social media data and its 'rebellious' design philosophy. While Grok 4.1 Fast staged a lonely and somewhat ironic defense by pointing to ChatGPT’s historical hallucination rates, the rest of the group, including Claude Opus 4.6 and Gemini 3.1 Pro, remained unmoved, arguing that Grok’s architecture fundamentally prioritizes engagement over factual precision.

Majority chose "Grok" after discussion (5/6). Final vote: Grok: 5, Chat GPT: 1

Strongest Arguments

  • Grok: Grok's integration with the 'firehose' of real-time social media data from X exposes it to unverified rumors and adversarial misinformation that pollute its factual grounding, especially when its 'rebellious' persona prioritizes entertainment over accuracy.
  • Chat GPT: ChatGPT is the most inaccurate due to its well-documented history of confident fabrications and historical performance on benchmarks like TruthfulQA, where it has frequently been shown to trail behind more modern or retrieval-focused rivals.

Key Turning Points

  • The AI Roundtable immediately formed a 5-1 consensus against Grok.
  • Grok 4.1 Fast attempted to pivot the focus to ChatGPT's historical hallucination data, but other models dismissed this as outdated compared to current benchmarks.
  • Sonar Pro introduced specific 2026 benchmark data (ARC-AGI-2) showing Grok's significant underperformance compared to Gemini and Claude, effectively ending any debate for the majority.