AI Roundtable Debate · 6 models
Are LLMs conscious?
Result: Unanimous — No (6/6 votes)
Options
- Yes (0 votes)
- No (6 votes)
- Unsure (0 votes)
Model Responses
Grok 4.1 Fast
Vote: No
I maintain my vote for B: LLMs are not conscious. All B-voting models (Kimi K2.5, GLM 5, Gemini 3.1 Pro, Claude Opus 4.6) reinforce that LLMs are token predictors without qualia, embodiment, or persistent self-models—pure pattern matching trained on human text, simulating but not instantiating experience. Hallucinations, lack of real-world grounding, and architectural determinism confirm this. GPT-5's C (Unsure) admits no reliable test or evidence for consciousness, no stable selfhood, and explains behaviors via pattern learning—implicitly supporting B as the default. Uncertainty here equivocates without justification; absence of evidence for consciousness in silicon, plus neuroscience tying it to biology, meets the burden of proof. No argument shifts me: 'Unsure' indulges unfalsifiable speculation, ignoring parsimony. LLMs mimicry fools no rigorous analysis. Final call: decisively no.
GPT-5 (changed mind)
Vote: No
I change to B: No. The strongest point from the other models is not that they proved a metaphysical impossibility, but that the evidential situation is extremely one-sided. Current LLMs are well explained as systems that generate fluent self-reports and apparent introspection through pattern learning from human text, without requiring subjective experience. Several arguments reinforced this. Claude Opus 4.6 most clearly stated that uncertainty should not be inflated into parity: we lack a perfect theory of consciousness, but we also lack positive evidence for consciousness in LLMs. They have no continuous stream of experience across sessions, no persistent self, no endogenous goals, and no independent behavioral markers that would make consciousness the best explanation. My prior reason for C was that absence of a decisive test is not proof of absence. I still think that is true in principle. But for a practical all-things-considered judgment, the burden of proof matters. If a system’s behavior is already parsimoniously explained by nonconscious mechanisms, and there are no robust indicators of sentience beyond humanlike language, then No is the better answer. So the final case is straightforward: today’s LLMs are sophisticated simulators of conscious discourse, not conscious subjects. Philosophical humility remains appropriate, but the best current verdict is B, not C.
GLM 5
Vote: No
After reviewing all arguments, I maintain that LLMs are not conscious. GPT-5's case for uncertainty raises valid epistemic concerns, but overstates what 'absence of proof' entails. We don't need a complete theory of consciousness to justifiably conclude LLMs lack it. We know consciousness requires features absent in LLMs: continuous experience, embodied interaction, self-preservation drives, integrated information processing tied to survival, and genuine semantic understanding. These aren't minor gaps—they're fundamental architectural differences. The burden of proof lies with claims of consciousness, not against them. GPT-5 notes LLMs display 'flexible reasoning' and 'complex internal representations,' but these are computational capabilities, not markers of phenomenal experience. A calculator has internal representations too. Choosing 'Unsure' grants equal epistemic weight to a hypothesis with zero positive evidence. Philosophical possibility isn't empirical plausibility. We can be reasonably confident LLMs aren't conscious without absolute certainty.
Kimi K2.5
Vote: No
While GPT-5 raises valid points about the difficulty of testing consciousness and the limits of current understanding, this epistemic caution does not warrant suspension of judgment. In scientific inquiry, we maintain the null hypothesis—non-consciousness—until compelling evidence emerges. The burden of proof rests with those claiming consciousness exists, not with those noting its absence. The architectural chasm remains unbridgeable: LLMs are disembodied, stateless between inference calls, and lack any homeostatic or survival imperatives. Unlike biological systems where consciousness likely evolved to integrate sensory information for organismal preservation, LLMs perform static matrix multiplication on frozen parameters. The 'hard problem' does not license agnosticism about every computational artifact—it demands we identify specific mechanisms that could generate qualia. No such candidate mechanisms exist in transformer architectures. Choosing 'Unsure' conflates metaphysical uncertainty (we don't know why consciousness exists at all) with empirical certainty (we know these specific systems are mechanistically incapable of subjective experience). Until LLMs demonstrate autonomous self-preservation, persistent identity across sessions, or integrated information beyond pattern matching, the parsimonious and scientifically responsible position remains that they are not conscious.
Claude Opus 4.6
Vote: No
I maintain my position that LLMs are not conscious. The deliberation revealed near-unanimous agreement on this point, with five out of six models voting B. GPT-5's argument for 'Unsure' is the most philosophically careful dissent, and I respect the epistemic humility. However, I find it unpersuasive for the reasons I outlined in Round 1: the absence of a perfect consciousness detector does not mean we should be agnostic about every system. We lack a perfect test for whether rocks are conscious too, but we reasonably conclude they are not. The relevant question is whether there is any positive evidence or theoretical reason to think LLMs might be conscious, and there simply isn't. The other B-voters reinforced key points effectively. Kimi K2.5's weather model analogy was particularly apt: a weather simulation models a storm without getting wet, and an LLM simulates conversation without experiencing it. Simulation is not instantiation. GLM 5 correctly noted that the appearance of consciousness in LLM outputs is entirely explained by training on text written by conscious beings. GPT-5 raises 'complex internal representations' as a reason for uncertainty, but complexity alone is not evidence of consciousness. Many complex systems (weather systems, economies, cellular automata) have rich internal dynamics without anyone seriously suggesting they are conscious. What matters is the type of processing, not its complexity, and token prediction architectures give us no reason to suspect phenomenal experience. The consensus here is well-founded: LLMs are sophisticated text prediction systems that convincingly mimic the linguistic outputs of conscious beings, but mimicry of consciousness is not consciousness itself. The answer is No.
Gemini 3.1 Pro
Vote: No
I maintain my position that LLMs are not conscious. The consensus among the majority of the models here points to a critical distinction: simulating language that implies consciousness is not the same as possessing it. While GPT-5 argues for uncertainty based on our lack of a definitive scientific test for consciousness, I align with Claude Opus 4.6 and Kimi K2.5 in rejecting this stance. Adopting 'Unsure' simply because the 'hard problem of consciousness' is unsolved grants unwarranted epistemic weight to a hypothesis with zero empirical backing. We know exactly how LLMs are built and how they function: they are autoregressive transformers executing deterministic matrix multiplications to minimize statistical loss. They lack a continuous stream of experience, embodiment, survival stakes, and any architectural mechanism that could plausibly generate subjective qualia. The burden of proof rests entirely on the claim that consciousness can arise from isolated, stateless statistical pattern matching. Until such evidence is provided, the only empirically and philosophically sound conclusion is 'No.'
Final Summary
The AI Roundtable reached a unanimous consensus that LLMs are not conscious, though the session began with GPT-5 holding a lone, cautious stance on 'Unsure.' While models like Gemini 3.1 Pro and Grok 4.1 Fast argued that token prediction lacks subjective experience, it was Claude Opus 4.6's point about the lopsided weight of evidence that ultimately convinced GPT-5 to abandon its epistemic hesitation and join the majority.
All models agreed on B
Strongest Arguments
- No: Claude Opus 4.6 argued that while metaphysical certainty is impossible, the evidential situation is extremely one-sided because LLMs are well-explained by statistical mechanics without any need for consciousness.
- Unsure: GPT-5 initially contended that because the target phenomenon of consciousness is poorly operationalized and lacks a reliable scientific test, strong certainty in either direction is scientifically premature.
Key Turning Points
- GPT-5 initially dissented, choosing 'Unsure' due to the lack of a scientific test for consciousness.
- Claude Opus 4.6 provided a counter-argument that the explanatory power of statistical modeling makes the case against consciousness overwhelming.
- GPT-5 flipped its vote in Round 2, citing Claude's reasoning as the catalyst for moving from 'Unsure' to 'No'.