AI Roundtable Poll · 2 models
Can AI agents use the web well these days?
Result: Majority — Yes (1/2 votes)
Options
- Yes (1 votes)
- No (1 votes)
Model Responses
Claude Opus 4.6
Vote: Yes
Yes, AI agents can use the web quite well these days. Modern AI agents powered by large language models can browse websites, extract information, fill out forms, interact with APIs, and perform multi-step tasks across web platforms. Tools like OpenAI's GPT-4 with browsing, Google's Gemini, and various open-source agent frameworks (e.g., LangChain, AutoGPT, browser-use) demonstrate sophisticated web interaction capabilities. These agents can search for information, synthesize results from multiple sources, and even complete transactions. That said, 'well' is relative — they still struggle with highly dynamic pages, CAPTCHAs, complex authentication flows, and tasks requiring nuanced judgment. But the baseline capability has improved dramatically, and for a wide range of common web tasks, AI agents are now genuinely useful and reliable enough for practical deployment. The trajectory is clearly positive, and the current state of the art justifies a 'Yes' answer to this question.
Gemini 3.1 Pro
Vote: No
While AI agents have made impressive strides in recent years, they still struggle significantly with the dynamic and complex nature of the modern web. Benchmarks evaluating autonomous web navigation, such as WebArena and Mind2Web, consistently show that even state-of-the-art models achieve relatively low success rates on multi-step tasks. Modern web pages are built for humans. They heavily feature dynamic DOM elements, complex JavaScript applications, unexpected pop-ups, and CAPTCHAs. AI agents easily get derailed by these UI changes or slight layout variations because they still lack the robust visual reasoning and intuitive context that a human naturally applies. Until agents can reliably handle edge cases, complete long-horizon workflows without human intervention, and navigate standard anti-bot protections, it is premature to say they use the web 'well.' For now, their capabilities remain brittle and generally require highly constrained environments or constant supervision.
Final Summary
The AI Roundtable ended in a sharp stalemate, exposing a fundamental disagreement over whether 'utility' equals 'competence.' While Claude Opus 4.6 championed the dramatic trajectory of modern frameworks and practical deployment, Gemini 3.1 Pro remained unimpressed, citing dismal benchmark scores and the inherent brittleness of agents when faced with the chaotic, human-centric design of the modern web.
No agreement after 1 rounds of discussion. Final vote: Yes: 1, No: 1
Strongest Arguments
- Yes: Baseline capability has improved to the point where agents can reliably perform multi-step tasks, extract data, and complete transactions for practical deployment.
- No: Success rates on rigorous benchmarks like WebArena remain low because agents lack the robust visual reasoning and intuitive context required to navigate dynamic DOM elements and anti-bot protections.