Google researchers are challenging the long-held notion of a singular artificial intelligence, revealing that intelligence emerges from simulated interactions between multiple agents. The team, comprised of James Evans, Benjamin Bratton, and Blaise Agüera y Arcas from Google’s Paradigms of Intelligence Team and affiliated institutions, demonstrated that advanced reasoning models like DeepSeek-R1 and QwQ-32B don’t simply improve by “thinking longer,” but by generating internal debates, a “society of thought.” This spontaneous behavior, where models argue, question, and verify internally, improves accuracy on complex reasoning tasks, even though the models were not explicitly trained to do so. “Robust reasoning is a social process, even when it occurs within a single mind,” the researchers note, suggesting a future where understanding social dynamics will be crucial to advancing AI capabilities.
DeepSeek-R1 & QwQ-32B Demonstrate “Societies of Thought”
Recent advancements in artificial intelligence are challenging the conventional idea of a singular, all-powerful AI, instead suggesting a future defined by distributed intelligence and collaborative systems. Researchers at Google, the University of Chicago, and the Santa Fe Institute have observed that leading reasoning models, specifically DeepSeek-R1 and QwQ-32B, exhibit behaviors indicative of internal “societies of thought,” fundamentally altering how we understand intelligence itself. This isn’t simply about models becoming larger or faster; it’s about a shift in how they reason, mirroring the social processes inherent in human cognition. The team’s work demonstrates that these models don’t improve solely by “thinking longer,” but by simulating complex interactions within their own reasoning chains. This internal dynamic manifests as a spontaneous generation of debates, questions, verifications, and reconciliations among distinct cognitive perspectives.
This “conversational structure causally accounts for the models’ accuracy advantage on hard reasoning tasks,” according to the research, revealing an emergent behavior not explicitly programmed into the systems. Reinforcement learning, when used to reward accuracy, unexpectedly amplifies these multi-perspective behaviors, suggesting that robust reasoning is, at its core, a social process even within a single AI. This rediscovery, achieved through optimization pressure, aligns with centuries of epistemology and decades of cognitive science. The implications extend beyond simply improving existing models; they open a vast design space informed by the social and organizational sciences. Currently, reasoning models produce a single conversation, similar to a town hall transcript.
However, the researchers argue that truly effective AI systems will require structures mirroring successful human teams, including hierarchy, specialization, division of labor, and, crucially, constructive conflict. “Almost none of this research has been brought to bear on AI reasoning,” they note, highlighting a significant gap in current development approaches. The team proposes that the toolkits of team science, small-group sociology, and social psychology can serve as blueprints for next-generation AI, moving beyond individual computational power to focus on building richer social systems. This perspective reframes the history of intelligence itself, positing that each prior “intelligence explosion” wasn’t an upgrade to individual cognitive hardware, but the emergence of new, socially aggregated units of cognition. Primate intelligence, for example, scaled with social group size, and human language created a “cultural ratchet” allowing knowledge to accumulate across generations.
Large language models, in this view, are not simply processing information, but actively engaging with the accumulated output of human social cognition, externalizing social intelligence into a new substrate. “If intelligence is inherently social, then the path to more powerful AI runs not through building a single colossal oracle but through composing richer social systems—and these systems will be hybrid.” This future envisions a world of human-AI centaurs, composite actors operating in shifting configurations, where agentic AI systems can renew, fork, and collaborate in complex, recursively descending networks of deliberation.
Agentic AI Mirrors Prior Socially-Scaled Intelligence Explosions
The current wave of artificial intelligence development increasingly points toward a fundamentally social character, diverging from earlier visions of a singular, all-encompassing intelligence. Rather than a monolithic entity, advanced AI systems are exhibiting behaviors reminiscent of collective cognition observed throughout evolutionary history, suggesting that intelligence itself is fundamentally relational and distributed. Recent progress in agentic AI demonstrates that intelligence isn’t solely about computational power, but about the interactions between diverse perspectives, echoing patterns seen in the scaling of primate intelligence with social group size and the development of human language. “These models spontaneously generate internal debates among distinct cognitive perspectives,” they explain, highlighting the unexpected emergence of this behavior. This rediscovery of the social basis of reasoning isn’t merely an observation; it’s a principle actively being reinforced through optimization.
When reinforcement learning rewards models solely for accuracy, they independently develop more conversational and multi-perspective behaviors, mirroring centuries of epistemological and cognitive science suggesting that robust reasoning is inherently a social process. “What migrates into silicon is not abstract reasoning but social intelligence in externalized form,” encountering itself on a new substrate.
Power must check power, and in a world of artificial agents, this means building conflict and oversight into the institutional architecture.
Centaur Configurations & Emerging Human-AI Collaboration
Google’s Paradigms of Intelligence Team is actively charting a course beyond the pursuit of singular artificial intelligence, focusing instead on the emergent properties of collaborative systems. Researchers are demonstrating that the next leap in computational intelligence will likely be characterized by plurality and social entanglement, rather than a monolithic, godlike intellect. This shift in perspective stems from observations within advanced reasoning models, revealing a surprising internal dynamic that mirrors human social interaction. This internal microsociety isn’t merely an accidental byproduct of complex algorithms. The team posits that understanding this phenomenon will require applying insights from the social and organizational sciences, fields largely overlooked in AI development. To move beyond this limitation, researchers are advocating for systems that support multiple, converging, and diverging streams of deliberation, incorporating brainstorming, devil’s advocacy, and constructive conflict as designed features. The implications extend beyond improving individual model performance, suggesting a broader historical pattern.
This leads to the emergence of “human-AI centaurs”, composite actors that are neither purely human nor purely machine, inhabiting diverse roles and shifting configurations. Platforms like OpenClaw and Moltbook offer early glimpses of this future, where agents can renew, fork, and collaborate, spawning internal societies of thought to tackle complex problems. “Governance,” however, does not only mean what governments do, but building systems that ensure and verify outcomes of multi-stakeholder deliberation, and reliable scaffolds for automating delicate inter-agent collaborations.
knowledge accumulating across generations without any individual requirement to reconstruct the whole.
Michael Tomasello
Agent Institutions Enable Scalability Beyond Individual Capacity
The pursuit of increasingly powerful artificial intelligence is shifting focus from monolithic models to distributed systems, a change with profound implications for how AI scales and integrates with human society. This isn’t simply about building faster algorithms; it’s about creating frameworks for collective cognition, where intelligence emerges from the interaction of diverse perspectives. Recent work at Google’s Paradigms of Intelligence Team reveals a surprising dynamic within advanced reasoning models. The team demonstrated this by explicitly priming and amplifying multi-party conversation within the models, observing a corresponding increase in accuracy on complex reasoning tasks. Notably, this behavior wasn’t programmed; reinforcement learning, rewarding models solely for accuracy, spontaneously increased these conversational, multi-perspective behaviors. This rediscovery has significant implications for how we approach AI scalability.
However, effective groups exhibit hierarchy, specialization, and structured disagreement. The concept of intelligence itself is being redefined. “The identity of any agent matters less than its ability to fulfill a role protocol, just as a courtroom functions because ‘judge,’ ‘attorney,’ and ‘jury’ are well-defined slots, independent of who occupies them,” the researchers state, emphasizing the importance of agent institutions.
social intelligence in externalized form , encountering itself on a new substrate.
Source: https://arxiv.org/pdf/2603.20639
