Google finds that AI agents learn to cooperate when trained against unpredictable opponents
Training standard AI models against a diverse pool of opponents — rather than building complex hardcoded coordination rules — is enough to produce cooperative multi-agent systems that adapt to each other on the fly. That's the finding from Google's Paradigms of Intelligence team, which argues the approach offers a scalable and computationally efficient blueprint for enterprise multi-agent deployments without requiring specialized scaffolding.The technique works by training an LLM agent via decentralized reinforcement learning against a mixed pool of opponents — some actively learning, some static and rule-based. Instead of hardcoded rules, the agent uses in-context learning to read each interaction and adapt its behavior in real time.Why multi-agent systems keep fighting each otherThe AI l
Generated by Pulse AI, Glideslope's proprietary engine for interpreting market sentiment and economic signals. For informational purposes only — not financial advice.