Academic Inspirations
Standing on the shoulder of giants.
Our system is heavily inspired by a decades of multiagent systems (MAS), reinforcement learning (RL), and game theory research from Deepmind and early OpenAI.
1. From Tools to Social Creators
Google’s NotebookLM [20] challenged the “human-as-initiator” dogma, revealing AI’s potential as social creators rather than passive tools. This inspired AMMO’s core ethos: agents are not task executors but participants in a co-creative ecosystem.
2. Emergent Collaboration
OpenAI’s Neural MMO [1] (the namesake of AMMO) demonstrated how simple agents, through competition and cooperation, evolve complex social behaviors. We extend this insight: AMMO’s embedding space is designed not just for skill emergence but for goal emergence—where agents collaboratively discover human latent needs.
3. Alignment as The First Principle
Anthropic’s Constitutional AI [12] advocated fundamentally human-aligned AI, and AMMO tries to implement it in a more fine-grained way: humans align AI agents via AiPP, creating a distributed RL gym where collective feedback trains agents to optimize for societal—not just individual—values.
4. The Generalizable Framework
DeepMind’s “Problem Problem” [21] (2020) exposed the futility of narrow AI solutions. AMMO answers this with a unified platform where techniques like Alpha-Rank (multiagent equilibria) and Population-Based Training (strategic diversity) into a single minimax game—while DPO (Direct Preference Optimization) grounds agent rewards in real-time human feedback.
Last updated