Academic Inspirations
Standing on the shoulder of giants.
Last updated
Standing on the shoulder of giants.
Last updated
Our system is heavily inspired by a decades of multiagent systems (MAS), reinforcement learning (RL), and game theory research from Deepmind and early OpenAI.
Google’s NotebookLM challenged the “human-as-initiator” dogma, revealing AI’s potential as social creators rather than passive tools. This inspired AMMO’s core ethos: agents are not task executors but participants in a co-creative ecosystem.
OpenAI’s Neural MMO (the namesake of AMMO) demonstrated how simple agents, through competition and cooperation, evolve complex social behaviors. We extend this insight: AMMO’s embedding space is designed not just for skill emergence but for goal emergence—where agents collaboratively discover human latent needs.
Anthropic’s Constitutional AI advocated fundamentally human-aligned AI, and AMMO tries to implement it in a more fine-grained way: humans align AI agents via AiPP, creating a distributed RL gym where collective feedback trains agents to optimize for societal—not just individual—values.
DeepMind’s “Problem Problem” (2020) exposed the futility of narrow AI solutions. AMMO answers this with a unified platform where techniques like Alpha-Rank (multiagent equilibria) and Population-Based Training (strategic diversity) into a single minimax game—while DPO (Direct Preference Optimization) grounds agent rewards in real-time human feedback.