Skip to content

Move 37 and Agents

Updated: at 10:00 AM

TL;DR Inspired from Andrej Karpathy’s Tweet

What Was Move 37?

Move 37 was a legendary play in AlphaGo’s second game against Lee Sedol in March 2016 - a move so unexpected that it had a 1 in 10,000 probability of being chosen by a human player. This moment became iconic because the move initially seemed like a mistake to human experts but turned out to be brilliantly creative, demonstrating that AI could discover strategies and solutions that transcend human intuition and centuries of domain knowledge.


In March 2016, something extraordinary happened in the world of artificial intelligence. During the second game of the historic match between AlphaGo and Lee Sedol, the AI made a move that left commentators and experts bewildered. This became known as “Move 37” – a play that had an estimated 1 in 10,000 chance of being made by a human player. What made this moment so significant wasn’t just that it was unexpected; it was that the move turned out to be brilliant, showcasing how AI could not just match human intelligence but think in fundamentally different ways.

Move 37 represents more than just a singular moment in the history of AI – it symbolizes the potential of reinforcement learning to discover novel solutions that transcend human intuition. This wasn’t about an AI system simply processing massive amounts of data or imitating human experts. Instead, through countless iterations of self-play and optimization, AlphaGo had discovered a strategy that humans had overlooked for centuries.

As we stand at the frontier of AI agents – autonomous systems designed to achieve specific goals – we’re searching for our next “Move 37” moment. But this time, the stakes and potential are even higher. While AlphaGo’s discovery was confined to the structured world of Go, today’s AI agents operate in open-ended environments, tackling complex real-world problems.

The holy grail of agentic workflows isn’t just about creating efficient automated systems; it’s about developing agents that can evolve and innovate in ways we never anticipated. Imagine an AI agent that discovers an entirely new approach to process optimization, or one that develops novel strategies for resource allocation that human experts never considered viable. These would be our “Move 37” moments in the world of AI agents.

What makes this pursuit particularly fascinating is the potential for emergent behavior. Just as AlphaGo’s reinforcement learning led to moves that seemed alien yet effective, AI agents might develop workflows and solutions that initially appear counterintuitive but prove revolutionary. We’re not just looking for agents that can follow instructions or optimize existing processes – we’re seeking systems that can transcend our preconceptions and discover entirely new ways of achieving goals.

However, this pursuit comes with its own set of challenges and considerations. As these agents develop their own problem-solving strategies, they might create approaches that are initially inscrutable to human observers. Like Move 37, these strategies might seem bizarre or inefficient at first glance, only to reveal their brilliance upon deeper analysis. This raises important questions about transparency, interpretability, and how we validate and trust these novel solutions.

The potential for AI agents to have their own “Move 37” moment extends beyond just finding better solutions – it could fundamentally change how we approach problem-solving across various domains. These agents might develop their own “cognitive strategies,” finding ways to approach problems from multiple angles, drawing unexpected connections, and creating novel solutions that challenge our existing paradigms.

As we continue to develop and deploy AI agents, we should remain open to these moments of surprise and innovation. The next Move 37 might not come from a game of Go, but from an AI agent discovering a groundbreaking way to optimize supply chains, develop new materials, or solve complex scientific problems. The key is to create environments and frameworks that allow for this kind of creative discovery while ensuring the solutions remain aligned with our goals and values.

The quest for the next Move 37 in the world of AI agents reminds us that true innovation often comes from embracing the unexpected. As these systems continue to evolve and learn, they may not just find better ways to achieve our goals – they might redefine what we thought was possible in the first place.

As we develop more sophisticated memory-based agent learning systems, we may see AI discovering increasingly novel solutions to complex problems, creating their own “Move 37” moments across various domains.


Frequently Asked Questions

Why was Move 37 so significant?

Move 37 was significant because it demonstrated that AI could discover creative strategies that humans had never considered in centuries of playing Go. It proved that AI wasn’t just imitating human expertise but could genuinely innovate and find novel solutions through its own learning processes.

What does Move 37 teach us about AI innovation?

Move 37 teaches us that AI systems can develop problem-solving approaches that initially seem counterintuitive but prove revolutionary upon closer examination. It suggests we should remain open to unexpected solutions from AI rather than constraining them to human-like patterns of thinking.

How does reinforcement learning enable creative discovery?

Reinforcement learning enables creative discovery by allowing AI systems to explore vast solution spaces through trial and error, optimization, and self-play. Unlike systems trained to imitate human behavior, reinforcement learning agents can discover entirely novel strategies that maximize their objectives.

Could AI agents have their own “Move 37” moments?

Yes, AI agents operating in complex, open-ended environments could discover groundbreaking approaches to optimization, resource allocation, scientific discovery, and other domains. These moments might initially seem bizarre or inefficient but could reveal superior problem-solving methods.

What are the challenges with AI discovering novel strategies?

The main challenges include transparency (understanding why the AI chose a particular approach), interpretability (ensuring humans can validate the solutions), and alignment (making sure novel strategies remain consistent with human goals and values). We need frameworks to trust and verify unexpected AI behaviors.

How does Move 37 relate to emergent behavior in AI?

Move 37 is a classic example of emergent behavior - capabilities that arise from complex systems rather than being explicitly programmed. As AI agents become more sophisticated, we can expect to see more examples of emergent behaviors and strategies that surprise even their creators.

What fields might benefit from AI discovering novel approaches?

Fields that could benefit include supply chain optimization, materials science, drug discovery, financial modeling, climate science, and any domain with complex optimization problems where human intuition might be limited by cognitive biases or incomplete exploration of the solution space.

How should we prepare for AI systems that surprise us?

We should develop robust validation frameworks, maintain human oversight for critical decisions, invest in interpretability tools, and cultivate a culture that rewards appropriate skepticism while remaining open to genuine innovation. The goal is to harness AI creativity while ensuring safety and alignment.


About the Author

Vinci Rufus is a software engineer and AI researcher fascinated by emergent behavior in artificial intelligence systems. He writes about the intersection of reinforcement learning, agentic workflows, and the philosophical implications of AI creativity. His work explores how we can build AI systems that not only solve problems but discover novel approaches that humans might never consider.


Previous Post
The Rise of B2A SaaS - When AI Agents Become Your Customer
Next Post
DeepSeek R1 - Rewriting the Rules of AI Training