
Sign up to save your podcasts
Or


Have you ever wondered why we, as humans, are so obsessed with games? From the strategic depth of Chess to the frantic tapping of Flappy Bird, we spend countless hours in digital and physical playgrounds. According to recent research, this isn't just about killing time—it’s actually a cornerstone of our General Intelligence.
Games are "structured microcosms" of the real world. When we play, we are actually practicing skills like resource management, social deduction, and physical navigation in a safe, fun environment. Now, researchers from institutions like MIT and Harvard are using this "Multiverse of Human Games" to see if artificial intelligence can finally keep up with us.
Evaluating how "smart" an AI really is has become a massive challenge. Traditional tests often focus on narrow tasks like solving a specific math problem or writing code. But being good at one thing doesn't mean a machine has the versatility of a human adult.
To bridge this gap, researchers built the AI GAMESTORE. This platform uses LLMs (Large Language Models) to automatically source and adapt popular games from the Apple App Store and Steam into standardized tests for machines. By having artificial intelligence play 100 different games—ranging from Angry Birds clones to complex puzzles—the researchers could measure its ability to learn and adapt just like a human would.
The researchers pitted seven of the world's most advanced LLMs (including frontier models like GPT-5.2 and Gemini 2.5 Pro) against 106 human players. The goal was simple: play the first two minutes of a new game and see who scores higher.
The results were a wake-up call for the tech world:
You might think a supercomputer could easily beat a human at a "casual" mobile game, but the AI GAMESTORE revealed three major "cognitive bottlenecks" where machines fail:
This research shows that while artificial intelligence is getting better at talking and coding, it still lacks the "cognitive versatility" of a typical human. The "Multiverse of Human Games" provides a way to track this progress through a "living" benchmark that can't be easily cheated or memorized.
The ultimate goal isn't just to build a better gamer. It’s to develop AI that can interact with the world as flexibly, safely, and intuitively as we do. Until then, it looks like your high score on the App Store is safe!
By mstraton8112Have you ever wondered why we, as humans, are so obsessed with games? From the strategic depth of Chess to the frantic tapping of Flappy Bird, we spend countless hours in digital and physical playgrounds. According to recent research, this isn't just about killing time—it’s actually a cornerstone of our General Intelligence.
Games are "structured microcosms" of the real world. When we play, we are actually practicing skills like resource management, social deduction, and physical navigation in a safe, fun environment. Now, researchers from institutions like MIT and Harvard are using this "Multiverse of Human Games" to see if artificial intelligence can finally keep up with us.
Evaluating how "smart" an AI really is has become a massive challenge. Traditional tests often focus on narrow tasks like solving a specific math problem or writing code. But being good at one thing doesn't mean a machine has the versatility of a human adult.
To bridge this gap, researchers built the AI GAMESTORE. This platform uses LLMs (Large Language Models) to automatically source and adapt popular games from the Apple App Store and Steam into standardized tests for machines. By having artificial intelligence play 100 different games—ranging from Angry Birds clones to complex puzzles—the researchers could measure its ability to learn and adapt just like a human would.
The researchers pitted seven of the world's most advanced LLMs (including frontier models like GPT-5.2 and Gemini 2.5 Pro) against 106 human players. The goal was simple: play the first two minutes of a new game and see who scores higher.
The results were a wake-up call for the tech world:
You might think a supercomputer could easily beat a human at a "casual" mobile game, but the AI GAMESTORE revealed three major "cognitive bottlenecks" where machines fail:
This research shows that while artificial intelligence is getting better at talking and coding, it still lacks the "cognitive versatility" of a typical human. The "Multiverse of Human Games" provides a way to track this progress through a "living" benchmark that can't be easily cheated or memorized.
The ultimate goal isn't just to build a better gamer. It’s to develop AI that can interact with the world as flexibly, safely, and intuitively as we do. Until then, it looks like your high score on the App Store is safe!