
Sign up to save your podcasts
Or


Stop everything! The AI world is experiencing a structural shift, often hailed as the "Sputnik Moment" for the industry. This episode breaks down how the small team at Poetiq achieved State-of-the-Art (SOTA) performance on the ARC-AGI-2 benchmark, hitting a verified 54% accuracy. This landmark score decisively surpasses the previous SOTA of 45% held by Gemini 3 Deep Think.
The breakthrough is profound because it proves that intelligence can be decoupled from model scale. Poetiq did not train a massive new foundation model. Instead, they implemented a Meta-System—an intelligent orchestration layer that uses frontier LLMs (like Gemini, GPT, and Claude) to write and execute code iteratively at runtime. This method validates the Neuro-Symbolic approach and demonstrates the power of explicit verification over implicit guessing.
Tune in to discover how Poetiq inverted the economics of AI reasoning, achieving superior results at less than half the cost ($30.57 per problem versus the previous $77.16). This marks the definitive shift from the era of "Training Compute" to "Test-Time Compute". We explore the architecture, the use of code generation as a core capability, and why this breakthrough has put the final nail in the coffin of the "Scale is All You Need" dogma.
Primary: Poetiq, ARC-AGI-2, SOTA, AI Orchestration
• Secondary: Meta-System, Test-Time Compute, Gemini 3 Deep Think, Neuro-Symbolic AI, Code Generation, LLM Reasoning, Cost Efficiency, Decoupling Intelligence from Scale
By NovCogStop everything! The AI world is experiencing a structural shift, often hailed as the "Sputnik Moment" for the industry. This episode breaks down how the small team at Poetiq achieved State-of-the-Art (SOTA) performance on the ARC-AGI-2 benchmark, hitting a verified 54% accuracy. This landmark score decisively surpasses the previous SOTA of 45% held by Gemini 3 Deep Think.
The breakthrough is profound because it proves that intelligence can be decoupled from model scale. Poetiq did not train a massive new foundation model. Instead, they implemented a Meta-System—an intelligent orchestration layer that uses frontier LLMs (like Gemini, GPT, and Claude) to write and execute code iteratively at runtime. This method validates the Neuro-Symbolic approach and demonstrates the power of explicit verification over implicit guessing.
Tune in to discover how Poetiq inverted the economics of AI reasoning, achieving superior results at less than half the cost ($30.57 per problem versus the previous $77.16). This marks the definitive shift from the era of "Training Compute" to "Test-Time Compute". We explore the architecture, the use of code generation as a core capability, and why this breakthrough has put the final nail in the coffin of the "Scale is All You Need" dogma.
Primary: Poetiq, ARC-AGI-2, SOTA, AI Orchestration
• Secondary: Meta-System, Test-Time Compute, Gemini 3 Deep Think, Neuro-Symbolic AI, Code Generation, LLM Reasoning, Cost Efficiency, Decoupling Intelligence from Scale