Tiny Recursive Model ; Small, Simple… and Surprisingly Strong 🤖🧩

Small, simple… and surprisingly strong.

A few months back, the Hierarchical Reasoning Model (HRM) (Paper) showed that a 27M-param network could beat much larger LLMs on grid-based puzzles like Sudoku, Maze, and ARC-AGI. Now a follow-up paper proposes the Tiny Recursive Model (TRM) (Paper): a 7M-param, single 2-layer network that out-generalizes HRM on the same tasks.

What “recursion” means here

TRM iteratively (1) updates a latent reasoning state from the current question + answer, then (2) refines the answer from that state—repeating up to 16 times.
Training runs several no-grad refinement loops and then backprops through one full recursion per step, which seems to help generalization.

How TRM differs from HRM

  • Architecture: one tiny 2-layer net vs. HRM’s two 4-layer nets.
  • Training: TRM backprops through the full recursion; HRM only backprops the final evaluations (earlier steps are detached).
  • Halting: TRM learns when to stop with a simple BCE “halt” head—no extra forward pass.

Results (test accuracy)

  • Sudoku-Extreme: TRM 87.4% (vs. HRM 55.0%)
  • Maze-Hard: TRM 85.3% (vs. 74.5%)
  • ARC-AGI-1/2: TRM 44.6% / 7.8% (vs. 40.3% / 5.0%)

Neat tidbits

  • Fewer layers ≠ worse: dropping from 4 to 2 layers improved Sudoku generalization (79.5% → 87.4%), likely reducing overfitting.
  • Replacing self-attention with an MLP boosted accuracy on fixed-size grids like Sudoku (74.7% → 87.4%), though attention still helps on larger 30×30 grids (Maze/ARC).

Comparing these tiny, task-specific models to general-purpose LLMs isn’t apples-to-apples. But TRM is a compelling proof that clever structure + recursion can rival (and sometimes surpass) sheer scale on the right problems.
Modules like this could easily become components within larger, tool-using AI systems.

Congrats to Alexia Jolicoeur-Martineau and the team!

Paper


#AI #MachineLearning #DeepLearning #Reasoning #Recursion #LLM #AgenticAI #GenAI #Walmart #WalmartGlobalTech




Enjoy Reading This Article?

Here are some more articles you might like to read next:

  • Google Gemini updates: Flash 1.5, Gemma 2 and Project Astra
  • Displaying External Posts on Your al-folio Blog
  • GLM-4.5 ; A Unified Open-Source Powerhouse for Agents, Reasoning & Coding 🤖✨
  • AlphaGo Moment for Model Architecture Discovery ; The Rise of Autonomous AI Scientists 🤖🚀
  • Reinforcement pre-training - baking the cherry into the cake