{"id":3835,"date":"2025-10-11T21:52:23","date_gmt":"2025-10-11T21:52:23","guid":{"rendered":"https:\/\/violethoward.com\/new\/samsung-ai-researchers-new-open-reasoning-model-trm-outperforms-models-10000x-larger-on-specific-problems\/"},"modified":"2025-10-11T21:52:23","modified_gmt":"2025-10-11T21:52:23","slug":"samsung-ai-researchers-new-open-reasoning-model-trm-outperforms-models-10000x-larger-on-specific-problems","status":"publish","type":"post","link":"https:\/\/violethoward.com\/new\/samsung-ai-researchers-new-open-reasoning-model-trm-outperforms-models-10000x-larger-on-specific-problems\/","title":{"rendered":"Samsung AI researcher's new, open reasoning model TRM outperforms models 10,000X larger \u2014 on specific problems"},"content":{"rendered":"


\n
<\/p>\n

The trend of AI researchers developing new, small open source generative models that outperform far larger, proprietary peers continued this week with yet another staggering advancement.<\/p>\n

Alexia Jolicoeur-Martineau<\/b>, Senior AI Researcher at Samsung's Advanced\u200b Institute of Technology (SAIT)<\/b> in Montreal, Canada,\u200b <\/b>has introduced the Tiny Recursion Model (TRM)<\/b> \u2014 a neural network so small it contains just 7 million parameters (internal model settings), yet it competes with or surpasses cutting-edge language models 10,000 times larger in terms of their parameter count, including OpenAI's o3-mini and Google's Gemini 2.5 Pro,<\/b> on some of the toughest reasoning benchmarks in AI research. <\/p>\n

The goal is to show that very highly performant new AI models can be created affordably without massive investments in the graphics processing units (GPUs) and power needed to train the larger, multi-trillion parameter flagship models powering many LLM chatbots today. The results were described in a research paper published on open access website arxiv.org, entitled "Less is More: Recursive Reasoning with Tiny Networks<\/i>."<\/p>\n

"The idea that one must rely on massive foundational models trained for millions of dollars by some big corporation in order to solve hard tasks is a trap," wrote Jolicoeur-Martineau on the social network X. "Currently, there is too much focus on exploiting LLMs rather than devising and expanding new lines of direction."<\/p>\n

<\/div>\n

Jolicoeur-Martineau also added: "With recursive reasoning, it turns out that 'less is more'. A tiny model pretrained from scratch, recursing on itself and updating its answers over time, can achieve a lot without breaking the bank."<\/b><\/p>\n

TRM's code is available now on Github under an enterprise-friendly, commercially viable MIT License \u2014 meaning anyone from researchers to companies can take, modify it, and deploy it for their own purposes, even commercial applications.<\/p>\n

One Big Caveat<\/b><\/h3>\n

However, readers should be aware that TRM was designed specifically to perform well on structured, visual, grid-based problems like Sudoku, mazes, and puzzles on the ARC (Abstract and Reasoning Corpus)-AGI benchmark, the latter which offers tasks that should be easy for humans but difficult for AI models, such sorting colors on a grid based on a prior, but not identical, solution. <\/p>\n

From Hierarchy to Simplicity<\/b><\/h3>\n

The TRM architecture represents a radical simplification. <\/p>\n

It builds upon a technique called Hierarchical Reasoning Model (HRM)<\/b> introduced earlier this year, which showed that small networks could tackle logical puzzles like Sudoku and mazes. <\/p>\n

HRM relied on two cooperating networks\u2014one operating at high frequency, the other at low\u2014supported by biologically inspired arguments and mathematical justifications involving fixed-point theorems. Jolicoeur-Martineau found this unnecessarily complicated.<\/p>\n

TRM strips these elements away. Instead of two networks, it uses a single two-layer model<\/b> that recursively refines its own predictions. <\/p>\n

The model begins with an embedded question and an initial answer, represented by variables x<\/b>, y<\/b>, and z<\/b>. Through a series of reasoning steps, it updates its internal latent representation z<\/b> and refines the answer y<\/b> until it converges on a stable output. Each iteration corrects potential errors from the previous step, yielding a self-improving reasoning process without extra hierarchy or mathematical overhead.<\/p>\n

How Recursion Replaces Scale<\/b><\/h3>\n

The core idea behind TRM is that recursion can substitute for depth and size.<\/i><\/p>\n

By iteratively reasoning over its own output, the network effectively simulates a much deeper architecture without the associated memory or computational cost. This recursive cycle, run over as many as sixteen supervision steps, allows the model to make progressively better predictions \u2014 similar in spirit to how large language models use multi-step \u201cchain-of-thought\u201d reasoning, but achieved here with a compact, feed-forward design.<\/p>\n

The simplicity pays off in both efficiency and generalization. The model uses fewer layers, no fixed-point approximations, and no dual-network hierarchy. A lightweight halting mechanism<\/b> decides when to stop refining, preventing wasted computation while maintaining accuracy.<\/p>\n

Performance That Punches Above Its Weight<\/b><\/h3>\n

Despite its small footprint, TRM delivers benchmark results that rival or exceed models millions of times larger. In testing, the model achieved:<\/p>\n