Wink Pings

Samsung's Tiny Model Leaves Giants Stumped

An AI model with only 7 million parameters has outperformed DeepSeek and Gemini 2.5 Pro on the ARC-AGI reasoning benchmark. Its success isn't due to scale, but to recursive thinking.

Samsung's Tiny Recursive Model (TRM) has only 7 million parameters, ten thousand times smaller than mainstream large models, yet it performs better on ARC-AGI 1 and 2.

It works differently. Large models predict the next word, while TRM first drafts an answer, then reasons on an internal "scratchpad," repeatedly critiquing and refining its logic for up to 16 cycles. Each cycle produces a better answer.

This proves that intelligence doesn't necessarily come from parameter scale; architecture and reasoning loops are equally crucial. Efficient models can run on inexpensive hardware, opening the door to high-quality reasoning for more applications.

![Abstract from a research paper titled Less is More: Recursive Reasoning with Tiny Networks by Alexia Jolicoeur-Martineau from Samsung AI. Hierarchical diagram illustrates the Tiny Recursive Model (TRM) process: input of prediction and latent vectors leads to steps including update x and z, apply N up to 16 times, and output prediction y. Includes add norm operations, self-attention, and application of N to predict y. Cross-entropy loss and reverse engineering elements are shown.](https://wink.run/image?url=https%3A%2F%2Fpbs.twimg.com%2Fmedia%2FG2rx44fXIAAZGRV%3Fformat%3Djpg%26name%3Dlarge)

Some are questioning if this is just overfitting or specialized for benchmarking. The authors responded with a Sudoku test: the model, trained on only 1,000 samples, achieved an 87.4% accuracy on 423,000 test puzzles, demonstrating a degree of generalization.

![这是一张关于“Less is More: Recursive Reasoning with Tiny Networks”的研究论文摘要页面的截图。图中展示了论文标题、作者、所属机构以及电子邮件地址。此外,还包含了一个流程图,描述了Hierarchical Reasoning Model(HRM)的工作原理。该模型通过两个在不同频率上递归的小型神经网络来实现复杂的推理任务。在实验部分,使用了Sudoku、Maze和ARC-AGI等难题来评估模型的性能。结果表明,尽管使用了较小的参数量(约27M),但HRM在这些任务上的表现优于大多数LLMs。最后,提出了一个更简单的递归推理方法——Tiny Recursive Model(TRM),它在保持高精度的同时减少了参数数量。](https://wink.run/image?url=https%3A%2F%2Fpbs.twimg.com%2Fmedia%2FG2qMIx_XkAAcUwK%3Fformat%3Djpg%26name%3Dlarge)

But there's a trade-off. A 7 million parameter model with 16 iterations versus a 70 billion parameter model with a single pass—which one wins on real-world latency? Efficiency could be an issue.

This forces a rethink of the path for AI development. The large model race might just be a competition for computational waste. Small, elegant solutions could be the future, especially those that can run locally on devices.

The paper and code are now public, and the community can verify it for themselves. Whether it's a breakthrough or just hype will soon be clear.

发布时间: 2025-10-08 04:21