MIT's AI Begins Writing Its Own Code for Upgrades, Even Outperforming GPT-4.1
MIT's research team has developed an AI called SEAL, which can read new knowledge, modify code, and perform gradient updates to achieve self-learning. The result is a 40% improvement in factual recall, and it surpasses GPT-4.1 using data generated by itself.
MIT researchers have recently created something new called SEAL (Self-Adapting Language Models). The unique aspect of this is that it doesn't require human fine-tuning; instead, it reads new information, rewrites it in its own words, and then runs gradient updates to optimize itself. Essentially, it's self-learning.
The results? A 40% improvement in factual recall, and after training with data it generated, its performance outperforms GPT-4.1. Moreover, it can learn new tasks entirely without human intervention.

Current AI models are fixed after training, but SEAL flips this process. It runs a reinforcement learning loop: first, it generates a 'self-editing' instruction to tell itself how to update; then, it tests the results; and finally, it reinforces only those parts that improve performance.

How does it work? For example, when SEAL reads new content about the Apollo program, it first reformats the information into concise logical key points, like taking study notes for itself. Then, it uses its notes to fine-tune itself.

Few-shot learning has also been upgraded. SEAL doesn't rely on fixed heuristic methods but decides its training strategy on its own: selecting what data augmentation methods to use, how to optimize, and even setting its own learning rate. The result is a 72.5% success rate, 3.6 times better than the baseline.

What's even more impressive is that after just two rounds of self-reinforcement, SEAL surpasses GPT-4.1 using data it generated. It has learned how to write data that is easier for itself to learn—by reformating facts into simple, atomic truths.

There's also a key point: during self-updates, SEAL retains most of what it previously learned. This is a major advancement for continuous learning. Although some forgetting still occurs, the retention curve looks promising.

Of course, there are concerns about what might happen if such an AI trains itself with biased or incorrect information. Indeed, without the ability to detect biased or erroneous information, LLMs could harm themselves rather than improve.
But regardless, the idea of large models that can fine-tune themselves is no longer science fiction. We have just entered the era of self-evolving models.
发布时间: 2025-10-13 15:13