{"id":2047,"date":"2025-06-24T02:36:09","date_gmt":"2025-06-24T02:36:09","guid":{"rendered":"https:\/\/violethoward.com\/new\/beyond-static-ai-mits-new-framework-lets-models-teach-themselves\/"},"modified":"2025-06-24T02:36:09","modified_gmt":"2025-06-24T02:36:09","slug":"beyond-static-ai-mits-new-framework-lets-models-teach-themselves","status":"publish","type":"post","link":"https:\/\/violethoward.com\/new\/beyond-static-ai-mits-new-framework-lets-models-teach-themselves\/","title":{"rendered":"Beyond static AI: MIT&#8217;s new framework lets models teach themselves"},"content":{"rendered":" \r\n<br><div>\n\t\t\t\t<div id=\"boilerplate_2682874\" class=\"post-boilerplate boilerplate-before\">\n<p><em>Join the event trusted by enterprise leaders for nearly two decades. VB Transform brings together the people building real enterprise AI strategy.\u00a0Learn more<\/em><\/p>\n\n\n\n<hr class=\"wp-block-separator has-css-opacity is-style-wide\"\/>\n<\/div><p>Researchers at MIT have developed a framework called Self-Adapting Language Models (SEAL) that enables large language models (LLMs) to continuously learn and adapt by updating their own internal parameters. SEAL teaches an LLM to generate its own training data and update instructions, allowing it to permanently absorb new knowledge and learn new tasks.<\/p>\n\n\n\n<p>This framework could be useful for enterprise applications, particularly for AI agents that operate in dynamic environments, where they must constantly process new information and adapt their behavior.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\" id=\"h-the-challenge-of-adapting-llms\">The challenge of adapting LLMs<\/h2>\n\n\n\n<p>While large language models have shown remarkable abilities, adapting them to specific tasks, integrating new information, or mastering novel reasoning skills remains a significant hurdle.<\/p>\n\n\n\n<p>Currently, when faced with a new task, LLMs typically learn from data \u201cas-is\u201d through methods like finetuning or in-context learning. However, the provided data is not always in an optimal format for the model to learn efficiently. Existing approaches don\u2019t allow the model to develop its own strategies for best transforming and learning from new information.<\/p>\n\n\n\n<p>\u201cMany enterprise use cases demand more than just factual recall\u2014they require deeper, persistent adaptation,\u201d Jyo Pari, PhD student at MIT and co-author of the paper, told VentureBeat. \u201cFor example, a coding assistant might need to internalize a company\u2019s specific software framework, or a customer-facing model might need to learn a user\u2019s unique behavior or preferences over time.\u201d\u00a0<\/p>\n\n\n\n<p>In such cases, temporary retrieval falls short, and the knowledge needs to be \u201cbaked into\u201d the model\u2019s weights so that it influences all future responses.\u00a0<\/p>\n\n\n\n<h2 class=\"wp-block-heading\" id=\"h-creating-self-adapting-language-models\">Creating self-adapting language models<\/h2>\n\n\n\n<p>\u201cAs a step towards scalable and efficient adaptation of language models, we propose equipping LLMs with the ability to generate their own training data and finetuning directives for using such data,\u201d the MIT researchers state in their paper.<\/p>\n\n\n\n<figure class=\"wp-block-image size-large\"><img fetchpriority=\"high\" decoding=\"async\" height=\"214\" width=\"800\" src=\"https:\/\/venturebeat.com\/wp-content\/uploads\/2025\/06\/image_b1ee4c.png?w=800\" alt=\"Overview of SEAL framework (source: arXiv)\" class=\"wp-image-3012752\" srcset=\"https:\/\/venturebeat.com\/wp-content\/uploads\/2025\/06\/image_b1ee4c.png 1202w, https:\/\/venturebeat.com\/wp-content\/uploads\/2025\/06\/image_b1ee4c.png?resize=300,80 300w, https:\/\/venturebeat.com\/wp-content\/uploads\/2025\/06\/image_b1ee4c.png?resize=768,206 768w, https:\/\/venturebeat.com\/wp-content\/uploads\/2025\/06\/image_b1ee4c.png?resize=800,214 800w, https:\/\/venturebeat.com\/wp-content\/uploads\/2025\/06\/image_b1ee4c.png?resize=400,107 400w, https:\/\/venturebeat.com\/wp-content\/uploads\/2025\/06\/image_b1ee4c.png?resize=750,201 750w, https:\/\/venturebeat.com\/wp-content\/uploads\/2025\/06\/image_b1ee4c.png?resize=578,155 578w, https:\/\/venturebeat.com\/wp-content\/uploads\/2025\/06\/image_b1ee4c.png?resize=930,249 930w, https:\/\/venturebeat.com\/wp-content\/uploads\/2025\/06\/image_b1ee4c.png?resize=1200,322 1200w\" sizes=\"(max-width: 800px) 100vw, 800px\"\/><figcaption class=\"wp-element-caption\"><em>Overview of SEAL framework Source: arXiv<\/em><\/figcaption><\/figure>\n\n\n\n<p>The researchers\u2019 solution is SEAL, short for Self-Adapting Language Models. It uses a reinforcement learning (RL) algorithm to train an LLM to generate \u201cself-edits\u201d\u2014natural-language instructions that specify how the model should update its own weights. These self-edits can restructure new information, create synthetic training examples, or even define the technical parameters for the learning process itself.<\/p>\n\n\n\n<p>Intuitively, SEAL teaches a model how to create its own personalized study guide. Instead of just reading a new document (the raw data), the model learns to rewrite and reformat that information into a style it can more easily absorb and internalize. This process brings together several key areas of AI research, including synthetic data generation, reinforcement learning and test-time training (TTT).<\/p>\n\n\n\n<p>The framework operates on a two-loop system. In an \u201cinner loop,\u201d the model uses a self-edit to perform a small, temporary update to its weights. In an \u201couter loop,\u201d the system evaluates whether that update improved the model\u2019s performance on a target task. If it did, the model receives a positive reward, reinforcing its ability to generate that kind of effective self-edit in the future. Over time, the LLM becomes an expert at teaching itself.<\/p>\n\n\n\n<p>In their study, the researchers used a single model for the entire SEAL framework. However, they also note that this process can be decoupled into a \u201cteacher-student\u201d model. A specialized teacher model could be trained to generate effective self-edits for a separate student model, which would then be updated. This approach could allow for more specialized and efficient adaptation pipelines in enterprise settings.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\" id=\"h-seal-in-action\">SEAL in action<\/h2>\n\n\n\n<p>The researchers tested SEAL in two key domains: knowledge incorporation (the ability to permanently integrate new facts) and few-shot learning (the ability to generalize from a handful of examples).<\/p>\n\n\n\n<figure class=\"wp-block-image size-large\"><img loading=\"lazy\" decoding=\"async\" height=\"152\" width=\"800\" src=\"https:\/\/venturebeat.com\/wp-content\/uploads\/2025\/06\/image_d93126.png?w=800\" alt=\"SEAL in knowledge incorporation (source: arXiv)\" class=\"wp-image-3012753\" srcset=\"https:\/\/venturebeat.com\/wp-content\/uploads\/2025\/06\/image_d93126.png 1220w, https:\/\/venturebeat.com\/wp-content\/uploads\/2025\/06\/image_d93126.png?resize=300,57 300w, https:\/\/venturebeat.com\/wp-content\/uploads\/2025\/06\/image_d93126.png?resize=768,146 768w, https:\/\/venturebeat.com\/wp-content\/uploads\/2025\/06\/image_d93126.png?resize=800,152 800w, https:\/\/venturebeat.com\/wp-content\/uploads\/2025\/06\/image_d93126.png?resize=400,76 400w, https:\/\/venturebeat.com\/wp-content\/uploads\/2025\/06\/image_d93126.png?resize=750,143 750w, https:\/\/venturebeat.com\/wp-content\/uploads\/2025\/06\/image_d93126.png?resize=578,110 578w, https:\/\/venturebeat.com\/wp-content\/uploads\/2025\/06\/image_d93126.png?resize=930,177 930w\" sizes=\"auto, (max-width: 800px) 100vw, 800px\"\/><figcaption class=\"wp-element-caption\"><em>SEAL in knowledge incorporation Source: arXiv<\/em><\/figcaption><\/figure>\n\n\n\n<p>For knowledge incorporation, the goal was to see if the model could answer questions about a text passage without having access to the passage during questioning. Finetuning Llama-3.2-1B on the raw text provided only a marginal improvement over the base model.\u00a0<\/p>\n\n\n\n<p>However, when the SEAL model created \u201cself-edits\u201d by generating several \u201cimplications\u201d from a passage and was trained on this synthetic data, its accuracy jumped to 47%. Notably, this outperformed results from using synthetic data generated by the much larger GPT-4.1, suggesting the model learned to create superior training material for itself.<\/p>\n\n\n\n<figure class=\"wp-block-image size-large\"><img loading=\"lazy\" decoding=\"async\" height=\"218\" width=\"800\" src=\"https:\/\/venturebeat.com\/wp-content\/uploads\/2025\/06\/image_1e87a7.png?w=800\" alt=\"SEAL in few-shot learning (source: arXiv)\" class=\"wp-image-3012754\" srcset=\"https:\/\/venturebeat.com\/wp-content\/uploads\/2025\/06\/image_1e87a7.png 1226w, https:\/\/venturebeat.com\/wp-content\/uploads\/2025\/06\/image_1e87a7.png?resize=300,82 300w, https:\/\/venturebeat.com\/wp-content\/uploads\/2025\/06\/image_1e87a7.png?resize=768,209 768w, https:\/\/venturebeat.com\/wp-content\/uploads\/2025\/06\/image_1e87a7.png?resize=800,218 800w, https:\/\/venturebeat.com\/wp-content\/uploads\/2025\/06\/image_1e87a7.png?resize=400,109 400w, https:\/\/venturebeat.com\/wp-content\/uploads\/2025\/06\/image_1e87a7.png?resize=750,204 750w, https:\/\/venturebeat.com\/wp-content\/uploads\/2025\/06\/image_1e87a7.png?resize=578,157 578w, https:\/\/venturebeat.com\/wp-content\/uploads\/2025\/06\/image_1e87a7.png?resize=930,253 930w\" sizes=\"auto, (max-width: 800px) 100vw, 800px\"\/><figcaption class=\"wp-element-caption\"><em>SEAL in few-shot learning Source: arXiv<\/em><\/figcaption><\/figure>\n\n\n\n<p>For few-shot learning, the researchers tested SEAL on examples from the Abstract Reasoning Corpus (ARC), where the model must solve visual puzzles. In the self-edit phase, the model had to generate the entire adaptation strategy, including which data augmentations and tools to use and what learning rate to apply.\u00a0<\/p>\n\n\n\n<p>SEAL achieved a 72.5% success rate, a dramatic improvement over the 20% rate achieved without RL training and the 0% rate of standard in-context learning.<\/p>\n\n\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter size-full\"><img loading=\"lazy\" decoding=\"async\" width=\"508\" height=\"510\" src=\"https:\/\/venturebeat.com\/wp-content\/uploads\/2025\/06\/image_969ef3.png\" alt=\"SEAL (red line) continues to improve across RL cycles (source: arXiv)\" class=\"wp-image-3012755\" srcset=\"https:\/\/venturebeat.com\/wp-content\/uploads\/2025\/06\/image_969ef3.png 508w, https:\/\/venturebeat.com\/wp-content\/uploads\/2025\/06\/image_969ef3.png?resize=300,301 300w, https:\/\/venturebeat.com\/wp-content\/uploads\/2025\/06\/image_969ef3.png?resize=52,52 52w, https:\/\/venturebeat.com\/wp-content\/uploads\/2025\/06\/image_969ef3.png?resize=160,160 160w, https:\/\/venturebeat.com\/wp-content\/uploads\/2025\/06\/image_969ef3.png?resize=400,402 400w\" sizes=\"auto, (max-width: 508px) 100vw, 508px\"\/><figcaption class=\"wp-element-caption\"><em>SEAL (red line) continues to improve across RL cycles Source: arXiv<\/em><\/figcaption><\/figure><\/div>\n\n\n<h2 class=\"wp-block-heading\" id=\"h-implications-for-the-enterprise\">Implications for the enterprise<\/h2>\n\n\n\n<p>Some experts project that the supply of high-quality, human-generated training data could be exhausted in the coming years. Progress may soon depend on \u201ca model\u2019s capacity to generate its own high-utility training signal,\u201d as the researchers put it. They add, \u201cA natural next step is to meta-train a dedicated SEAL synthetic-data generator model that produces fresh pretraining corpora, allowing future models to scale and achieve greater data efficiency without relying on additional human text.\u201d<\/p>\n\n\n\n<p>For example, the researchers propose that an LLM could ingest complex documents like academic papers or financial reports and autonomously generate thousands of explanations and implications to deepen its understanding.\u00a0<\/p>\n\n\n\n<p>\u201cThis iterative loop of self-expression and self-refinement could allow models to keep improving on rare or underrepresented topics even in the absence of additional external supervision,\u201d the researchers explain.<\/p>\n\n\n\n<p>This capability is especially promising for building AI agents. Agentic systems must incrementally acquire and retain knowledge as they interact with their environment. SEAL provides a mechanism for this. After an interaction, an agent could synthesize a self-edit to trigger a weight update, allowing it to internalize the lessons learned. This enables the agent to evolve over time, improve its performance based on experience, and reduce its reliance on static programming or repeated human guidance.<\/p>\n\n\n\n<p>\u201cSEAL demonstrates that large language models need not remain static after pretraining,\u201d the researchers write. \u201cBy learning to generate their own synthetic self-edit data and to apply it through lightweight weight updates, they can autonomously incorporate new knowledge and adapt to novel tasks.\u201d<\/p>\n\n\n\n<h2 class=\"wp-block-heading\" id=\"h-limitations-of-seal\">Limitations of SEAL<\/h2>\n\n\n\n<p>That said, SEAL is not a universal solution. For example, it can suffer from \u201ccatastrophic forgetting,\u201d where constant retraining cycles can result in the model learning its earlier knowledge.<\/p>\n\n\n\n<p>\u201cIn our current implementation, we encourage a hybrid approach,\u201d Pari said. \u201cEnterprises should be selective about what knowledge is important enough to integrate permanently.\u201d\u00a0<\/p>\n\n\n\n<p>Factual and evolving data can remain in external memory through RAG, while long-lasting, behavior-shaping knowledge is better suited for weight-level updates via SEAL.\u00a0<\/p>\n\n\n\n<p>\u201cThis kind of hybrid memory strategy ensures the right information is persistent without overwhelming the model or introducing unnecessary forgetting,\u201d he said.<\/p>\n\n\n\n<p>It is also worth noting that SEAL takes a non-trivial amount of time to tune the self-edit examples and train the model. This makes continuous, real-time editing infeasible in most production settings.<\/p>\n\n\n\n<p>\u201cWe envision a more practical deployment model where the system collects data over a period\u2014say, a few hours or a day\u2014and then performs targeted self-edits during scheduled update intervals,\u201d Pari said. \u201cThis approach allows enterprises to control the cost of adaptation while still benefiting from SEAL\u2019s ability to internalize new knowledge.\u201d<\/p>\n<div id=\"boilerplate_2660155\" class=\"post-boilerplate boilerplate-after\"><div class=\"Boilerplate__newsletter-container vb\">\n<div class=\"Boilerplate__newsletter-main\">\n<p><strong>Daily insights on business use cases with VB Daily<\/strong><\/p>\n<p class=\"copy\">If you want to impress your boss, VB Daily has you covered. We give you the inside scoop on what companies are doing with generative AI, from regulatory shifts to practical deployments, so you can share insights for maximum ROI.<\/p>\n<p class=\"Form__newsletter-legal\">Read our Privacy Policy<\/p>\n<p class=\"Form__success\" id=\"boilerplateNewsletterConfirmation\">\n\t\t\t\t\tThanks for subscribing. Check out more VB newsletters here.\n\t\t\t\t<\/p>\n<p class=\"Form__error\">An error occured.<\/p>\n<\/p><\/div>\n<div class=\"image-container\">\n\t\t\t\t\t<img decoding=\"async\" src=\"https:\/\/venturebeat.com\/wp-content\/themes\/vb-news\/brand\/img\/vb-daily-phone.png\" alt=\"\"\/>\n\t\t\t\t<\/div>\n<\/p><\/div>\n<\/div>\t\t\t<\/div>\r\n<br>\r\n<br><a href=\"https:\/\/venturebeat.com\/ai\/beyond-static-ai-mits-new-framework-lets-models-teach-themselves\/\">Source link <\/a>","protected":false},"excerpt":{"rendered":"<p>Join the event trusted by enterprise leaders for nearly two decades. VB Transform brings together the people building real enterprise AI strategy.\u00a0Learn more Researchers at MIT have developed a framework called Self-Adapting Language Models (SEAL) that enables large language models (LLMs) to continuously learn and adapt by updating their own internal parameters. SEAL teaches an [&hellip;]<\/p>\n","protected":false},"author":1,"featured_media":2048,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"_jetpack_memberships_contains_paid_content":false,"footnotes":""},"categories":[33],"tags":[],"class_list":["post-2047","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-ai-automation"],"aioseo_notices":[],"jetpack_featured_media_url":"https:\/\/violethoward.com\/new\/wp-content\/uploads\/2025\/06\/self-adapting-language-models.png","jetpack_sharing_enabled":true,"_links":{"self":[{"href":"https:\/\/violethoward.com\/new\/wp-json\/wp\/v2\/posts\/2047","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/violethoward.com\/new\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/violethoward.com\/new\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/violethoward.com\/new\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/violethoward.com\/new\/wp-json\/wp\/v2\/comments?post=2047"}],"version-history":[{"count":0,"href":"https:\/\/violethoward.com\/new\/wp-json\/wp\/v2\/posts\/2047\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/violethoward.com\/new\/wp-json\/wp\/v2\/media\/2048"}],"wp:attachment":[{"href":"https:\/\/violethoward.com\/new\/wp-json\/wp\/v2\/media?parent=2047"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/violethoward.com\/new\/wp-json\/wp\/v2\/categories?post=2047"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/violethoward.com\/new\/wp-json\/wp\/v2\/tags?post=2047"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}<!-- This website is optimized by Airlift. Learn more: https://airlift.net. Template:. Learn more: https://airlift.net. Template: 69e302c146fa5c92dc28ac12. Config Timestamp: 2026-04-18 04:04:16 UTC, Cached Timestamp: 2026-04-29 10:27:14 UTC -->