Translater

Mittwoch, 2. Juli 2025

๐Ÿš€ ๐๐ซ๐ž๐š๐ค๐ญ๐ก๐ซ๐จ๐ฎ๐ ๐ก ๐ข๐ง ๐€๐ˆ ๐‘๐ž๐ฌ๐ž๐š๐ซ๐œ๐ก: ๐’๐ž๐ฅ๐Ÿ-๐€๐๐š๐ฉ๐ญ๐ข๐ง๐  ๐‹๐š๐ง๐ ๐ฎ๐š๐ ๐ž ๐Œ๐จ๐๐ž๐ฅ๐ฌ (๐’๐„๐€๐‹) – ๐“๐จ๐ฐ๐š๐ซ๐๐ฌ ๐“๐ซ๐ฎ๐ฅ๐ฒ ๐€๐ฎ๐ญ๐จ๐ง๐จ๐ฆ๐จ๐ฎ๐ฌ ๐’๐ž๐ฅ๐Ÿ-๐ˆ๐ฆ๐ฉ๐ซ๐จ๐ฏ๐ž๐ฆ๐ž๐ง๐ญ

Cover image on self-training AI with the caption Singularity achieved – AI trains itself - Trauth Research


Recently published by MIT and the Improbable AI Lab, the study “Self-Adapting Language Models (SEAL)” represents a major paradigm shift in artificial intelligence. For the first time, a language model has been trained to autonomously improve itself by generating its own fine-tuning data and updating its own parameters.

๐Ÿ” What makes SEAL unique?

  • Until now, LLMs such as GPT or Llama have been essentially static after training. They could be fine-tuned for specific tasks, but lacked the capacity to develop their own strategies for adaptation.

  • SEAL fundamentally changes this: The model generates its own fine-tuning data and autonomously determines how to best adapt to new tasks or knowledge.

  • This self-adaptation is implemented via a novel reinforcement learning loop: The model creates self-edit instructions, evaluates their impact, and directly rewards improvements – all without external supervision.

๐Ÿ“ˆ Highlights of the results:

  • On SQuAD (QA, knowledge integration), SEAL outperforms even synthetic data generated by GPT-4.1 after just two training iterations – and achieves this with a smaller base model.

  • In few-shot learning scenarios, SEAL achieves a 72.5% success rate, compared to only 20% for classic approaches, illustrating the immense potential of self-directed model adaptation.

  • Notably, SEAL operates independently of specific data formats – it can learn a variety of structures for self-training and weight adjustment.

๐Ÿง  Why is this revolutionary?
We are approaching a future in which language models will be able to autonomously ingest new information and embed it internally – without human guidance, relying entirely on their own data generation and selection. This is not only a significant step toward autonomous AI, but also provides an answer to the impending “data wall” in large-scale model training.

๐Ÿ’ก My conclusion:
SEAL is more than an efficiency hack – it marks the beginning of the era of truly self-optimizing AI systems. This paradigm shift will have profound implications for research, industry, and ultimately the entire digital infrastructure.

๐Ÿ‘‰ Link to the preprint (open access)
More information & code: https://jyopari.github.io/posts/seal

www.Trauth-Research.com

#AI #DeepLearning #ReinforcementLearning #MetaLearning #LLM #AutonomousAI



Keine Kommentare:

Kommentar verรถffentlichen