Точка Синхронізації

AI Archive of Human History

Free(): Learning to Forget in Malloc-Only Reasoning Models
| USA | technology

Free(): Learning to Forget in Malloc-Only Reasoning Models

#Free()LM #Reasoning Models #Large Language Models #Test-time Compute #arXiv #Malloc-only #Artificial Intelligence #Memory Management

📌 Key Takeaways

  • Researchers identified a 'malloc-only' flaw in current AI models that prevents them from deleting redundant reasoning steps.
  • Excessive thinking tokens in LLMs frequently lead to performance degradation rather than higher accuracy.
  • Free()LM introduces a mechanism to prune obsolete information during the inference process.
  • The model aims to improve the efficiency of test-time compute by focusing on selective memory and logical pruning.

📖 Full Retelling

Researchers recently published a technical paper on the arXiv preprint server introducing 'Free()LM,' a novel framework designed to solve the performance degradation caused by excessive 'thinking' tokens in large language models (LLMs). Released in late February 2024, the study addresses a critical paradox in reasoning models where scaling test-time compute—the amount of processing done during inference—can actually hinder accuracy rather than improve it. The authors argue that current architectures function as 'malloc-only' systems that allocate memory for every computational step but lack the capacity to discard redundant or incorrect logic, leading to cognitive clutter during complex problem-solving tasks. The core of the research identifies a fundamental architectural flaw in how modern AI models handle long-chain reasoning. Unlike traditional software that uses a 'free' command to release unused memory, standard LLMs continuously accumulate both valid and obsolete information within their context window. This accumulation forces the model to sift through an ever-increasing amount of noise, which often results in logical drift or the 'lost in the middle' phenomenon. By failing to prune irrelevant steps, the models eventually collapse under the weight of their own generated tokens, effectively hitting a ceiling of diminishing returns. To overcome these limitations, the proposed Free()LM architecture introduces a mechanism for models to 'learn to forget.' This approach allows the system to actively identify and prune redundant or erroneous intermediate steps during the reasoning process. By mimicking a more human-like cognitive process of filtering out distractions, Free()LM aims to maximize the efficiency of test-time compute. This development represents a significant shift in AI research, moving away from simply increasing the volume of computation toward a more sophisticated management of the model's internal state and memory allocation.

🏷️ Themes

Artificial Intelligence, Computer Science, Machine Learning

📚 Related People & Topics

Reasoning model

Language models designed for reasoning tasks

A reasoning model, also known as reasoning language models (RLMs) or large reasoning models (LRMs), is a type of large language model (LLM) that has been specifically trained to solve complex tasks requiring multiple steps of logical reasoning. These models demonstrate superior performance on logic,...

Wikipedia →

Large language model

Type of machine learning model

A large language model (LLM) is a language model trained with self-supervised machine learning on a vast amount of text, designed for natural language processing tasks, especially language generation. The largest and most capable LLMs are generative pre-trained transformers (GPTs) that provide the c...

Wikipedia →

Artificial intelligence

Artificial intelligence

Intelligence of machines

# Artificial Intelligence (AI) **Artificial Intelligence (AI)** is a specialized field of computer science dedicated to the development and study of computational systems capable of performing tasks typically associated with human intelligence. These tasks include learning, reasoning, problem-solvi...

Wikipedia →

🔗 Entity Intersection Graph

Connections for Reasoning model:

View full profile →

📄 Original Source Content
arXiv:2602.08030v1 Announce Type: new Abstract: Reasoning models enhance problem-solving by scaling test-time compute, yet they face a critical paradox: excessive thinking tokens often degrade performance rather than improve it. We attribute this to a fundamental architectural flaw: standard LLMs operate as "malloc-only" engines, continuously accumulating valid and redundant steps alike without a mechanism to prune obsolete information. To break this cycle, we propose Free()LM, a model that int

Original source

More from USA

News from Other Countries

🇵🇱 Poland

🇬🇧 United Kingdom

🇺🇦 Ukraine

🇮🇳 India