SP
BravenNow
Learning Physical Principles from Interaction: Self-Evolving Planning via Test-Time Memory
| USA | technology | ✓ Verified - arxiv.org

Learning Physical Principles from Interaction: Self-Evolving Planning via Test-Time Memory

#PhysMem #robot planning #vision-language models #physical principles #robot manipulation #embodied AI #test-time learning

📌 Key Takeaways

  • Researchers developed PhysMem, a memory framework for robot planners to learn physical principles from interaction
  • The system uses verification before application to test hypotheses against new observations
  • PhysMem achieved 76% success rate on brick insertion task compared to 23% for direct experience retrieval
  • Real-world experiments showed consistent improvements over 30-minute deployment sessions

📖 Full Retelling

Researchers Haoyang Li, Yang You, Hao Su, and Leonidas Guibas developed PhysMem, a memory framework that enables robot planners to learn physical principles from interaction during test time, without updating model parameters, in a paper submitted to arXiv on February 23, 2026. The research addresses a critical limitation in vision-language model (VLM) planners, which can reason about physical properties like friction and stability in general terms but often fail when predicting how specific objects will interact with particular environments without direct experience. PhysMem represents a significant advancement in robotic manipulation capabilities by creating a system that records experiences, generates candidate hypotheses, and verifies them through targeted interaction before promoting validated knowledge to guide future decisions. The central innovation of PhysMem lies in its verification-before-application approach, where the system tests hypotheses against new observations rather than applying retrieved experience directly. This design reduces rigid reliance on prior experience when physical conditions change, allowing robots to adapt to novel situations. The researchers evaluated their framework on three real-world manipulation tasks and simulation benchmarks across four VLM backbones. Results demonstrated that on a controlled brick insertion task, principled abstraction achieved a 76% success rate compared to only 23% for direct experience retrieval. Real-world experiments further showed consistent performance improvements over 30-minute deployment sessions, demonstrating the practical value of this approach. This research contributes to the growing field of embodied AI by creating a more flexible and adaptable robotic planning system. By enabling robots to learn from their interactions during deployment rather than requiring extensive pre-training or model updates, PhysMem opens new possibilities for autonomous systems operating in dynamic, unstructured environments. The framework's ability to abstract physical principles from specific experiences represents a step toward more generalizable robotic intelligence that can transfer knowledge across different scenarios while remaining responsive to immediate environmental conditions.

🏷️ Themes

Robotics, Artificial Intelligence, Machine Learning

Entity Intersection Graph

No entity connections available yet for this article.

Original Source
--> Computer Science > Robotics arXiv:2602.20323 [Submitted on 23 Feb 2026] Title: Learning Physical Principles from Interaction: Self-Evolving Planning via Test-Time Memory Authors: Haoyang Li , Yang You , Hao Su , Leonidas Guibas View a PDF of the paper titled Learning Physical Principles from Interaction: Self-Evolving Planning via Test-Time Memory, by Haoyang Li and 3 other authors View PDF HTML Abstract: Reliable object manipulation requires understanding physical properties that vary across objects and environments. Vision-language model planners can reason about friction and stability in general terms; however, they often cannot predict how a specific ball will roll on a particular surface or which stone will provide a stable foundation without direct experience. We present PhysMem, a memory framework that enables VLM robot planners to learn physical principles from interaction at test time, without updating model parameters. The system records experiences, generates candidate hypotheses, and verifies them through targeted interaction before promoting validated knowledge to guide future decisions. A central design choice is verification before application: the system tests hypotheses against new observations rather than applying retrieved experience directly, reducing rigid reliance on prior experience when physical conditions change. We evaluate PhysMem on three real-world manipulation tasks and simulation benchmarks across four VLM backbones. On a controlled brick insertion task, principled abstraction achieves 76% success compared to 23% for direct experience retrieval, and real-world experiments show consistent improvement over 30-minute deployment sessions. Subjects: Robotics (cs.RO) ; Artificial Intelligence (cs.AI) Cite as: arXiv:2602.20323 [cs.RO] (or arXiv:2602.20323v1 [cs.RO] for this version) https://doi.org/10.48550/arXiv.2602.20323 Focus to learn more arXiv-issued DOI via DataCite (pending registration) Submission history From: Haoyang Li [ view ...
Read full article at source

Source

arxiv.org

More from USA

News from Other Countries

🇬🇧 United Kingdom

🇺🇦 Ukraine