DiscoverBest AI papers explainedMemento: Fine-tuning LLM Agents without Fine-tuning LLMs
Memento: Fine-tuning LLM Agents without Fine-tuning LLMs

Memento: Fine-tuning LLM Agents without Fine-tuning LLMs

Update: 2025-09-01
Share

Description

The research introduces Memento, a novel approach for adaptive Large Language Model (LLM) agents that enables continuous learning without requiring fine-tuning of the base LLM parameters. This method leverages a memory-based online reinforcement learning framework, formally defined as a Memory-augmented Markov Decision Process (M-MDP), which stores past experiences in an episodic memory and continually updates a neural case-selection policy. Memento utilizes a planner-executor architecture and a comprehensive suite of tools, demonstrating state-of-the-art performance on various benchmarks, including GAIA, DeepResearcher, and SimpleQA. The ablation studies confirm that both parametric and non-parametric case-based reasoning (CBR) are crucial for significant performance gains and effective generalization to out-of-distribution tasks.

Comments 
In Channel
loading
00:00
00:00
x

0.5x

0.8x

1.0x

1.25x

1.5x

2.0x

3.0x

Sleep Timer

Off

End of Episode

5 Minutes

10 Minutes

15 Minutes

30 Minutes

45 Minutes

60 Minutes

120 Minutes

Memento: Fine-tuning LLM Agents without Fine-tuning LLMs

Memento: Fine-tuning LLM Agents without Fine-tuning LLMs

Enoch H. Kang