AI Agents

Agent-Omit: Training Efficient LLM Agents for Adaptive Thought and Observation Omission via Agentic Reinforcement Learning

YYansong NingJJun FangNNaiqiang TanHHao Liu
Published
February 4, 2026
Authors
4
Word Count
10,925
Code
Includes code

Efficient LLM agents through adaptive thought and observation omission.

Abstract

Managing agent thought and observation during multi-turn agent-environment interactions is an emerging strategy to improve agent efficiency. However, existing studies treat the entire interaction trajectories equally, overlooking the thought necessity and observation utility varies across turns. To this end, we first conduct quantitative investigations into how thought and observation affect agent effectiveness and efficiency. Based on our findings, we propose Agent-Omit, a unified training framework that empowers LLM agents to adaptively omit redundant thoughts and observations. Specifically, we first synthesize a small amount of cold-start data, including both single-turn and multi-turn omission scenarios, to fine-tune the agent for omission behaviors. Furthermore, we introduce an omit-aware agentic reinforcement learning approach, incorporating a dual sampling mechanism and a tailored omission reward to incentivize the agent's adaptive omission capability. Theoretically, we prove that the deviation of our omission policy is upper-bounded by KL-divergence. Experimental results on five agent benchmarks show that our constructed Agent-Omit-8B could obtain performance comparable to seven frontier LLM agent, and achieve the best effectiveness-efficiency trade-off than seven efficient LLM agents methods. Our code and data are available at https://github.com/usail-hkust/Agent-Omit.

Key Takeaways

  • 1

    Agent-Omit improves LLM agent efficiency via adaptive omission.

  • 2

    Framework includes Agent Omission Behavior Synthesis and Omit-Aware RL.

  • 3

    Efficient agents lead to faster responses and lower costs.

Limitations

  • Requires synthetic dataset for fine-tuning.

  • May not generalize well to all tasks.

Keywords

LLM agentsmulti-turn interactionsthought necessityobservation utilitycold-start datafine-tuneagentic reinforcement learningdual sampling mechanismomission rewardKL-divergence

More in AI Agents

View all
Agent-Omit: Training Efficient LLM Agents for Adaptive Thought and Observation Omission via Agentic Reinforcement Learning | Paperchime