Rick-Brick
Four Axes Redefining LLMs as Autonomous AI Agents — A Systematization of Agentic Reasoning Frameworks

Title

Four Axes Redefining LLMs as Autonomous AI Agents — A Systematization of Agentic Reasoning Frameworks

Summary

Systematizing agentic reasoning along four axes: planning, tool use, memory, and self-improvement, based on the ‘Large Language Model Agent’ survey (arXiv

.21460). Explores the frontier of LLM agent research with Silo-Bench, MC-Search, etc.

Tags

[“AI Agents”,“Reasoning”,“LLM”,“Multi-agent”,“MCP”]

Body

In March 2025, “Large Language Model Agent: A Survey on Methodology, Applications and Challenges” (arXiv

.21460) appeared on arXiv, presenting a comprehensive survey that organizes the methodologies, applications, and challenges of LLM agents across 329 papers. This paper, involving 26 researchers led by Junyu Luo and positioning the field as a crucial research area for AGI (Artificial General Intelligence), provides new guidance for the systematization of agentic reasoning.

This article, using this survey as a basis, will organize agentic reasoning along the four axes of planning, tool use, memory, and self-improvement, while also explaining the frontiers of LLM agent research, such as Silo-Bench and MC-Search.

Why is “Agentic Reasoning” Gaining Attention?

From Scaling Laws to Agentification

From the late 2010s to the early 2020s, the performance improvements of LLMs were driven by scaling laws—increasing model size, data volume, and computation. However, the cost of scaling increases exponentially, making it difficult to achieve equivalent performance gains with the same methods.

As of 2026, the research focus has clearly shifted from “how large should the model be” to “how should the model be used.” A survey paper organizing the landscape of agentic reinforcement learning (arXiv

.02547) describes this transition as “from LLMs as passive sequence generators to autonomous decision-making agents embedded in complex dynamic environments.”

A Paradigm Shift from “Generation” to “Action”

Traditional LLMs were designed as “closed generation systems” that receive a prompt and return text. Agentic LLMs fundamentally overturn this design.

  • Given a goal, they autonomously create action plans.
  • They invoke external tools (search engines, code interpreters, APIs, etc.).
  • They incorporate execution results as feedback and revise plans.
  • They maintain long-term memory and adapt across multiple sessions.

This is a shift from “answering questions” to “completing tasks,” representing a fundamental paradigm transformation for AI systems.

Four Core Axes of Agentic Reasoning

The four axes defined by multiple surveys, including “LLM-based Agentic Reasoning Frameworks: A Survey from Methods to Scenarios” (arXiv

.17692) and arXiv
.02547, have become a common framework in the current research community.

┌────────────────────────────────────────┐
│         LLM Agent                      │
│                                        │
│  ┌──────────┐    ┌──────────┐          │
│  │ Planning │    │ Tool Use │          │
│  └────┬─────┘    └────┬─────┘          │
│       │               │                │
│  ┌────▼─────┐    ┌────▼─────┐          │
│  │ Memory   │    │ Self-    │          │
│  │          │    │ Improve  │          │
│  └──────────┘    └──────────┘          │
└────────────────────────────────────────┘

Axis 1: Planning

Planning is the starting point of agentic reasoning. It refers to the ability to decompose a given goal and organize it as a sequence of executable subtasks.

There has been a gradual evolution in task decomposition methods. While Chain-of-Thought (CoT) became popular as a simple planning representation, more sophisticated methods have emerged in recent years.

  • Tree-of-Thoughts (ToT): Represents plans in a tree structure, exploring and evaluating multiple candidate paths.
  • Graph-of-Thoughts (GoT): Introduces a graph structure, enabling plan reuse and branching.
  • Hierarchical Architectures: Higher-level agents formulate strategic plans and delegate specific subtasks to lower-level agents.

The accuracy of planning is the most critical factor determining the overall performance of the agent. Without appropriate subtask decomposition, subsequent tool use and memory utilization cannot be effective.

Axis 2: Tool Use

Tool use is the functionality that allows agents to interact with external systems. LLMs autonomously select and invoke diverse tools, such as issuing queries to search engines, executing code, referencing databases, and calling external APIs.

The advent of the Model Context Protocol (MCP) is of historical significance in standardizing this tool use. Proposed by Anthropic in November 2024, this protocol is an open standard that enables “plug-and-play” connectivity between LLMs and toolkits, often referred to as the “USB-C for AI applications.”

The adoption of MCP has spread rapidly, with monthly SDK downloads exceeding 97 million by the end of 2025, and OpenAI, Google, and Microsoft have also announced their adoption. In December 2025, Anthropic donated MCP to the Agentic AI Foundation (AAIF) under the Linux Foundation, establishing it as a true industry standard.

Methodologies for tool selection can be categorized into the following three types:

Selection MethodContentApplication Scenarios
Autonomous SelectionLLM selects tools based on situational judgmentGeneral-purpose agents
Rule-BasedSelection based on predefined rulesTasks with clear constraints
Learning-BasedOptimizes tool selection through reinforcement learningRepetitive tasks

Axis 3: Memory

Memory is the foundation supporting agent autonomy. Since the context window of LLMs is finite, external memory mechanisms are necessary for long-term information retention.

“Memory for Autonomous LLM Agents: Mechanisms, Evaluation, and Emerging Frontiers” (arXiv

.07670) comprehensively reviews memory research from 2022 to early 2026 and defines a four-layer memory architecture.

Memory TypeContentImplementation Examples
Working MemoryContext of the current task (context window)LLM input buffer
Episodic MemoryPast events/experiences (with timestamps)Vector databases
Semantic MemoryAbstract concepts, knowledge, factsKnowledge graphs, RAG
Procedural MemoryExecutable skills, plan templatesFine-tuning, code

A significant finding revealed by research is the reality that “examples of ideal integration of the four-layer structure are rare.” Most current systems effectively implement two layers, with transitions between layers handled heuristically. Research such as A-Mem (arXiv

.12110) explores advanced memory architectures combining vector search and graph structures, but complete four-layer integration remains an open research challenge.

Axis 4: Self-Improvement

The ability of agents to learn from experience and improve themselves is the fourth axis. The three paradigms of self-improvement organized by arXiv

.17692 are as follows:

Reflection: A mechanism by which agents review past actions and outcomes to extract lessons. Reflexion and Self-Refine are representative frameworks implementing this idea. Reflexion stores action history as episodic memory and inserts a reflection process before the next trial. Self-Refine generates feedback on its generated output and uses that feedback to iteratively improve the output.

Iterative Optimization: An approach that iteratively refines prompts and tool selection strategies rather than updating the weights of the entire model.

Interactive Learning: Dynamically adjusts goals through continuous interaction with the environment. It has high affinity with reinforcement learning, and integration with agentic RL is progressing.

Agent Basic Loop from an Implementation Perspective

From an implementation standpoint, the agent’s basic loop, representing the interaction of the four axes, can be described as follows:

# Agentic Reasoning Basic Loop (Pseudocode)
def agent_loop(goal, tools, memory):
    while not goal_achieved(goal):
        # 1. Observe (Perceive) from environment
        observation = perceive(environment)

        # 2. Retrieve relevant information from memory (Memory)
        relevant_context = memory.retrieve(observation, top_k=5)

        # 3. Generate plan (Planning)
        plan = llm.plan(goal, observation, relevant_context)

        # 4. Select and execute tool (Tool Use)
        action = plan.next_action()
        result = tools.execute(action)

        # 5. Update memory (Memory -> Material for Self-Improvement)
        memory.store(episode={
            "action": action,
            "result": result,
            "timestamp": now()
        })

        # 6. Reflect and self-correct (Self-Improvement)
        if result.is_failure():
            reflection = llm.reflect(action, result)
            plan.revise(reflection)

In this loop, the four axes do not function as independent modules but form a dynamic system that provides feedback to each other.

Multi-Agent: The Fifth Dimension

Multi-agent systems address challenges beyond the capabilities of a single agent. The survey in arXiv

.21460 positions inter-agent coordination mechanisms as an important pillar of architecture.

MultiAgentBench: Development of Evaluation Frameworks

MultiAgentBench (arXiv

.01935, accepted to ACL 2025) contributes to accelerating research in multi-agent systems. This framework for quantitatively evaluating the collaboration and competition of LLM agents has the following features:

  • Measures not only task completion rate but also collaboration quality using milestone-based KPIs.
  • Evaluates four types of collaboration topologies: Star, Chain, Tree, and Graph.
  • Validates innovative strategies such as group discussion and cognitive planning.
  • Key Finding: Graph structures showed the highest performance in research scenarios, and cognitive planning improved milestone achievement rate by 3%.

Designing Collaboration Topologies

The organizational structure of multi-agent systems can be classified into three types.

Centralized          Distributed             Hierarchical
     A                A  B                 Leader
   / | \              |\/|                /   |   \
  B  C  D             C  D              Sub1 Sub2 Sub3
                                        / \       / \
                                       E   F     G   H

Results from MultiAgentBench indicate that the optimal topology varies depending on the nature of the task. For complex research tasks, Graph structures are superior, while for simple execution tasks, Star or Chain structures are efficient.

MC-Search (arXiv

.00873, submitted to ICLR 2026), introduced in March 2026, is a framework for evaluating and enhancing agent capabilities in complex information search tasks that span text and images.

Dataset Scale and Characteristics:

  • Includes 3,333 high-quality examples.
  • Step-wise reasoning chains with an average of 3.7 hops.
  • Quality assurance through HAVE (Hop-wise Attribution and Verification of Evidence).

Innovative Evaluation Metrics (three process-level evaluations beyond traditional answer accuracy):

  1. LLM-as-a-Judge: Evaluation of open-ended reasoning quality.
  2. Structure-Aware per Step Hit Rate: Measurement of step-wise search accuracy.
  3. Rollout Deviation: Quantification of execution drift (deviation from the plan).

Search-Align: A framework for improving the planning and search accuracy of open-source MLLMs through process-monitoring fine-tuning using verified reasoning chains.

The eight systematic error patterns revealed by MC-Search (e.g., over-searching/under-searching, modality mismatch in planning) specifically illustrate typical failures that implementers should avoid.

Challenges and Limitations of Agentic Reasoning

Amplification of Reliability Issues and Hallucinations

When agents act autonomously over multiple steps, errors in intermediate steps can propagate to subsequent steps, increasing the risk of amplifying the final mistake.

CARE-RFT (arXiv

.00085) directly addresses this trade-off. While reinforcement fine-tuning (RFT) to enhance reasoning performance tends to amplify hallucinations, this problem can be mitigated through a reliability-based penalty design using skewed inverse KL divergence, as demonstrated by the research.

Cost and Latency

Each time an agent repeats the planning, execution, and reflection loop, LLM inference costs are incurred. Complex tasks may require dozens of LLM calls, posing practical limitations.

Security and Prompt Injection

Agents that reference external data are vulnerable to “prompt injection” attacks, where malicious content can induce unintended actions. Sandbox design and the principle of least privilege are crucial.

Difficulty of Evaluation

As MC-Search demonstrates, evaluating agent performance is significantly more challenging than single-turn question answering. Designing appropriate process-level metrics is itself an important research challenge.

Application Domains: How Agents Are Changing Fields

Software Engineering

This is one of the most active application areas for agentic reasoning. Agents that autonomously perform code generation, debugging, and refactoring over multiple steps are emerging, and rapid performance improvements have been recorded on benchmarks like SWE-Bench. The role of engineers is shifting from “people who write code” to “people who provide goals to agents and verify the outputs.”

Scientific Discovery

Agents that autonomously repeat experimental design, literature review, hypothesis generation, and result analysis are changing the pace of scientific research. Applications in drug discovery and materials science are increasing, and accelerated discovery through collaboration with human researchers is expected.

Interaction Economy Among AI Agents

Meta’s acquisition of the AI agent-exclusive platform “Moltbook” signifies the dawn of an “agent economy” where agents communicate and collaborate. Infrastructure development to realize agent identity verification and linkage to human owners is emerging as the next challenge.

Conclusion: Design Guidelines Indicated by the Four Axes

The current state of LLM agent research organized by arXiv

.21460 shows that the four axes of planning, tool use, memory, and self-improvement form a system that complements each other, rather than being independent modules.

Planning  ──────→ Tool Use
  ↑               │
  │               ↓
Self-Improvement ←──── Memory

Insights gained from tool use (external operations based on planning) are stored as memory, accumulated memory becomes material for self-improvement, and improved capabilities refine the next plan. This cycle is the core of agentic AI.

While benchmarks like MultiAgentBench and MC-Search are advancing, the four-layer integration of memory, countermeasures against prompt injection, and process-level evaluation methods remain open challenges.

In the post-scaling era, LLM agents are not just technical advancements but are redefining the very nature of human-AI collaboration. A systematic understanding of the four axes has become indispensable foundational knowledge for everyone designing and utilizing agents.

References

TitleSourceDateURL
Large Language Model Agent: A Survey on Methodology, Applications and ChallengesarXiv2025/03/27https://arxiv.org/abs/2503.21460
LLM-based Agentic Reasoning Frameworks: A Survey from Methods to ScenariosarXiv2025/08https://arxiv.org/html/2508.17692v1
The Landscape of Agentic Reinforcement Learning for LLMs: A SurveyarXiv2025/09https://arxiv.org/abs/2509.02547
Memory for Autonomous LLM Agents: Mechanisms, Evaluation, and Emerging FrontiersarXiv2026/03https://arxiv.org/html/2603.07670
MC-Search: Evaluating and Enhancing Multimodal Agentic Search with Structured Long Reasoning ChainsarXiv2026/03/01https://arxiv.org/abs/2603.00873
MultiAgentBench: Evaluating the Collaboration and Competition of LLM agentsarXiv / ACL 20252025/03/03https://arxiv.org/abs/2503.01935
A-Mem: Agentic Memory for LLM AgentsarXiv2025/02https://arxiv.org/pdf/2502.12110
Model Context Protocol — WikipediaWikipedia2025https://en.wikipedia.org/wiki/Model_Context_Protocol
A Year of MCP: From Internal Experiment to Industry StandardPento Blog2025https://www.pento.ai/blog/a-year-of-mcp-2025-review
Agentic LLMs in 2025: How AI Is Becoming Self-Directed, Tool-Using & AutonomousData Science Dojo2025https://datasciencedojo.com/blog/agentic-llm-in-2025/

This article was automatically generated by LLM. It may contain errors.