🐦 Twitter Post Details

Viewing enriched Twitter post

@dair_ai

First comprehensive framework for how AI agents actually improve through adaptation. While there is a lot of hype about building bigger models, the research reveals a different lever: systematic adaptation of agents and their tools. Researchers from many universities surveyed the rapidly expanding landscape of agentic AI adaptation. What they found: a fragmented field with no unified understanding of how agents learn to use tools, when to adapt the agent versus the tool, and which strategies work for which scenarios. These are all important for building production-ready AI agents. Adaptation in agentic AI follows four distinct paradigms that most practitioners conflate or ignore entirely. The framework organizes all adaptation strategies into two dimensions. > Agent Adaptation (A1, A2): modifying the agent's parameters, representations, or policies. > Tool Adaptation (T1, T2): optimizing external components like retrievers, planners, and memory modules while keeping the agent frozen. Let's discuss each in more detail: A1: Tool Execution Signaled Agent Adaptation. The agent learns from verifiable outcomes produced by tools it invokes. This involves code sandbox results, retrieval relevance scores, and API call outcomes. Methods like Toolformer, ToolLLM, and DeepRetrieval also fall here. The signal comes from whether the tool execution succeeded, not whether the final answer was correct. A2: Agent Output Signaled Agent Adaptation. The agent optimizes based on evaluations of its own final outputs. This includes both tool-free reasoning (DeepSeek-R1, Kimi-1.5) and tool-augmented adaptation (ReTool, Search-R1). The signal comes from answer correctness or preference scores, not intermediate tool calls. T1: Agent-Agnostic Tool Adaptation. This involves tools trained independently of any specific agent, including HuggingGPT, ViperGPT, and classic ML tools that serve as plug-and-play modules. These tools generalize well across different agents but may not be optimized for any particular one. T2: Agent-Supervised Tool Adaptation. Tools adapted using signals from a frozen agent's outputs. Includes reward-driven retriever tuning, adaptive search subagents, and memory-update modules like Reflexion and Memento. The agent stays fixed while tools learn to better support its reasoning. The trade-offs between paradigms are explicit. Cost and flexibility: A1/A2 require substantial compute for training billion-parameter models but offer maximal flexibility. T1/T2 optimize external components at a lower cost but may hit ceilings set by the frozen agent's capabilities. Generalization patterns differ significantly. T1 tools trained on broad distributions generalize well across agents and tasks. A1 methods risk overfitting to specific environments unless carefully regularized. T2 approaches enable independent tool upgrades without agent retraining, facilitating continuous improvement. The researchers identify when each paradigm fits. A1 suits scenarios with verifiable tool outputs like code execution or database queries. A2 works when only the final answer quality matters. T1 applies when tools must serve multiple agents. T2 excels when the agent is fixed, but tool performance is the bottleneck. State-of-the-art systems increasingly combine paradigms. A deep research system might use T1-style pretrained retrievers, T2-style adaptive search agents trained via frozen LLM feedback, and A1-style reasoning agents fine-tuned with execution feedback in a cascaded architecture. Four open challenges remain unsolved: - Co-adaptation: jointly optimizing agents and tools remains underexplored. - Continual adaptation: enabling lifelong learning without catastrophic forgetting. - Safe adaptation: preventing harmful behaviors during optimization. - Efficient adaptation: reducing computational costs while maintaining performance. The choice of adaptation paradigm fundamentally shapes what an agentic system can learn, how fast it improves, and whether improvements transfer across tasks. Teams building production agents need a principled framework for these decisions, not ad-hoc choices. Report: https://t.co/o2KPQLLQsZ Learn to build effective AI agents in our academy: https://t.co/g1Ijo0S5AA

Media 1
Media 2
Media 3

📊 Media Metadata

{
  "media": [
    {
      "type": "photo",
      "url": "https://crmoxkoizveukayfjuyo.supabase.co/storage/v1/object/public/media/posts/1999878493090599296/media_0.jpg?",
      "filename": "media_0.jpg"
    },
    {
      "type": "photo",
      "url": "https://crmoxkoizveukayfjuyo.supabase.co/storage/v1/object/public/media/posts/1999878493090599296/media_1.jpg?",
      "filename": "media_1.jpg"
    },
    {
      "type": "photo",
      "url": "https://crmoxkoizveukayfjuyo.supabase.co/storage/v1/object/public/media/posts/1999878493090599296/media_2.png?",
      "filename": "media_2.png"
    }
  ],
  "processed_at": "2025-12-13T16:36:36.959725",
  "pipeline_version": "2.0"
}

🔧 Raw API Response

{
  "type": "tweet",
  "id": "1999878493090599296",
  "url": "https://x.com/dair_ai/status/1999878493090599296",
  "twitterUrl": "https://twitter.com/dair_ai/status/1999878493090599296",
  "text": "First comprehensive framework for how AI agents actually improve through adaptation.\n\nWhile there is a lot of hype about building bigger models, the research reveals a different lever: systematic adaptation of agents and their tools.\n\nResearchers from many universities surveyed the rapidly expanding landscape of agentic AI adaptation.\n\nWhat they found: a fragmented field with no unified understanding of how agents learn to use tools, when to adapt the agent versus the tool, and which strategies work for which scenarios.\n\nThese are all important for building production-ready AI agents.\n\nAdaptation in agentic AI follows four distinct paradigms that most practitioners conflate or ignore entirely.\n\nThe framework organizes all adaptation strategies into two dimensions.\n\n> Agent Adaptation (A1, A2): modifying the agent's parameters, representations, or policies.\n> Tool Adaptation (T1, T2): optimizing external components like retrievers, planners, and memory modules while keeping the agent frozen.\n\nLet's discuss each in more detail:\n\nA1: Tool Execution Signaled Agent Adaptation.\nThe agent learns from verifiable outcomes produced by tools it invokes.  This involves code sandbox results, retrieval relevance scores, and API call outcomes. Methods like Toolformer, ToolLLM, and DeepRetrieval also fall here. The signal comes from whether the tool execution succeeded, not whether the final answer was correct.\n\nA2: Agent Output Signaled Agent Adaptation.\nThe agent optimizes based on evaluations of its own final outputs. This includes both tool-free reasoning (DeepSeek-R1, Kimi-1.5) and tool-augmented adaptation (ReTool, Search-R1). The signal comes from answer correctness or preference scores, not intermediate tool calls.\n\nT1: Agent-Agnostic Tool Adaptation.\nThis involves tools trained independently of any specific agent, including HuggingGPT, ViperGPT, and classic ML tools that serve as plug-and-play modules. These tools generalize well across different agents but may not be optimized for any particular one.\n\nT2: Agent-Supervised Tool Adaptation. Tools adapted using signals from a frozen agent's outputs. Includes reward-driven retriever tuning, adaptive search subagents, and memory-update modules like Reflexion and Memento. The agent stays fixed while tools learn to better support its reasoning.\n\nThe trade-offs between paradigms are explicit. Cost and flexibility: A1/A2 require substantial compute for training billion-parameter models but offer maximal flexibility. T1/T2 optimize external components at a lower cost but may hit ceilings set by the frozen agent's capabilities.\n\nGeneralization patterns differ significantly. T1 tools trained on broad distributions generalize well across agents and tasks. A1 methods risk overfitting to specific environments unless carefully regularized. T2 approaches enable independent tool upgrades without agent retraining, facilitating continuous improvement.\n\nThe researchers identify when each paradigm fits. A1 suits scenarios with verifiable tool outputs like code execution or database queries. A2 works when only the final answer quality matters. T1 applies when tools must serve multiple agents. T2 excels when the agent is fixed, but tool performance is the bottleneck.\n\nState-of-the-art systems increasingly combine paradigms. A deep research system might use T1-style pretrained retrievers, T2-style adaptive search agents trained via frozen LLM feedback, and A1-style reasoning agents fine-tuned with execution feedback in a cascaded architecture.\n\nFour open challenges remain unsolved:\n- Co-adaptation: jointly optimizing agents and tools remains underexplored.\n- Continual adaptation: enabling lifelong learning without catastrophic forgetting.\n- Safe adaptation: preventing harmful behaviors during optimization.\n- Efficient adaptation: reducing computational costs while maintaining performance.\n\nThe choice of adaptation paradigm fundamentally shapes what an agentic system can learn, how fast it improves, and whether improvements transfer across tasks. Teams building production agents need a principled framework for these decisions, not ad-hoc choices.\n\nReport: https://t.co/o2KPQLLQsZ\n\nLearn to build effective AI agents in our academy: https://t.co/g1Ijo0S5AA",
  "source": "Twitter for iPhone",
  "retweetCount": 3,
  "replyCount": 0,
  "likeCount": 9,
  "quoteCount": 0,
  "viewCount": 347,
  "createdAt": "Sat Dec 13 16:26:03 +0000 2025",
  "lang": "en",
  "bookmarkCount": 8,
  "isReply": false,
  "inReplyToId": null,
  "conversationId": "1999878493090599296",
  "displayTextRange": [
    0,
    279
  ],
  "inReplyToUserId": null,
  "inReplyToUsername": null,
  "author": {
    "type": "user",
    "userName": "dair_ai",
    "url": "https://x.com/dair_ai",
    "twitterUrl": "https://twitter.com/dair_ai",
    "id": "889050642903293953",
    "name": "DAIR.AI",
    "isVerified": false,
    "isBlueVerified": true,
    "verifiedType": null,
    "profilePicture": "https://pbs.twimg.com/profile_images/1643277398522187778/31dedbLo_normal.jpg",
    "coverPicture": "https://pbs.twimg.com/profile_banners/889050642903293953/1742055232",
    "description": "",
    "location": "",
    "followers": 83147,
    "following": 1,
    "status": "",
    "canDm": true,
    "canMediaTag": true,
    "createdAt": "Sun Jul 23 09:12:45 +0000 2017",
    "entities": {
      "description": {
        "urls": []
      },
      "url": {}
    },
    "fastFollowersCount": 0,
    "favouritesCount": 3881,
    "hasCustomTimelines": true,
    "isTranslator": false,
    "mediaCount": 87,
    "statusesCount": 2645,
    "withheldInCountries": [],
    "affiliatesHighlightedLabel": {},
    "possiblySensitive": false,
    "pinnedTweetIds": [
      "1999488933412110456"
    ],
    "profile_bio": {
      "description": "Democratizing AI research, education, and technologies.",
      "entities": {
        "description": {},
        "url": {
          "urls": [
            {
              "display_url": "dair.ai",
              "expanded_url": "https://www.dair.ai/",
              "indices": [
                0,
                23
              ],
              "url": "https://t.co/lkqPZtMmfU"
            }
          ]
        }
      }
    },
    "isAutomated": false,
    "automatedBy": null
  },
  "extendedEntities": {
    "media": [
      {
        "display_url": "pic.twitter.com/fTjp4yR8Ny",
        "expanded_url": "https://twitter.com/dair_ai/status/1999878493090599296/photo/1",
        "ext_media_availability": {
          "status": "Available"
        },
        "features": {
          "large": {},
          "orig": {}
        },
        "id_str": "1999878489319817216",
        "indices": [
          280,
          303
        ],
        "media_key": "3_1999878489319817216",
        "media_results": {
          "id": "QXBpTWVkaWFSZXN1bHRzOgwAAQoAARvA/uPl2kAACgACG8D+5Mab0YAAAA==",
          "result": {
            "__typename": "ApiMedia",
            "id": "QXBpTWVkaWE6DAABCgABG8D+4+XaQAAKAAIbwP7kxpvRgAAA",
            "media_key": "3_1999878489319817216"
          }
        },
        "media_url_https": "https://pbs.twimg.com/media/G8D-4-XaQAAnP_t.jpg",
        "original_info": {
          "focus_rects": [
            {
              "h": 739,
              "w": 1320,
              "x": 0,
              "y": 60
            },
            {
              "h": 1320,
              "w": 1320,
              "x": 0,
              "y": 0
            },
            {
              "h": 1505,
              "w": 1320,
              "x": 0,
              "y": 0
            },
            {
              "h": 1562,
              "w": 781,
              "x": 273,
              "y": 0
            },
            {
              "h": 1562,
              "w": 1320,
              "x": 0,
              "y": 0
            }
          ],
          "height": 1562,
          "width": 1320
        },
        "sizes": {
          "large": {
            "h": 1562,
            "w": 1320
          }
        },
        "type": "photo",
        "url": "https://t.co/fTjp4yR8Ny"
      }
    ]
  },
  "card": null,
  "place": {},
  "entities": {
    "urls": [
      {
        "display_url": "github.com/pat-jj/Awesome…",
        "expanded_url": "https://github.com/pat-jj/Awesome-Adaptation-of-Agentic-AI/blob/main/paper.pdf",
        "indices": [
          4151,
          4174
        ],
        "url": "https://t.co/o2KPQLLQsZ"
      },
      {
        "display_url": "dair-ai.thinkific.com/courses/buildi…",
        "expanded_url": "https://dair-ai.thinkific.com/courses/building-effective-ai-agents-2",
        "indices": [
          4227,
          4250
        ],
        "url": "https://t.co/g1Ijo0S5AA"
      }
    ]
  },
  "quoted_tweet": null,
  "retweeted_tweet": null,
  "isLimitedReply": false,
  "article": null
}