🐦 Twitter Post Details

Viewing enriched Twitter post

@omarsar0

Banger paper for agent builders. Multi-agent systems often underdeliver. The problem isn't how the agents themselves are built. It's how they're organized. They are mostly built with fixed chains, trees, and graphs that can't adapt as tasks evolve. But what if the system could learn its own coordination patterns? This new research introduces Puppeteer, a framework that learns to orchestrate agents dynamically rather than relying on handcrafted topologies. Instead of pre-defining collaboration structures, an orchestrator selects which agent speaks next based on the evolving conversation state. The policy is trained with REINFORCE, optimizing directly for task success. Rather than searching over complex graph topologies, they serialize everything into sequential agent selections. This reframing sidesteps combinatorial complexity. What emerges is surprising: compact cyclic patterns develop naturally. Not sprawling graphs, but tight loops where 2-3 agents handle most of the work. The remarkable part is that the system discovers efficiency on its own. Results: - On GSM-Hard math problems: 70% accuracy (up from 13.5% for the base model alone). - On MMLU-Pro: 83% (vs 76% baseline). - On SRDD software development: 76.4% (vs 60.6% baseline). These gains come with reduced token consumption. The paper shows that token costs consistently decrease throughout training while performance improves. They also prove the agent selection process satisfies Markov properties, meaning the current state alone determines the optimal next agent. No need to track full history. Why it matters for AI devs: learned simplicity beats engineered complexity. A trained router with a handful of specialized agents can outperform elaborate handcrafted workflows while cutting computational overhead.

Media 1

📊 Media Metadata

{
  "media": [
    {
      "url": "https://crmoxkoizveukayfjuyo.supabase.co/storage/v1/object/public/media/posts/1995553529436783096/media_0.jpg?",
      "media_url": "https://crmoxkoizveukayfjuyo.supabase.co/storage/v1/object/public/media/posts/1995553529436783096/media_0.jpg?",
      "type": "photo",
      "filename": "media_0.jpg"
    }
  ],
  "processed_at": "2025-12-04T20:37:57.448512",
  "pipeline_version": "2.0"
}

🔧 Raw API Response

{
  "type": "tweet",
  "id": "1995553529436783096",
  "url": "https://x.com/omarsar0/status/1995553529436783096",
  "twitterUrl": "https://twitter.com/omarsar0/status/1995553529436783096",
  "text": "Banger paper for agent builders.\n\nMulti-agent systems often underdeliver. The problem isn't how the agents themselves are built. It's how they're organized.\n\nThey are mostly built with fixed chains, trees, and graphs that can't adapt as tasks evolve.\n\nBut what if the system could learn its own coordination patterns?\n\nThis new research introduces Puppeteer, a framework that learns to orchestrate agents dynamically rather than relying on handcrafted topologies.\n\nInstead of pre-defining collaboration structures, an orchestrator selects which agent speaks next based on the evolving conversation state. The policy is trained with REINFORCE, optimizing directly for task success.\n\nRather than searching over complex graph topologies, they serialize everything into sequential agent selections. This reframing sidesteps combinatorial complexity.\n\nWhat emerges is surprising: compact cyclic patterns develop naturally. Not sprawling graphs, but tight loops where 2-3 agents handle most of the work.\n\nThe remarkable part is that the system discovers efficiency on its own.\n\nResults:\n- On GSM-Hard math problems: 70% accuracy (up from 13.5% for the base model alone).\n- On MMLU-Pro: 83% (vs 76% baseline).\n- On SRDD software development: 76.4% (vs 60.6% baseline).\n\nThese gains come with reduced token consumption. The paper shows that token costs consistently decrease throughout training while performance improves.\n\nThey also prove the agent selection process satisfies Markov properties, meaning the current state alone determines the optimal next agent. No need to track full history.\n\nWhy it matters for AI devs: learned simplicity beats engineered complexity. A trained router with a handful of specialized agents can outperform elaborate handcrafted workflows while cutting computational overhead.",
  "source": "Twitter for iPhone",
  "retweetCount": 136,
  "replyCount": 36,
  "likeCount": 835,
  "quoteCount": 5,
  "viewCount": 50386,
  "createdAt": "Mon Dec 01 18:00:11 +0000 2025",
  "lang": "en",
  "bookmarkCount": 940,
  "isReply": false,
  "inReplyToId": null,
  "conversationId": "1995553529436783096",
  "displayTextRange": [
    0,
    281
  ],
  "inReplyToUserId": null,
  "inReplyToUsername": null,
  "author": {
    "type": "user",
    "userName": "omarsar0",
    "url": "https://x.com/omarsar0",
    "twitterUrl": "https://twitter.com/omarsar0",
    "id": "3448284313",
    "name": "elvis",
    "isVerified": false,
    "isBlueVerified": true,
    "verifiedType": null,
    "profilePicture": "https://pbs.twimg.com/profile_images/939313677647282181/vZjFWtAn_normal.jpg",
    "coverPicture": "https://pbs.twimg.com/profile_banners/3448284313/1565974901",
    "description": "",
    "location": "DAIR.AI Academy",
    "followers": 277867,
    "following": 724,
    "status": "",
    "canDm": true,
    "canMediaTag": true,
    "createdAt": "Fri Sep 04 12:59:26 +0000 2015",
    "entities": {
      "description": {
        "urls": []
      },
      "url": {}
    },
    "fastFollowersCount": 0,
    "favouritesCount": 33719,
    "hasCustomTimelines": true,
    "isTranslator": true,
    "mediaCount": 4356,
    "statusesCount": 16656,
    "withheldInCountries": [],
    "affiliatesHighlightedLabel": {},
    "possiblySensitive": false,
    "pinnedTweetIds": [
      "1996595107924263287"
    ],
    "profile_bio": {
      "description": "Building agents @dair_ai • Ex Meta AI, Elastic, PhD • Sharing research & insights on AI Agents • New cohort: https://t.co/tn8LKG5d20",
      "entities": {
        "description": {
          "urls": [
            {
              "display_url": "dair-ai.thinkific.com/courses/claude…",
              "expanded_url": "https://dair-ai.thinkific.com/courses/claude-code",
              "indices": [
                109,
                132
              ],
              "url": "https://t.co/tn8LKG5d20"
            }
          ],
          "user_mentions": [
            {
              "id_str": "0",
              "indices": [
                16,
                24
              ],
              "name": "",
              "screen_name": "dair_ai"
            }
          ]
        },
        "url": {
          "urls": [
            {
              "display_url": "dair-ai.thinkific.com",
              "expanded_url": "https://dair-ai.thinkific.com/",
              "indices": [
                0,
                23
              ],
              "url": "https://t.co/JBU5beHQNs"
            }
          ]
        }
      }
    },
    "isAutomated": false,
    "automatedBy": null
  },
  "extendedEntities": {
    "media": [
      {
        "display_url": "pic.twitter.com/sH5KTEOQVI",
        "expanded_url": "https://twitter.com/omarsar0/status/1995553529436783096/photo/1",
        "ext_media_availability": {
          "status": "Available"
        },
        "features": {
          "large": {
            "faces": [
              {
                "h": 215,
                "w": 215,
                "x": 865,
                "y": 1057
              }
            ]
          },
          "orig": {
            "faces": [
              {
                "h": 215,
                "w": 215,
                "x": 865,
                "y": 1057
              }
            ]
          }
        },
        "id_str": "1995553526022635522",
        "indices": [
          282,
          305
        ],
        "media_key": "3_1995553526022635522",
        "media_results": {
          "id": "QXBpTWVkaWFSZXN1bHRzOgwAAQoAARuxoVvo24ACCgACG7GhXLRbQfgAAA==",
          "result": {
            "__typename": "ApiMedia",
            "id": "QXBpTWVkaWE6DAABCgABG7GhW+jbgAIKAAIbsaFctFtB+AAA",
            "media_key": "3_1995553526022635522"
          }
        },
        "media_url_https": "https://pbs.twimg.com/media/G7GhW-jbgAIkszD.jpg",
        "original_info": {
          "focus_rects": [
            {
              "h": 892,
              "w": 1592,
              "x": 0,
              "y": 0
            },
            {
              "h": 1592,
              "w": 1592,
              "x": 0,
              "y": 0
            },
            {
              "h": 1798,
              "w": 1577,
              "x": 15,
              "y": 0
            },
            {
              "h": 1798,
              "w": 899,
              "x": 674,
              "y": 0
            },
            {
              "h": 1798,
              "w": 1592,
              "x": 0,
              "y": 0
            }
          ],
          "height": 1798,
          "width": 1592
        },
        "sizes": {
          "large": {
            "h": 1798,
            "w": 1592
          }
        },
        "type": "photo",
        "url": "https://t.co/sH5KTEOQVI"
      }
    ]
  },
  "card": null,
  "place": {},
  "entities": {},
  "quoted_tweet": null,
  "retweeted_tweet": null,
  "isLimitedReply": false,
  "article": null
}