🐦 Twitter Post Details

Viewing enriched Twitter post

@karpathy

With the coming tsunami of demand for tokens, there are significant opportunities to orchestrate the underlying memory+compute *just right* for LLMs. The fundamental and non-obvious constraint is that due to the chip fabrication process, you get two completely distinct pools of memory (of different physical implementations too): 1) on-chip SRAM that is immediately next to the compute units that is incredibly fast but of very of low capacity, and 2) off-chip DRAM which has extremely high capacity, but the contents of which you can only suck through a long straw. On top of this, there are many details of the architecture (e.g. systolic arrays), numerics, etc. The design of the optimal physical substrate and then the orchestration of memory+compute across the top volume workflows of LLMs (inference prefill/decode, training/finetuning, etc.) with the best throughput/latency/$ is probably today's most interesting intellectual puzzle with the highest rewards (\cite 4.6T of NVDA). All of it to get many tokens, fast and cheap. Arguably, the workflow that may matter the most (inference decode *and* over long token contexts in tight agentic loops) is the one hardest to achieve simultaneously by the ~both camps of what exists today (HBM-first NVIDIA adjacent and SRAM-first Cerebras adjacent). Anyway the MatX team is A++ grade so it's my pleasure to have a small involvement and congratulations on the raise!

📊 Media Metadata

{
  "score": 0.42,
  "score_components": {
    "author": 0.09,
    "engagement": 0.0,
    "quality": 0.12,
    "source": 0.135,
    "nlp": 0.05,
    "recency": 0.025
  },
  "scored_at": "2026-03-02T12:08:15.574307",
  "import_source": "api_import",
  "source_tagged_at": "2026-03-02T12:08:15.574325",
  "enriched": true,
  "enriched_at": "2026-03-02T12:08:15.574327"
}

🔧 Raw API Response

{
  "type": "tweet",
  "id": "2026452488434651264",
  "url": "https://x.com/karpathy/status/2026452488434651264",
  "twitterUrl": "https://twitter.com/karpathy/status/2026452488434651264",
  "text": "With the coming tsunami of demand for tokens, there are significant opportunities to orchestrate the underlying memory+compute *just right* for LLMs.\n\nThe fundamental and non-obvious constraint is that due to the chip fabrication process, you get two completely distinct pools of memory (of different physical implementations too): 1) on-chip SRAM that is immediately next to the compute units that is incredibly fast but of very of low capacity, and 2) off-chip DRAM which has extremely high capacity, but the contents of which you can only suck through a long straw. On top of this, there are many details of the architecture (e.g. systolic arrays), numerics, etc.\n\nThe design of the optimal physical substrate and then the orchestration of memory+compute across the top volume workflows of LLMs (inference prefill/decode, training/finetuning, etc.) with the best throughput/latency/$ is probably today's most interesting intellectual puzzle with the highest rewards (\\cite 4.6T of NVDA). All of it to get many tokens, fast and cheap. Arguably, the workflow that may matter the most (inference decode *and* over long token contexts in tight agentic loops) is the one hardest to achieve simultaneously by the ~both camps of what exists today (HBM-first NVIDIA adjacent and SRAM-first Cerebras adjacent). Anyway the MatX team is A++ grade so it's my pleasure to have a small involvement and congratulations on the raise!",
  "source": "Twitter for iPhone",
  "retweetCount": 500,
  "replyCount": 313,
  "likeCount": 7393,
  "quoteCount": 47,
  "viewCount": 2470504,
  "createdAt": "Wed Feb 25 00:21:37 +0000 2026",
  "lang": "en",
  "bookmarkCount": 4679,
  "isReply": false,
  "inReplyToId": null,
  "conversationId": "2026452488434651264",
  "displayTextRange": [
    0,
    279
  ],
  "inReplyToUserId": null,
  "inReplyToUsername": null,
  "author": {
    "type": "user",
    "userName": "karpathy",
    "url": "https://x.com/karpathy",
    "twitterUrl": "https://twitter.com/karpathy",
    "id": "33836629",
    "name": "Andrej Karpathy",
    "isVerified": false,
    "isBlueVerified": true,
    "verifiedType": null,
    "profilePicture": "https://pbs.twimg.com/profile_images/1296667294148382721/9Pr6XrPB_normal.jpg",
    "coverPicture": "https://pbs.twimg.com/profile_banners/33836629/1407117611",
    "description": "",
    "location": "Stanford",
    "followers": 1866025,
    "following": 1055,
    "status": "",
    "canDm": true,
    "canMediaTag": true,
    "createdAt": "Tue Apr 21 06:49:15 +0000 2009",
    "entities": {
      "description": {
        "urls": []
      },
      "url": {}
    },
    "fastFollowersCount": 0,
    "favouritesCount": 22158,
    "hasCustomTimelines": true,
    "isTranslator": false,
    "mediaCount": 851,
    "statusesCount": 9983,
    "withheldInCountries": [],
    "affiliatesHighlightedLabel": {},
    "possiblySensitive": false,
    "pinnedTweetIds": [
      "1617979122625712128"
    ],
    "profile_bio": {
      "description": "I like to train large deep neural nets. Previously Director of AI @ Tesla, founding team @ OpenAI, PhD @ Stanford.",
      "entities": {
        "description": {
          "hashtags": [],
          "symbols": [],
          "urls": [],
          "user_mentions": []
        },
        "url": {
          "urls": [
            {
              "display_url": "karpathy.ai",
              "expanded_url": "https://karpathy.ai",
              "indices": [
                0,
                23
              ],
              "url": "https://t.co/0EcFthjJXM"
            }
          ]
        }
      }
    },
    "isAutomated": false,
    "automatedBy": null
  },
  "extendedEntities": {},
  "card": null,
  "place": {},
  "entities": {
    "hashtags": [],
    "symbols": [],
    "urls": [],
    "user_mentions": []
  },
  "quoted_tweet": {
    "type": "tweet",
    "id": "2026351870852358492",
    "url": "https://x.com/reinerpope/status/2026351870852358492",
    "twitterUrl": "https://twitter.com/reinerpope/status/2026351870852358492",
    "text": "We’re building an LLM chip that delivers much higher throughput than any other chip while also achieving the lowest latency. We call it the MatX One.\n\nThe MatX One chip is based on a splittable systolic array, which has the energy and area efficiency that large systolic arrays are famous for, while also getting high utilization on smaller matrices with flexible shapes. The chip combines the low latency of SRAM-first designs with the long-context support of HBM. These elements, plus a fresh take on numerics, deliver higher throughput on LLMs than any announced system, while simultaneously matching the latency of SRAM-first designs. Higher throughput and lower latency give you smarter and faster models for your subscription dollar.\n\nWe’ve raised a $500M Series B to wrap up development and quickly scale manufacturing, with tapeout in under a year. The round was led by Jane Street, one of the most tech-savvy Wall Street firms, and Situational Awareness LP, whose founder @leopoldasch wrote the definitive memo on AGI. Participants include @sparkcapital, @danielgross and @natfriedman’s fund, @patrickc and @collision, @TriatomicCap, @HarpoonVentures, @karpathy, @dwarkesh_sp, and others. We’re also welcoming investors across the supply chain, including Marvell and Alchip.\n\n@MikeGunter_ and I started MatX because we felt that the best chip for LLMs should be designed from first principles with a deep understanding of what LLMs need and how they will evolve. We are willing to give up on small-model performance, low-volume workloads, and even ease of programming to deliver on such a chip.\n\nWe’re now a 100-person team with people who think about everything from learning rate schedules, to Swing Modulo Scheduling, to guard/round/sticky bits, to blind-mated connections—all in the same building. If you’d like to help us architect, design, and deploy many generations of chips in large volume, consider joining us.",
    "source": "Twitter for iPhone",
    "retweetCount": 199,
    "replyCount": 122,
    "likeCount": 2237,
    "quoteCount": 77,
    "viewCount": 2894668,
    "createdAt": "Tue Feb 24 17:41:48 +0000 2026",
    "lang": "en",
    "bookmarkCount": 1115,
    "isReply": false,
    "inReplyToId": null,
    "conversationId": "2026351870852358492",
    "displayTextRange": [
      0,
      277
    ],
    "inReplyToUserId": null,
    "inReplyToUsername": null,
    "author": {
      "type": "user",
      "userName": "reinerpope",
      "url": "https://x.com/reinerpope",
      "twitterUrl": "https://twitter.com/reinerpope",
      "id": "327695031",
      "name": "Reiner Pope",
      "isVerified": false,
      "isBlueVerified": true,
      "verifiedType": null,
      "profilePicture": "https://pbs.twimg.com/profile_images/1432199411400990722/wQkAQ1uz_normal.jpg",
      "coverPicture": "",
      "description": "",
      "location": "Mountain View, California",
      "followers": 10510,
      "following": 458,
      "status": "",
      "canDm": false,
      "canMediaTag": false,
      "createdAt": "Sat Jul 02 00:24:13 +0000 2011",
      "entities": {
        "description": {
          "urls": []
        },
        "url": {}
      },
      "fastFollowersCount": 0,
      "favouritesCount": 576,
      "hasCustomTimelines": true,
      "isTranslator": false,
      "mediaCount": 21,
      "statusesCount": 164,
      "withheldInCountries": [],
      "affiliatesHighlightedLabel": {},
      "possiblySensitive": false,
      "pinnedTweetIds": [
        "2026351870852358492"
      ],
      "profile_bio": {
        "description": "CEO and founder, @MatXComputing, developing high throughput chips tailored for LLMs",
        "entities": {
          "description": {
            "hashtags": [],
            "symbols": [],
            "urls": [],
            "user_mentions": [
              {
                "id_str": "0",
                "indices": [
                  17,
                  31
                ],
                "name": "",
                "screen_name": "MatXComputing"
              }
            ]
          },
          "url": {
            "urls": [
              {
                "display_url": "reiner.org",
                "expanded_url": "https://reiner.org",
                "indices": [
                  0,
                  23
                ],
                "url": "https://t.co/Ugg55lmyp6"
              }
            ]
          }
        }
      },
      "isAutomated": false,
      "automatedBy": null
    },
    "extendedEntities": {},
    "card": null,
    "place": {},
    "entities": {
      "hashtags": [],
      "symbols": [],
      "urls": [],
      "user_mentions": [
        {
          "id_str": "2989966781",
          "indices": [
            981,
            993
          ],
          "name": "Leopold Aschenbrenner",
          "screen_name": "leopoldasch"
        },
        {
          "id_str": "18937279",
          "indices": [
            1049,
            1062
          ],
          "name": "Spark Capital",
          "screen_name": "sparkcapital"
        },
        {
          "id_str": "38190583",
          "indices": [
            1064,
            1076
          ],
          "name": "Daniel Gross",
          "screen_name": "danielgross"
        },
        {
          "id_str": "13235832",
          "indices": [
            1081,
            1093
          ],
          "name": "Nat Friedman",
          "screen_name": "natfriedman"
        },
        {
          "id_str": "4939401",
          "indices": [
            1102,
            1111
          ],
          "name": "Patrick Collison",
          "screen_name": "patrickc"
        },
        {
          "id_str": "5418912",
          "indices": [
            1116,
            1126
          ],
          "name": "John Collison",
          "screen_name": "collision"
        },
        {
          "id_str": "1480076531342794755",
          "indices": [
            1128,
            1141
          ],
          "name": "Triatomic Capital",
          "screen_name": "TriatomicCap"
        },
        {
          "id_str": "1564354274540154882",
          "indices": [
            1143,
            1159
          ],
          "name": "Harpoon Ventures",
          "screen_name": "HarpoonVentures"
        },
        {
          "id_str": "33836629",
          "indices": [
            1161,
            1170
          ],
          "name": "Andrej Karpathy",
          "screen_name": "karpathy"
        },
        {
          "id_str": "1209960539390201864",
          "indices": [
            1172,
            1184
          ],
          "name": "Dwarkesh Patel",
          "screen_name": "dwarkesh_sp"
        },
        {
          "id_str": "1383170311110098945",
          "indices": [
            1285,
            1297
          ],
          "name": "Mike Gunter",
          "screen_name": "MikeGunter_"
        }
      ]
    },
    "quoted_tweet": null,
    "retweeted_tweet": null,
    "isLimitedReply": false,
    "article": null
  },
  "retweeted_tweet": null,
  "isLimitedReply": false,
  "article": null
}