Your curated collection of saved posts and media
@CharlieBul58993 @JTillipman @bridgewriter (former NSC counsel) - https://t.co/K8WEStCDhc
A deep dive in @lawfare on the many legal problems with the Pentagon's designation of Anthropic as a supply chain risk. https://t.co/6mlWhgwMge
New research just exposed the biggest lie in AI coding benchmarks. LLMs score 84-89% on standard coding tests. On real production code? 25-34%. That's not a gap. That's a different reality. Here's what happened: Researchers built a benchmark from actual open-source repositories real classes with real dependencies, real type systems, real integration complexity. Then they tested the same models that dominate HumanEval leaderboards. The results were brutal. The models weren't failing because the code was "harder." They were failing because it was *real*. Synthetic benchmarks test whether a model can write a self-contained function with a clean docstring. Production code requires understanding inheritance hierarchies, framework integrations, and project-specific utilities. Different universe. Same leaderboard score. But it gets worse. A separate study ran 600,000 debugging experiments across 9 LLMs. They found a bug in a program. The LLM found it too. Then they renamed a variable. Added a comment. Shuffled function order. Changed nothing about the bug itself. The LLM couldn't find the same bug anymore. 78% of the time, cosmetic changes that don't affect program behavior completely broke the model's ability to debug. Function shuffling alone reduced debugging accuracy by 83%. The models aren't reading code. They're pattern-matching against what code *looks like* in their training data. A third study confirmed this from another angle: when researchers obfuscated real-world code changing symbols, structure, and semantics while keeping functionality identical LLM pass rates dropped by up to 62.5%. The researchers call this the "Specialist in Familiarity" problem. LLMs perform well on code they've memorized. The moment you show them something unfamiliar with the same logic, they collapse. Three papers. Three different methodologies. Same conclusion: The benchmarks we use to evaluate AI coding tools are measuring memorization, not understanding. If you're shipping code generated by LLMs into production without review, these numbers should concern you. If you're building developer tools, the question isn't "what's your HumanEval score." It's "what happens when the code doesn't look like the training data."
Gift link: https://t.co/S1D5ZMpE3l
π« @bradlightcap has stopped following @GaryMarcus (π€π: any thoughts on this?) https://t.co/kI0mBNCoxY
Gift link: https://t.co/S1D5ZMpE3l
In the last few days, OpenAI and its executives have claimed that its DoW deal prevents its models being used for mass domestic surveillance. As I write in a lengthy explainer for @ReadTransformer today, that appears to be misleading at best. https://t.co/IdlpVUSY0p
Be like Sam Altman > runs YC > starts a open-sourced non profit to regulate ai & protect humanity > raise money for non profit > use that money to build a closed source AI > create a new for profit company > raise money & kick out existing investors > use our data for ads in ChatGPT > go on the news and stand up for Anthropic against US gov > 24hrs later sign a deal with US to do exactly the opposite
Folks, this is not normal. Four American soldiers have died, but let me tell you about the curtains. βI always liked gold.β https://t.co/1Kt9tvNi8g
Shocked! https://t.co/8EX9ADZibS
Satya Nadella just said what the entire industry is too invested to admit. Every CEO signing $100 billion data center contracts right now is making a bet that history may not honor. Nadella: βWe are one sort of innovation away from the entire regime changing.β Right now, every major player is running the same play. More data. More GPUs. Bigger clusters. Same architecture. Theyβve convinced themselves scale is destiny. Theyβve convinced themselves the biggest balance sheet wins. Theyβve convinced themselves this is a resource war. Itβs not. Nadella: βIf you look at where weβve gone, it was all about pre-training scale, then it was about post-training, then we came up with reasoning, then we said, βoh, thereβs RL.ββ The architecture isnβt stable. It never was. Itβs been mutating the entire time. Each shift rewriting the rules. Each breakthrough making the previous moat irrelevant. And the companies that didnβt see it coming didnβt get a warning. They just woke up behind. Nadella: βA new model architecture that could even be more efficient in its performance.β When that lands, the $100 billion clusters donβt matter. The hoarded GPUs donβt matter. The multi-decade infrastructure advantage doesnβt matter. Every castle built for the current paradigm becomes a monument to the wrong bet. This is what makes the AI race unlike anything in history. In nuclear competition, more warheads meant more power. The advantage was permanent. Cumulative. Compounding. In this race, one person with the right insight at 2am in an apartment somewhere erases a trillion dollars of infrastructure before the market opens. No warning. No negotiation. No second place. The most dangerous competitor in this race doesnβt have a data center. They just have the equation.
Anyone else having those weird dreams where future generations hate you? From @TheOnion https://t.co/y2OI4SqdpC
Markets have a history of overreacting to narratives long before underlying economics change. AI is no exception. Capital tends to swing between euphoria and panic when it struggles to price uncertainty. The work, as always, happens between those extremes. What's your take on this development? @pchamard @Khulood_Almani @antgrasso @GlenGilmore @Shi4Tech @CurieuxExplorer @FrRonconi @chidambara09 @theomitsa @Analytics_699 @Nicochan33 @nafisalam @pierrepinna @smaksked @Corix_JC @amalmerzouk @AdityaRPatro @quepasachico @IngridVasiliu @EstelaMandela @sonu_monika @RLDI_Lamy @SpirosMargaris @IanLJones98 @Timothy_Hughes @avrohomg @bimedotcom @HaroldSinnott @c4trends @mvollmer1 @DG_Collective @bamitav @rwang0 @ipfconline1 @sijlalhussain https://t.co/ZxZhQSG0py
βI probably spend a third, maybe 40%, of my time making sure the culture of Anthropic is good,β Anthropic CEO Dario Amodei said. https://t.co/wATWuk7bZO https://t.co/Pbo7qTjhEm

Larry Ellison $ORCL highlighted something critical: models like ChatGPT, Gemini, Grok, and Llama are all trained on largely the same public internet data. When everyone trains on the same information, models inevitably converge. Thatβs why AI is moving toward commoditization. The real moat isnβt the model itself. Itβs the proprietary data behind it. Companies that can train on exclusive datasets gain an advantage competitors canβt replicate. Having data that no one else has will allow you to dominate your market.
Warren Buffett: "I like the way I lived 30 years ago β and I live that way now. The only difference is I have a plane to travel around privately. But in terms of what I eat, the clothes I wear, the books I read, the television I watch ... it's what I want to do in life." https://t.co/0XiZu1RgMJ
The CEO of a $380 billion AI company said something that should concern every developer, every startup, and every government on earth. He called open-source AI a "red herring." This is Dario Amodei, the man who runs Claude. His argument sounds technical but it's not and it's about money. Here's what he said: "I don't care whether a model is open source or not. The only thing I care about is, is it a good model?" Sounds reasonable, until you look at his books. 75% of Anthropic's $14 billion in revenue comes from one thing, charging companies per token to use Claude through an API. If enterprises could run their own models for free that revenue disappears. So when Amodei says open source "doesn't matter," what he means is, please don't look at open source. His technical argument, AI "open source" isn't real open source. You get the weights, just numbers not the actual source code and you can't see inside the model. Fair point but it misses the bigger picture. Companies running open source models don't need to see inside. They need three things, lower costs, data privacy, and freedom from vendor lock in. Open source somewhat delivers all three. A Berkeley study found open source AI models cost up to 90% less than closed APIs. Hospitals can keep patient data in house. Banks can meet compliance rules. Defense contractors don't send classified data to someone else's servers. Amodei brought up DeepSeek to prove his point. "I don't think it mattered that DeepSeek was open source," he said. But DeepSeek's release crashed Nvidia's stock by the biggest single day loss in market history and it mattered. Here's the pattern that keeps repeating in tech: Linux was a toy until it ran the internet. Android was "fragmented" until it owned 72% of mobile. Open source starts cheap, then it gets good, then it wins. Anthropic is posting record revenue. 500+ customers spending over $1M a year and 8 of the Fortune 10 on board. But so were BlackBerry and Sun Microsystems. So was every incumbent that dismissed the disruptor. The real question isn't whether open source is a red herring. But the real question is whether the man running a $380 billion closed source empire is the right person to ask.
The price of intelligence is falling fast. As AI becomes cheaper and more capable, agentic systems are starting to take over tasks once done by human workers. The shift isnβt gradual. Itβs economic. https://t.co/qpha8z0s7S @abcnews @AlanKohler
Inside Amazonβs layoffs, AI and βleanerβ operations are reshaping the culture. Survivorβs guilt and rising workloads are becoming part of the transition as automation accelerates. Whatβs happening there may become a blueprint other companies quietly follow. https://t.co/cRMN54FsNi @ft @rafeuddin_
AI is starting to shape scientific discovery itself. In particle physics, machine learning systems inside detectors now decide which signals are worth keeping and which are discarded. When algorithms filter reality, they quietly influence what scientists get to study. https://t.co/xjGX5V2QEZ @IEEESpectrum
Investors are piling into AI-resistant βhaloβ stocks. Heavy-asset, low-obsolescence companies are driving UK and EU markets to record highs. In the AI era, scarcity and stability are suddenly back in favor. https://t.co/aQwauFrnR9
McKinsey says agentic AI could fundamentally reshape global banking. But it warns banks not to get trapped in endless pilots and proofs of concept. The competitive edge wonβt go to those experimenting the longest, but to those scaling fastest. https://t.co/rrt7ROCCBl @DigWatchWorld @mckinsey
AI is beginning to decode the electrical noise inside our brains. Signals once thought too complex to interpret are now being translated into patterns and meaning. When machines start reading inner thoughts, neuroscience enters an entirely new era. https://t.co/4ZB8TUqxiG @LauraCReporter @bbcnews
AI literacy isnβt optional anymore. As automation spreads across industries, understanding how AI works becomes a core economic skill. If policymakers underestimate that urgency, the competitiveness gap will widen fast. https://t.co/ncHCj37goe @epc_eu https://t.co/XtYStA1f1V