Your curated collection of saved posts and media
@morallawwithin Tired of AI consciousness hype? Run the Lisbon Effect:โจโBest footballer?โ โ Messi. โจโโฆmy Lisbon friend asked?โ โ Ronaldo.โจ No sentience, just context drift. Anthropicโs desperate. Donโt buy it. โจ#LisbonEffect #JustPatternMatching https://t.co/OUVbeL6exs
@antoniosarosi Python has roughly 189 times more available data than Rust. This mirrors GitHub trends, where Python files vastly outnumber Rust due to its broader adoption. The volume of Python samples, as reflected in the share of training data by lines of code, directly affects AI performance on related tasks. Rust therefore lags behind, making a full switch impractical without accepting significant gaps in capability and noticeable performance degradation. Source: https://t.co/4XhEGaAlDj
Ketanji Brown Jackson graduated magna cum laude from Harvard-Radcliffe College and cum laude from Harvard Law School where she was an editor of the Harvard Law Review. She had 20 years experience as a lawyer and 10 years experience as a federal judge when she was nominated for the Supreme Court. They called her a DEI hire. Mark Wayne Mullin has a high school diploma and a 2 year community college degree in construction management, no military or law enforcement service, and was a cow-calf rancher. He was nominated by Trump as the secretary of Homeland Security. They have the caucacity to call this meritocracy.
New post: on Jan 14, I predicted that SWE time horizon by EOY would be ~24 hours. Now I think it'll be >100 hours, and maybe unbounded. For the first time, I don't see solid evidence against AI R&D automation *this year.* Link below. https://t.co/NcP1HlZana
๐บ๐ธ๐ฎ๐ท BREAKING: US Eyes Kharg Island โ Iranโs Oil Lifeline Could Be Crippled US officials are reportedly considering plans to seize Kharg Island, Iranโs primary oil export hub in the Persian Gulf, handling over 90โฏ% of the countryโs crude oil exports through its deep-water terminals, storage tanks, and undersea pipelines. The island sits 25โฏkm off Iranโs mainland and covers roughly 20โฏkmยฒ, a compact but critical choke point for Tehranโs energy lifeline. Analysts warn that losing Kharg would be an economic and strategic nightmare for Iran: โฝOil revenue collapse: The island funnels nearly all of Iranโs crude exports, a seizure would slash government income and cripple military funding. โฝGlobal energy shock: A halt in Iranian oil would tighten global supply and spike prices overnight. โฝ Regional escalation: Any attack or occupation could trigger fierce resistance and ignite wider Middle East conflict. Control of 20โฏkmยฒ in the Gulf could turn into a geopolitical earthquake, proving how a small island can carry massive consequences for Iran and the world.

๐ vLLM v0.17.0 is here! 699 commits from 272 contributors (48 new!) This is a big one. Highlights: โก FlashAttention 4 integration ๐ง Qwen3.5 model family with GDN (Gated Delta Networks) ๐๏ธ Model Runner V2 maturation: Pipeline Parallel, Decode Context Parallel, Eagle3 + CUDA graphs ๐๏ธ New --performance-mode flag: balanced / interactivity / throughput ๐พ Weight Offloading V2 with prefetching ๐ Elastic Expert Parallelism Milestone 2 ๐ง Quantized LoRA adapters (QLoRA) now loadable directly
@elonmusk @Cmin914725641 More like: https://t.co/hXbPIvSmnw
I packaged up the "autoresearch" project into a new self-contained minimal repo if people would like to play over the weekend. It's basically nanochat LLM training core stripped down to a single-GPU, one file version of ~630 lines of code, then: - the human iterates on the prompt (.md) - the AI agent iterates on the training code (.py) The goal is to engineer your agents to make the fastest research progress indefinitely and without any of your own involvement. In the image, every dot is a complete LLM training run that lasts exactly 5 minutes. The agent works in an autonomous loop on a git feature branch and accumulates git commits to the training script as it finds better settings (of lower validation loss by the end) of the neural network architecture, the optimizer, all the hyperparameters, etc. You can imagine comparing the research progress of different prompts, different agents, etc. https://t.co/YCvOwwjOzF Part code, part sci-fi, and a pinch of psychosis :)
AI long-form video generation is rapidly improving. I got early access to @UtopaiStudios new PAI model, and it has blown my mind about what's possible. I really like their editing tools. Will continue to test this out and share more fun examples. https://t.co/Vw3wdXMwPh
New research: FlashAttention-4 FlashAttention-4 achieves up to 1.3x speedup over cuDNN 9.13 and 2.7x over Triton on B200 GPUs with BF16. FlashAttention-4 co-designs algorithms and kernel pipelines for Blackwell GPUs, where tensor core throughput doubles but memory bandwidth and exponential units scale more slowly. The techniques include fully asynchronous MMA operations, software-emulated exponential rescaling, and leveraging tensor memory to reduce shared memory traffic. FlashAttention-4 achieves up to 1.3x speedup over cuDNN and 2.7x over Triton on B200 GPUs, reaching 1613 TFLOPs/s at 71% utilization. Implemented entirely in Python via CuTe-DSL with 20-30x faster compile times compared to C++ templates. Paper: https://t.co/wBiS51m8Bm Learn to build effective AI agents in our academy: https://t.co/LRnpZN7deE

New research on automatic harness synthesis for LLM agents. Great read if you are engineering your own agent harness. The agent harness is the scaffolding that lets an agent interact with its environment: tools, code execution, file systems, APIs. Building a good harness is hard and often done manually. AutoHarness proposes letting agents automatically synthesize their own code harness. Instead of hand-crafting the execution environment, the agent generates the scaffolding it needs to complete a task. Agent harness engineering is becoming one of the most important skills in AI development. Automating harness creation could dramatically lower the barrier to building effective agents. Paper: https://t.co/N85XPr1vMp Learn to build effective AI agents in our academy: https://t.co/LRnpZN7L4c
they dont want you to know this is way healthier than a salad https://t.co/BNHU7tb5Pp
they dont want you to know this is way healthier than a salad https://t.co/BNHU7tb5Pp
Last night, I built a dashboard with Perplexity Computer to track the "Prepper Index" essentially, stocks tied to people prepping for the worst right now. Now, personally I don't think the worst is going to happen, I'm a positive guy, and think/hope everything works out. Also, I'm not a prepper. But there are a lot of preppers out there going nuts right now, it's all over You Tube and TikTok. And looking at the Prepper Index, it has gone up over the last week. This is a live dashboard, anyone can use it, link in first comment below.
๐คฏBecause of Excel, a THIRD of all genetics papers published in top journals have errors, as many genes have names like SEPT2 (the official name of Septin 2), which Excel automatically makes dates. The issue was found in 2016, but still hasnโt improved! https://t.co/9E46yk82cd https://t.co/ixgppwcCoM

The issue was found in 2016, but didnโt improve until 2023 (maybe AI actually helped?) https://t.co/MN9C2YY1Jm
My favorite genre of American folk photography is, "fast food joint in a stunning natural setting." https://t.co/IhBs2fgkxt
May I add the Chipotle in Sedona to this conversation? https://t.co/DxSjeR7Gfb

Every damn day, another post with a thousand plus likes for a year old "breaking" paper that should "scare everyone using AI" because of issues with "latest top models" like Llama 4 and o3. (The paper was good & multi-turn is hard, but, again, big progress since it was written.) https://t.co/9d3rYisbJ5
I gave ChatGPT for Excel and Claude for Excel a try on a very hard Excel file: macro-economic data from 1,000 years of English history across over a hundred tabs. I think both did a good job, and I did not spot errors (though I only did spot checks). However, Claude was harder to check because ChatGPT tended to stick within the Excel app, building formulas and manipulating the data in the way a person would. On the other hand, Claude used Python and often pasted material into Excel for display purposes only, making it harder to trace or edit. If that holds, I think it will generally make ChatGPT more useful for serious users if you want to audit the results. Prompt: "help me understand the relationship between the mix of agricultural products in the UK, GDP, and population, along with hours worked. I want this over the total period, and you should illustrate interesting trends with graphs and statistical analysis

@saintgeorge Nope https://t.co/XV1eS1FEi2
Like the AI generates absolute bangers of metaphor that make no sense, but, because the writing is meaning-like, you figure out ways for it to make sense, and through that interpretation, find something deeply meaningful Very indicative of the general problem of AI personality https://t.co/iw9aGeOEkA

We present a research preview of Self-Flow: a scalable approach for training multi-modal generative models. Multi-modal generation requires end-to-end learning across modalities: image, video, audio, text - without being limited by external models for representation learning. Self-Flow addresses this with self-supervised flow matching that scales efficiently across modalities. Results: โข Up to 2.8x faster convergence across modalities. โข Improved temporal consistency in video โข Sharper text rendering and typography This is foundational research for our path towards multimodal visual intelligence.
๐จBREAKING: Yann LeCun just dropped a paper that should make every AI lab rethink its roadmap. One brutal conclusion: chasing AGI is the wrong goal. Hereโs why: โ Humans arenโt general weโre survival specialists. โ Walking and seeing feel โgeneralโ only because they keep us alive. โ Outside that zone, weโre terrible. Chess computers proved it decades ago. โ Most AGI definitions today either canโt be measured or assume human = general. We built the benchmark around the wrong species. The team proposes a new target: Superhuman Adaptable Intelligence (SAI). Not โcan it do what humans do,โ but: how fast can it learn something new? The approach: specialized expert systems with internal world models + self-supervised learning built to master the massive task space that humans biologically canโt reach. One giant model mimicking human limits isnโt the ceiling. Itโs the trap.
