Your curated collection of saved posts and media
ethan never got a real answer to his question and it shows https://t.co/DT5OJc7kOW
If you ever want to see a really interesting AI thinking trace, push it really hard on literature or poetry suggestions. Here is Claude 4.6 Opus working through poetry when I asked it to find something that captures the feeling of AI while avoiding its usual favorites (eg Rilke) https://t.co/B0phEv3qIN

95% of cancer drugs that work in mice fail in humans. Weβve spent 50 years testing drugs on the wrong species. Itβs time to talk about this. π§΅ https://t.co/MKp9akwBQJ
I used to marvel at new tech and now I'm just over it. This encapsulates how I feel about any new app or tech device these days. https://t.co/LIB7tGn6cz
Trump (in 2011): βOur president will start a war with Iran because he has absolutely no ability to negotiate. Heβs weak and heβs ineffective.β https://t.co/Y1sER7ENTO
Pretty crazy when you realize just how flagrant the scaremongering has been about data center water usage: https://t.co/U7rFRgKPko
@newrepublic @Prof_Sugon_Deez https://t.co/KTMPjgMkp1
@newrepublic @Prof_Sugon_Deez https://t.co/KTMPjgMkp1
JD Vance, October 2024: βOut interests, I think, very much, is a not going to war in Iran." βKamala Harris kinda likes warβ¦They seem to be sleepwalking us into a war with Iran." https://t.co/ptmhOuJJhW
This is the AI that will be taking our jobs https://t.co/nycRqJimm6
sneak peek of Anthropic's 2026 Super Bowl ad https://t.co/6v5ZFiHYgO
sneak peek of Anthropic's 2026 Super Bowl ad https://t.co/6v5ZFiHYgO
Video generation models are improving fastβreal-time autoregressive models now deliver high quality at low latency, and theyβre quickly being adopted for world models and robotics applications. So whatβs the problem? Theyβre still too slow on consumer hardware. π What if we told you that we can get true real-time 16 FPS video generation on a single RTX 5090? (1.5-12x over FA 2/3/4 on 5090, H100, B200) Today we release MonarchRT π¦, an efficient video attention that parameterizes attention maps as (tiled) Monarch matrices and delivers real E2E gains. π Paper: https://t.co/d1AAMIseow π Website: https://t.co/41mqriKekx π GitHub: https://t.co/hp5iJttviA π§΅1/n
We identified an issue with the Mamba-2 π initialization in HuggingFace and FlashLinearAttention repository (dt_bias being incorrectly initialized). This bug is related to 2 main issues: 1. init being incorrect (torch.ones) if Mamba-2 layers are used in isolation without the Mamba2ForCausalLM model class (this has been already fixed: https://t.co/oahfxjIsKb). 2. Skipping initialization due to meta device init for DTensors with FSDP-2 (https://t.co/hLC8nnQFc3 will fix this issue upon merging). The difference is substantial. Mamba-2 seems to be quite sensitive to the initialization. Check out our experiments at the 7B MoE scale: https://t.co/n8iuUICRux Special thanks to @kevinyli_, @bharatrunwal2, @HanGuo97, @tri_dao and @_albertgu π Also thanks to @SonglinYang4 for quickly helping in merging the PR.

@Kaivalya_in @milindmghosh Cohere: https://t.co/eQGWmi0eM6 Sarashina: https://t.co/plQG6qPAqT but it looks like the first in Japan was actually Stockmark100B which beat it by a few months: https://t.co/q3sGmwogg1

@paws4puzzles @milindmghosh Firstly, DeepSeek and Qwen have both released multiple much larger and more powerful models with MIT and Apache 2.0 licenses. Secondly, Falcon 180B doesn't have an Apache 2.0 license: https://t.co/lYDKI8ox1f
@milindmghosh Update: the Stockmark 100B model by Stockmark is actually the first 100B model from Japan, coming out in May as opposed to Sarashina2 in November. This doesn't change the order becasue Cohere Command R+ came out in April. https://t.co/q3sGmwogg1
In this amazing multidisciplinary collaboration, we report our early experience with the @openclaw -> https://t.co/THXYyajfQB
Some real cognitive dissonance happening with takes saying "but Anthropic HAD to drop their safety measures, they're the good guys you see!" Anyway from our paper last year: https://t.co/d0yyWfx0fe
A new filing just dropped in the Musk v. Altman case, and it may be the most brazen and cynical document OpenAI has produced yet. It's a motion to exclude the testimony of Stuart Russell, but their attacks blatantly contradict things @OpenAI itself has said for years. π§΅ https://t.co/WSPSpNiYqV
@UnslothAI Btw, I have some anecdotal evidence that disabling thinking for GLM-4.7-Flash improves performance for agentic coding stuff. Haven't evaluated in detail yet (only opencode) as it takes time, but would be interest to know if you give it a try and share your observations. To disable thinking with llama.cpp add this to the llama-server command: --chat-template-kwargs "{\"enable_thinking\": false}" Here is my config for reference:
Introducing LlamaBarn β a tiny macOS menu bar app for running local LLMs Open source, built on llama.cpp https://t.co/F1Z3DVl9Kg
I am deeply thankful to the Hugging Face team for this opportunity. With their support I will be able to continue my work on the projects and I feel optimistic about the great stuff that we are going to create with the community! https://t.co/Gl95jBPhly
Introducing LM Link β¨ Connect to remote instances of LM Studio, securely. π End-to-end encrypted π‘ Load models locally, use them on the go π₯οΈ Use local devices, LLM rigs, or cloud VMs Launching in partnership with @Tailscale Try it now: https://t.co/Vl2vr6HlF5