Your curated collection of saved posts and media

Showing 32 posts Β· last 7 days Β· newest first
πŸ”jxnlco retweeted
S
Simon Willison
@simonw
πŸ“…
Mar 17, 2026
1h ago
πŸ†”68704943
⭐0.34

... and a follow-up chapter about Subagents, now a feature of Codex and Claude Code and Gemini CLI and Mistral Vibe and OpenCode and VS Code and Cursor https://t.co/suGmK4g3Hp

❀️17
likes
πŸ”3
retweets
πŸ”ivanleomk retweeted
A
agrim singh
@agrimsingh
πŸ“…
Mar 17, 2026
37m ago
πŸ†”08839930
⭐0.38

i dj'd a set this weekend and planned half of it with an app i built on @GoogleDeepMind's new multimodal embeddings model. it understands what music actually sounds like - not bpm, not genre tags. raw audio in, vibes out. built with @cursor_ai + @openai gpt 5.4, gemini embeddings and @convex to keep everything running smoothly @ericzakariasson @DynamicWebPaige @waynesutton @gabrielchua here's what it does (demo below) 🧡

❀️6
likes
πŸ”3
retweets
A
agrimsingh
@agrimsingh
πŸ“…
Mar 17, 2026
37m ago
πŸ†”08839930

i dj'd a set this weekend and planned half of it with an app i built on @GoogleDeepMind's new multimodal embeddings model. it understands what music actually sounds like - not bpm, not genre tags. raw audio in, vibes out. built with @cursor_ai + @openai gpt 5.4, gemini embeddings and @convex to keep everything running smoothly @ericzakariasson @DynamicWebPaige @waynesutton @gabrielchua here's what it does (demo below) 🧡

πŸ–ΌοΈ Media
πŸ”_akhaliq retweeted
N
Niels Rogge
@NielsRogge
πŸ“…
Mar 17, 2026
56m ago
πŸ†”69938152

Tried the viral poster-skill with Claude Code on the trending Moonshot paper :) Not too bad! https://t.co/I8lb0aUrbT

Media 1
❀️10
likes
πŸ”3
retweets
πŸ–ΌοΈ Media
N
NielsRogge
@NielsRogge
πŸ“…
Mar 17, 2026
56m ago
πŸ†”69938152

Tried the viral poster-skill with Claude Code on the trending Moonshot paper :) Not too bad! https://t.co/I8lb0aUrbT

@Kimi_Moonshot β€’ Mon Mar 16 03:03

Introducing π‘¨π’•π’•π’†π’π’•π’Šπ’π’ π‘Ήπ’†π’”π’Šπ’…π’–π’‚π’π’”: Rethinking depth-wise aggregation. Residual connections have long relied on fixed, uniform accumulation. Inspired by the duality of time and depth, we introduce Attention Residuals, replacing standard depth-wise recurrence with learned, input-dep

Media 1
πŸ–ΌοΈ Media
πŸ”Modular retweeted
F
Forward Future
@ForwardFuture
πŸ“…
Mar 13, 2026
3d ago
πŸ†”14582366
⭐0.32

β€œEveryone should be a GPU programmer.” @clattner_llvm's goal with @Modular: β€œWhat Modular is doing is opening up the box. We’re fixing the language problem and the platform problem. "The goal is to let more developers learn modern compute. And to give developers real choice in the hardware they use.” β€œThose two things unlock the ecosystem.”

❀️6
likes
πŸ”1
retweets
S
simonw
@simonw
πŸ“…
Mar 17, 2026
1h ago
πŸ†”68704943
⭐0.38

... and a follow-up chapter about Subagents, now a feature of Codex and Claude Code and Gemini CLI and Mistral Vibe and OpenCode and VS Code and Cursor https://t.co/suGmK4g3Hp

H
HamelHusain
@HamelHusain
πŸ“…
Mar 17, 2026
1h ago
πŸ†”89874302
⭐0.36

@matijagrcic @jxnlco I don’t doubt that the internal benchmarks are done rigorously with intellectual honesty The sad part is that many coding agents report their own internal benchmarks for which they are SOTA So it’s hard for the outside observer to put any stock in these despite best intentions

R
rasbt
@rasbt
πŸ“…
Mar 17, 2026
2h ago
πŸ†”71094774
⭐0.32

@SalajSonar1086 Already done :). The respective tutorial articles are linked via the β€œView in Article” links there

πŸ”huggingface retweeted
G
Gabriele Berton
@gabriberton
πŸ“…
Mar 16, 2026
15h ago
πŸ†”45334177
⭐0.34

VisMatch is on pypi! VisMatch is a wrapper for image matching models, like LightGlue, RoMa-v2, MASt3R, LoFTR, and 50+ more! It's literally as simple as: pip install vismatch vismatch-match --inputs img0 img1 --matcher choose_any To run image matching on any 2 images [1/4] https://t.co/dIr2YapWak

❀️183
likes
πŸ”28
retweets
πŸ”Scobleizer retweeted
R
Ray Fernando
@RayFernando1337
πŸ“…
Mar 16, 2026
16h ago
πŸ†”10226271
⭐0.34

Nvidia GTC 2026 OpenClaw Setup on DGX Spark IRL https://t.co/zQwwfCF9XP

❀️28
likes
πŸ”3
retweets
πŸ”Scobleizer retweeted
L
lucas
@lucas_flatwhite
πŸ“…
Mar 17, 2026
10h ago
πŸ†”99053607
⭐0.36

πŸ› οΈ Claude Code "opusplan" 말 κ·ΈλŒ€λ‘œ ν•˜μ΄λΈŒλ¦¬λ“œ λͺ¨λΈ.. κ³΅μ‹μž„! Claude Codeμ—λŠ” opusplan λͺ¨λΈμ„ 선택할 수 μžˆμ–΄μš”. > /model opusplan ν•˜μ΄λΈŒλ¦¬λ“œ λͺ¨λΈ alias인데, μž‘μ—… 단계에 따라 μžλ™μœΌλ‘œ λͺ¨λΈμ„ μ „ν™˜ν•΄μš”. λ³΅μž‘ν•œ 좔둠을 μœ„ν•œ ν”Œλžœ λͺ¨λ“œμ—μ„œλŠ” Opusλ₯Ό μ‹€ν–‰ λ‹¨κ³„μ—μ„œλŠ” Sonnet으둜 μžλ™ μ „ν™˜λ©λ‹ˆλ‹€! Opus둜 κ³„νšν•˜κ³  κ΅¬ν˜„κΉŒμ§€ ν•˜λŠ” 것도 λ¬Όλ‘  κ°€λŠ₯ν•΄μš”. ν•˜μ§€λ§Œ 이미 νƒ„νƒ„ν•œ κ³„νšμ΄ μžˆλ‹€λ©΄, 싀행은 SonnetμœΌλ‘œλ„ μΆ©λΆ„ν•˜κ³  더 μ €λ ΄ν•  수 μžˆμ–΄μš”. 각 μž‘μ—…μ— λ§žλŠ” λͺ¨λΈμ„ μ“°λŠ” 것 = 효율 πŸš€ ν”Œλž˜λ‹κ³Ό 싀행은 μš”κ΅¬λ˜λŠ” 인지 λΆ€ν•˜κ°€ λ‹¬λΌμš”. Opus의 κΉŠμ€ μΆ”λ‘  λŠ₯λ ₯은 κ³„νš 수립 λ‹¨κ³„μ—μ„œ κ°€μž₯ λΉ›λ‚˜κ³ , 일단 νƒ„νƒ„ν•œ κ³„νšμ΄ μ„Έμ›Œμ§„ μ΄ν›„μ˜ 싀행은 Sonnet으둜 μΆ©λΆ„νžˆ 컀버될 수 μžˆμ–΄μš”. μ–Έμ œ μ“°λ©΄ μ’‹λƒκ΅¬μš”? - λ³΅μž‘ν•œ κΈ°λŠ₯ 섀계같이 μ•„ν‚€ν…μ²˜ 결정이 μ€‘μš”ν•œ μž‘μ—… - λ¦¬νŒ©ν† λ§ κ³„νšκ°™μ€ 영ν–₯ λ²”μœ„ 뢄석이 ν•„μš”ν•œ 경우 - Opus ν’€νŒŒμ›Œ μ‚¬μš© λŒ€λΉ„ λΉ„μš© 절감이 ν•„μš”ν•  λ•Œ 이거 이제 많이 ν™œμš©ν•˜μ‹€ λ“―!!!

❀️22
likes
πŸ”5
retweets
D
drmapavone
@drmapavone
πŸ“…
Mar 17, 2026
10h ago
πŸ†”58775875

Jensen today announced Alpamayo 1.5 at #NVIDIAGTC! #Alpamayo 1.5 is a major update to Alpamayo 1β€”@nvidia’s open 10B-parameter chain-of-thought reasoning VLA model, first introduced at #CES. Built on the #Cosmos-Reason2 VLM backbone and post-trained with RL, it adds support for navigation guidance, flexible multi-camera setups, configurable camera parameters, and user question answering. The result is an interactive, steerable reasoning engine for the AV community. We’re also releasing post-training scripts to help researchers and developers adapt the model. Additionally, we’ve significantly expanded the Alpamayo open platform across data and simulation, including releasing highly requested reasoning labels for the PhysicalAI Autonomous Vehicles dataset (https://t.co/fD9eUcndya), as well as our chain-of-causation auto-labeling pipeline. πŸ”Ž Learn more about Alpamayo 1.5 and the latest extensions to the Alpamayo open platform: https://t.co/P0nuqkwBab (please note that most of the links will become active in the next few days.) Happy buildingβ€”and stay tuned for more in the coming months! @NVIDIADRIVE @NVIDIAAI

Media 2
+1 more
πŸ–ΌοΈ Media
L
lucas_flatwhite
@lucas_flatwhite
πŸ“…
Mar 17, 2026
10h ago
πŸ†”99053607

πŸ› οΈ Claude Code "opusplan" 말 κ·ΈλŒ€λ‘œ ν•˜μ΄λΈŒλ¦¬λ“œ λͺ¨λΈ.. κ³΅μ‹μž„! Claude Codeμ—λŠ” opusplan λͺ¨λΈμ„ 선택할 수 μžˆμ–΄μš”. > /model opusplan ν•˜μ΄λΈŒλ¦¬λ“œ λͺ¨λΈ alias인데, μž‘μ—… 단계에 따라 μžλ™μœΌλ‘œ λͺ¨λΈμ„ μ „ν™˜ν•΄μš”. λ³΅μž‘ν•œ 좔둠을 μœ„ν•œ ν”Œλžœ λͺ¨λ“œμ—μ„œλŠ” Opusλ₯Ό μ‹€ν–‰ λ‹¨κ³„μ—μ„œλŠ” Sonnet으둜 μžλ™ μ „ν™˜λ©λ‹ˆλ‹€! Opus둜 κ³„νšν•˜κ³  κ΅¬ν˜„κΉŒμ§€ ν•˜λŠ” 것도 λ¬Όλ‘  κ°€λŠ₯ν•΄μš”. ν•˜μ§€λ§Œ 이미 νƒ„νƒ„ν•œ κ³„νšμ΄ μžˆλ‹€λ©΄, 싀행은 SonnetμœΌλ‘œλ„ μΆ©λΆ„ν•˜κ³  더 μ €λ ΄ν•  수 μžˆμ–΄μš”. 각 μž‘μ—…μ— λ§žλŠ” λͺ¨λΈμ„ μ“°λŠ” 것 = 효율 πŸš€ ν”Œλž˜λ‹κ³Ό 싀행은 μš”κ΅¬λ˜λŠ” 인지 λΆ€ν•˜κ°€ λ‹¬λΌμš”. Opus의 κΉŠμ€ μΆ”λ‘  λŠ₯λ ₯은 κ³„νš 수립 λ‹¨κ³„μ—μ„œ κ°€μž₯ λΉ›λ‚˜κ³ , 일단 νƒ„νƒ„ν•œ κ³„νšμ΄ μ„Έμ›Œμ§„ μ΄ν›„μ˜ 싀행은 Sonnet으둜 μΆ©λΆ„νžˆ 컀버될 수 μžˆμ–΄μš”. μ–Έμ œ μ“°λ©΄ μ’‹λƒκ΅¬μš”? - λ³΅μž‘ν•œ κΈ°λŠ₯ 섀계같이 μ•„ν‚€ν…μ²˜ 결정이 μ€‘μš”ν•œ μž‘μ—… - λ¦¬νŒ©ν† λ§ κ³„νšκ°™μ€ 영ν–₯ λ²”μœ„ 뢄석이 ν•„μš”ν•œ 경우 - Opus ν’€νŒŒμ›Œ μ‚¬μš© λŒ€λΉ„ λΉ„μš© 절감이 ν•„μš”ν•  λ•Œ 이거 이제 많이 ν™œμš©ν•˜μ‹€ λ“―!!!

@dani_avila7 β€’ Mon Mar 16 15:58

Did you know about the opusplan model in Claude Code? /model opusplan It's a hybrid alias that automatically uses Opus in plan mode for complex reasoning, then switches to Sonnet for execution. Best of both worlds: Opus thinks, Sonnet builds https://t.co/r7un0X5bVg

πŸ–ΌοΈ Media
G
gdb
@gdb
πŸ“…
Mar 17, 2026
10h ago
πŸ†”37895367
⭐0.32

Subagents are now supported in Codex. They're very fun and make it possible to get large amounts of work done *quickly*:

@OpenAIDevs β€’ Mon Mar 16 20:09

Subagents are now available in Codex. You can accelerate your workflow by spinning up specialized agents to: β€’ Keep your main context window clean β€’ Tackle different parts of a task in parallel β€’ Steer individual agents as work unfolds https://t.co/QJC2ZYtYcA

πŸ”omarsar0 retweeted
O
elvis
@omarsar0
πŸ“…
Mar 16, 2026
1d ago
πŸ†”09077648
⭐0.38

Banger report from the Kimi team: Attention Residuals Residual connections made deep Transformers trainable. But they also force uncontrolled hidden-state growth with depth. This work proposes a cleaner alternative. It introduces Attention Residuals, which replace fixed residual accumulation with softmax attention over previous layer outputs. Instead of blindly summing everything, each layer selectively retrieves the earlier representations it actually needs. To keep this practical at scale, they add a blockwise version that compresses layers into block summaries, recovering most of the gains with minimal systems overhead. Why does it matter? Residual paths have barely changed across modern LLMs, even though they govern how information moves through depth. This paper shows that making the mixing content-dependent improves scaling laws, matches a baseline trained with 1.25x more compute, boosts GPQA-Diamond by +7.5 and HumanEval by +3.1, while keeping inference overhead under 2%. Paper: https://t.co/04IG6FDiVr Learn to build effective AI agents in our academy: https://t.co/1e8RZKs4uX

❀️130
likes
πŸ”17
retweets
πŸ”ivanleomk retweeted
Y
Yoeven
@yoeven
πŸ“…
Mar 17, 2026
13h ago
πŸ†”65291100

The moment he realised that https://t.co/vWmBsnR1nt isn't fully built on transformers and we can run on a single GPU with high accuracy and lower cost https://t.co/ZJYuL62UB8

Media 1
❀️4
likes
πŸ”1
retweets
πŸ–ΌοΈ Media
Y
yoeven
@yoeven
πŸ“…
Mar 17, 2026
13h ago
πŸ†”65291100

The moment he realised that https://t.co/vWmBsnR1nt isn't fully built on transformers and we can run on a single GPU with high accuracy and lower cost https://t.co/ZJYuL62UB8

Media 1Media 2
πŸ–ΌοΈ Media
πŸ”_akhaliq retweeted
H
Haocheng Xi
@HaochengXiUCB
πŸ“…
Mar 17, 2026
14h ago
πŸ†”24284251
⭐0.34

Thanks for sharing our newest work @_akhaliq ! Classic algorithms like K-Means deserve to be revisited in the era of massive datasets and GPUs. Flash-KMeans rethinks the algorithm from a systems perspective to make exact K-Means fast and memory-efficient on modern hardware.

❀️9
likes
πŸ”1
retweets
H
HamelHusain
@HamelHusain
πŸ“…
Mar 17, 2026
14h ago
πŸ†”89510900
⭐0.32

Claude Code CLI > Codex CLI Codex Desktop > Claude Code Desktop It’s a jagged UX frontier

H
HaochengXiUCB
@HaochengXiUCB
πŸ“…
Mar 17, 2026
14h ago
πŸ†”24284251
⭐0.38

Thanks for sharing our newest work @_akhaliq ! Classic algorithms like K-Means deserve to be revisited in the era of massive datasets and GPUs. Flash-KMeans rethinks the algorithm from a systems perspective to make exact K-Means fast and memory-efficient on modern hardware.

@_akhaliq β€’ Thu Mar 12 16:44

Flash-KMeans Fast and Memory-Efficient Exact K-Means paper: https://t.co/Yy7V7L12Bn https://t.co/c1mGipQl3f

C
code
@code
πŸ“…
Mar 17, 2026
14h ago
πŸ†”94910880

🌐 Agentic Browser Tools (Experimental) in @code! Agents can now open pages, read content, click elements, and verify changes directly in the integrated browser while building your web app. Enable βš™οΈ workbench.browser.enableChatTools to try it out. Learn mode: https://t.co/kNwugFcbIA

Media 2
πŸ–ΌοΈ Media
N
nvidianewsroom
@nvidianewsroom
πŸ“…
Mar 16, 2026
14h ago
πŸ†”99149451
⭐0.42

#NVIDIAGTC news: NVIDIA Dynamo 1.0 enters production as the broadly adopted inference operating system for AI factories. Dynamo 1.0 boosts Blackwell inference performance by up to 7x. The industry is scaling on NVIDIA. ⬇️https://t.co/Iaq2H2SmhR

P
PyTorch
@PyTorch
πŸ“…
Mar 16, 2026
15h ago
πŸ†”18333110

#ExecuTorch addresses fragmented native deployment for #AI agents as a #PyTorch native platform. It enables voice models across CPU, GPU, and NPU on Android, iOS, Linux, macOS & Windows πŸ”— https://t.co/NeQQyUniL4 https://t.co/O3itnoQFoG

Media 1
πŸ–ΌοΈ Media
G
gabriberton
@gabriberton
πŸ“…
Mar 16, 2026
15h ago
πŸ†”45334177

VisMatch is on pypi! VisMatch is a wrapper for image matching models, like LightGlue, RoMa-v2, MASt3R, LoFTR, and 50+ more! It's literally as simple as: pip install vismatch vismatch-match --inputs img0 img1 --matcher choose_any To run image matching on any 2 images [1/4] https://t.co/dIr2YapWak

Media 1
πŸ–ΌοΈ Media
πŸ”jxnlco retweeted
E
edwin
@edwinarbus
πŸ“…
Mar 16, 2026
19h ago
πŸ†”50334333
⭐0.34

Matt Maher tested frontier models in Cursor v. other harnesses. Cursor boosted model performance by 11% on average: Gemini: 52% β†’ 57% GPT-5.4: 82% β†’ 88% Opus: 77% β†’ 93% His benchmark measures how well models implement a 100-feature PRD. @cursor_ai consistently outperformed. https://t.co/hrjCmWMNKN

❀️176
likes
πŸ”17
retweets
Z
ZGojcic
@ZGojcic
πŸ“…
Mar 16, 2026
16h ago
πŸ†”28929828
⭐0.40

A new generation in AV simulation is here! We are announcing AlpaDreams, a real time interactive generative world model for AV simualtion! Just a year ago it took minutes to generate a few seconds of video, today it is real time and interactive! https://t.co/FbhKu3PMqe

R
RayFernando1337
@RayFernando1337
πŸ“…
Mar 16, 2026
16h ago
πŸ†”10226271
⭐0.34

Nvidia GTC 2026 OpenClaw Setup on DGX Spark IRL https://t.co/zQwwfCF9XP

J
jasteinerman
@jasteinerman
πŸ“…
Mar 16, 2026
17h ago
πŸ†”75976987
⭐0.32

Love this submission from our world models hackathon this weekend - a generative FPS!

@AnshulDhawan001 β€’ Mon Mar 16 21:08

Spent the weekend hacking at the Worlds in Action hackathon at @fdotinc by @SensAIHackademy. It was so much fun playing with the world models by @theworldlabs . I believe generative games are the future where characters, rules and even parts of the world can be generated and ad

A
ArtificialAnlys
@ArtificialAnlys
πŸ“…
Mar 16, 2026
18h ago
πŸ†”52868861

NVIDIA has released Nemotron 3 VoiceChat! A ~12B parameter Speech to Speech model that leads our open weights Conversational Dynamics vs. Speech Reasoning pareto frontier Understanding Speech to Speech model performance is multidimensional - two key and distinct dimensions are raw intelligence and conversational dynamics: how well a model handles the natural rhythms of human conversation such as turn-taking, interruptions. Amongst full duplex open weights models, NVIDIA’s new Nemotron 3 VoiceChat, V1, leads in balancing these dimensions, setting itself apart from other models on the Conversational Dynamics vs. Speech Reasoning pareto frontier. Key benchmarking results: ➀ Conversational Dynamics (Full Duplex Bench): Nemotron 3 VoiceChat (V1) scores 77.8%, second among open weights speech to speech models behind NVIDIA's own PersonaPlex (91.0%) and ahead of FLM-Audio (62.0%), Moshi (61.0%) and Freeze-Omni (58.7%) ➀ Speech Reasoning (Big Bench Audio): Nemotron 3 VoiceChat (V1) scores 29.2%, second among open weights speech to speech models behind Freeze-Omni (33.9%) and well ahead of PersonaPlex (12.6%), FLM-Audio (5.3%) and Moshi (1.7%) ➀ Pareto leader: While Freeze-Omni leads on speech reasoning and PersonaPlex leads on conversational dynamics, Nemotron 3 VoiceChat (V1) is the only open weights model that performs amongst the top 3 on both - making it the clear leader on the pareto frontier between these two critical dimensions ➀ Larger than other open weights models but still relatively small compared to LLMs: Nemotron 3 VoiceChat (V1) has 12B parameters, making it one of the larger open weights speech to speech models, while NVIDIA's PersonaPlex is ~7B. While larger compared to other larger open weights speech to speech models the model still is relatively small compared to leading LLMs ➀ Context vs. proprietary models: While this release materially advances open weights performance, open weights speech to speech models still significantly underperform leading proprietary offerings. For comparison, proprietary models on our Big Bench Audio benchmark score substantially higher - Step-Audio R1.1 at 96%, Grok Voice Agent at 92%, Gemini 2.5 Flash (Thinking) at 92%, and Nova 2.0 Sonic at 87%. The gap between open weights and proprietary remains large in this modality. As the capability and adoption of Speech to Speech models increases, we expect to expand our set of benchmarks to include elements such as tool-calling and multi-turn instruction following. See more details below ⬇️

Media 1
πŸ–ΌοΈ Media
πŸ”jeremyphoward retweeted
R
raia hadsell
@RaiaHadsell
πŸ“…
Mar 16, 2026
21h ago
πŸ†”56989392
⭐0.36

It's been about 20 years since I first started working on embeddings with Yann LeCun (siamese networks!), and I've been fascinated ever since. Gemini Embeddings 2 approaches the platonic ideal: native embedding of text, image, video, audio, and docs to a single space.

❀️277
likes
πŸ”24
retweets
H
HuggingPapers
@HuggingPapers
πŸ“…
Mar 16, 2026
18h ago
πŸ†”83694046

OmniForcing unlocks real-time joint audio-visual generation Achieves ~25 FPS with 0.7s latencyβ€”a 35Γ— speedup over offline diffusion modelsβ€”by distilling bidirectional LTX-2 into a causal streaming generator with maintained multi-modal fidelity. https://t.co/UGYGMyTQOs

Media 1
πŸ–ΌοΈ Media