Your curated collection of saved posts and media

Showing 32 posts Β· last 7 days Β· newest first
πŸ”huggingface retweeted
G
Gabriele Berton
@gabriberton
πŸ“…
Mar 16, 2026
10h ago
πŸ†”45334177
⭐0.34

VisMatch is on pypi! VisMatch is a wrapper for image matching models, like LightGlue, RoMa-v2, MASt3R, LoFTR, and 50+ more! It's literally as simple as: pip install vismatch vismatch-match --inputs img0 img1 --matcher choose_any To run image matching on any 2 images [1/4] https://t.co/dIr2YapWak

❀️183
likes
πŸ”28
retweets
πŸ”Scobleizer retweeted
R
Ray Fernando
@RayFernando1337
πŸ“…
Mar 16, 2026
11h ago
πŸ†”10226271
⭐0.34

Nvidia GTC 2026 OpenClaw Setup on DGX Spark IRL https://t.co/zQwwfCF9XP

❀️28
likes
πŸ”3
retweets
πŸ”Scobleizer retweeted
L
lucas
@lucas_flatwhite
πŸ“…
Mar 17, 2026
5h ago
πŸ†”99053607
⭐0.36

πŸ› οΈ Claude Code "opusplan" 말 κ·ΈλŒ€λ‘œ ν•˜μ΄λΈŒλ¦¬λ“œ λͺ¨λΈ.. κ³΅μ‹μž„! Claude Codeμ—λŠ” opusplan λͺ¨λΈμ„ 선택할 수 μžˆμ–΄μš”. > /model opusplan ν•˜μ΄λΈŒλ¦¬λ“œ λͺ¨λΈ alias인데, μž‘μ—… 단계에 따라 μžλ™μœΌλ‘œ λͺ¨λΈμ„ μ „ν™˜ν•΄μš”. λ³΅μž‘ν•œ 좔둠을 μœ„ν•œ ν”Œλžœ λͺ¨λ“œμ—μ„œλŠ” Opusλ₯Ό μ‹€ν–‰ λ‹¨κ³„μ—μ„œλŠ” Sonnet으둜 μžλ™ μ „ν™˜λ©λ‹ˆλ‹€! Opus둜 κ³„νšν•˜κ³  κ΅¬ν˜„κΉŒμ§€ ν•˜λŠ” 것도 λ¬Όλ‘  κ°€λŠ₯ν•΄μš”. ν•˜μ§€λ§Œ 이미 νƒ„νƒ„ν•œ κ³„νšμ΄ μžˆλ‹€λ©΄, 싀행은 SonnetμœΌλ‘œλ„ μΆ©λΆ„ν•˜κ³  더 μ €λ ΄ν•  수 μžˆμ–΄μš”. 각 μž‘μ—…μ— λ§žλŠ” λͺ¨λΈμ„ μ“°λŠ” 것 = 효율 πŸš€ ν”Œλž˜λ‹κ³Ό 싀행은 μš”κ΅¬λ˜λŠ” 인지 λΆ€ν•˜κ°€ λ‹¬λΌμš”. Opus의 κΉŠμ€ μΆ”λ‘  λŠ₯λ ₯은 κ³„νš 수립 λ‹¨κ³„μ—μ„œ κ°€μž₯ λΉ›λ‚˜κ³ , 일단 νƒ„νƒ„ν•œ κ³„νšμ΄ μ„Έμ›Œμ§„ μ΄ν›„μ˜ 싀행은 Sonnet으둜 μΆ©λΆ„νžˆ 컀버될 수 μžˆμ–΄μš”. μ–Έμ œ μ“°λ©΄ μ’‹λƒκ΅¬μš”? - λ³΅μž‘ν•œ κΈ°λŠ₯ 섀계같이 μ•„ν‚€ν…μ²˜ 결정이 μ€‘μš”ν•œ μž‘μ—… - λ¦¬νŒ©ν† λ§ κ³„νšκ°™μ€ 영ν–₯ λ²”μœ„ 뢄석이 ν•„μš”ν•œ 경우 - Opus ν’€νŒŒμ›Œ μ‚¬μš© λŒ€λΉ„ λΉ„μš© 절감이 ν•„μš”ν•  λ•Œ 이거 이제 많이 ν™œμš©ν•˜μ‹€ λ“―!!!

❀️22
likes
πŸ”5
retweets
D
drmapavone
@drmapavone
πŸ“…
Mar 17, 2026
5h ago
πŸ†”58775875

Jensen today announced Alpamayo 1.5 at #NVIDIAGTC! #Alpamayo 1.5 is a major update to Alpamayo 1β€”@nvidia’s open 10B-parameter chain-of-thought reasoning VLA model, first introduced at #CES. Built on the #Cosmos-Reason2 VLM backbone and post-trained with RL, it adds support for navigation guidance, flexible multi-camera setups, configurable camera parameters, and user question answering. The result is an interactive, steerable reasoning engine for the AV community. We’re also releasing post-training scripts to help researchers and developers adapt the model. Additionally, we’ve significantly expanded the Alpamayo open platform across data and simulation, including releasing highly requested reasoning labels for the PhysicalAI Autonomous Vehicles dataset (https://t.co/fD9eUcndya), as well as our chain-of-causation auto-labeling pipeline. πŸ”Ž Learn more about Alpamayo 1.5 and the latest extensions to the Alpamayo open platform: https://t.co/P0nuqkwBab (please note that most of the links will become active in the next few days.) Happy buildingβ€”and stay tuned for more in the coming months! @NVIDIADRIVE @NVIDIAAI

Media 2
+1 more
πŸ–ΌοΈ Media
L
lucas_flatwhite
@lucas_flatwhite
πŸ“…
Mar 17, 2026
5h ago
πŸ†”99053607

πŸ› οΈ Claude Code "opusplan" 말 κ·ΈλŒ€λ‘œ ν•˜μ΄λΈŒλ¦¬λ“œ λͺ¨λΈ.. κ³΅μ‹μž„! Claude Codeμ—λŠ” opusplan λͺ¨λΈμ„ 선택할 수 μžˆμ–΄μš”. > /model opusplan ν•˜μ΄λΈŒλ¦¬λ“œ λͺ¨λΈ alias인데, μž‘μ—… 단계에 따라 μžλ™μœΌλ‘œ λͺ¨λΈμ„ μ „ν™˜ν•΄μš”. λ³΅μž‘ν•œ 좔둠을 μœ„ν•œ ν”Œλžœ λͺ¨λ“œμ—μ„œλŠ” Opusλ₯Ό μ‹€ν–‰ λ‹¨κ³„μ—μ„œλŠ” Sonnet으둜 μžλ™ μ „ν™˜λ©λ‹ˆλ‹€! Opus둜 κ³„νšν•˜κ³  κ΅¬ν˜„κΉŒμ§€ ν•˜λŠ” 것도 λ¬Όλ‘  κ°€λŠ₯ν•΄μš”. ν•˜μ§€λ§Œ 이미 νƒ„νƒ„ν•œ κ³„νšμ΄ μžˆλ‹€λ©΄, 싀행은 SonnetμœΌλ‘œλ„ μΆ©λΆ„ν•˜κ³  더 μ €λ ΄ν•  수 μžˆμ–΄μš”. 각 μž‘μ—…μ— λ§žλŠ” λͺ¨λΈμ„ μ“°λŠ” 것 = 효율 πŸš€ ν”Œλž˜λ‹κ³Ό 싀행은 μš”κ΅¬λ˜λŠ” 인지 λΆ€ν•˜κ°€ λ‹¬λΌμš”. Opus의 κΉŠμ€ μΆ”λ‘  λŠ₯λ ₯은 κ³„νš 수립 λ‹¨κ³„μ—μ„œ κ°€μž₯ λΉ›λ‚˜κ³ , 일단 νƒ„νƒ„ν•œ κ³„νšμ΄ μ„Έμ›Œμ§„ μ΄ν›„μ˜ 싀행은 Sonnet으둜 μΆ©λΆ„νžˆ 컀버될 수 μžˆμ–΄μš”. μ–Έμ œ μ“°λ©΄ μ’‹λƒκ΅¬μš”? - λ³΅μž‘ν•œ κΈ°λŠ₯ 섀계같이 μ•„ν‚€ν…μ²˜ 결정이 μ€‘μš”ν•œ μž‘μ—… - λ¦¬νŒ©ν† λ§ κ³„νšκ°™μ€ 영ν–₯ λ²”μœ„ 뢄석이 ν•„μš”ν•œ 경우 - Opus ν’€νŒŒμ›Œ μ‚¬μš© λŒ€λΉ„ λΉ„μš© 절감이 ν•„μš”ν•  λ•Œ 이거 이제 많이 ν™œμš©ν•˜μ‹€ λ“―!!!

@dani_avila7 β€’ Mon Mar 16 15:58

Did you know about the opusplan model in Claude Code? /model opusplan It's a hybrid alias that automatically uses Opus in plan mode for complex reasoning, then switches to Sonnet for execution. Best of both worlds: Opus thinks, Sonnet builds https://t.co/r7un0X5bVg

πŸ–ΌοΈ Media
G
gdb
@gdb
πŸ“…
Mar 17, 2026
5h ago
πŸ†”37895367
⭐0.32

Subagents are now supported in Codex. They're very fun and make it possible to get large amounts of work done *quickly*:

@OpenAIDevs β€’ Mon Mar 16 20:09

Subagents are now available in Codex. You can accelerate your workflow by spinning up specialized agents to: β€’ Keep your main context window clean β€’ Tackle different parts of a task in parallel β€’ Steer individual agents as work unfolds https://t.co/QJC2ZYtYcA

πŸ”omarsar0 retweeted
O
elvis
@omarsar0
πŸ“…
Mar 16, 2026
19h ago
πŸ†”09077648
⭐0.38

Banger report from the Kimi team: Attention Residuals Residual connections made deep Transformers trainable. But they also force uncontrolled hidden-state growth with depth. This work proposes a cleaner alternative. It introduces Attention Residuals, which replace fixed residual accumulation with softmax attention over previous layer outputs. Instead of blindly summing everything, each layer selectively retrieves the earlier representations it actually needs. To keep this practical at scale, they add a blockwise version that compresses layers into block summaries, recovering most of the gains with minimal systems overhead. Why does it matter? Residual paths have barely changed across modern LLMs, even though they govern how information moves through depth. This paper shows that making the mixing content-dependent improves scaling laws, matches a baseline trained with 1.25x more compute, boosts GPQA-Diamond by +7.5 and HumanEval by +3.1, while keeping inference overhead under 2%. Paper: https://t.co/04IG6FDiVr Learn to build effective AI agents in our academy: https://t.co/1e8RZKs4uX

❀️130
likes
πŸ”17
retweets
πŸ”ivanleomk retweeted
Y
Yoeven
@yoeven
πŸ“…
Mar 17, 2026
8h ago
πŸ†”65291100

The moment he realised that https://t.co/vWmBsnR1nt isn't fully built on transformers and we can run on a single GPU with high accuracy and lower cost https://t.co/ZJYuL62UB8

Media 1
❀️4
likes
πŸ”1
retweets
πŸ–ΌοΈ Media
Y
yoeven
@yoeven
πŸ“…
Mar 17, 2026
8h ago
πŸ†”65291100

The moment he realised that https://t.co/vWmBsnR1nt isn't fully built on transformers and we can run on a single GPU with high accuracy and lower cost https://t.co/ZJYuL62UB8

Media 1Media 2
πŸ–ΌοΈ Media
πŸ”_akhaliq retweeted
H
Haocheng Xi
@HaochengXiUCB
πŸ“…
Mar 17, 2026
9h ago
πŸ†”24284251
⭐0.34

Thanks for sharing our newest work @_akhaliq ! Classic algorithms like K-Means deserve to be revisited in the era of massive datasets and GPUs. Flash-KMeans rethinks the algorithm from a systems perspective to make exact K-Means fast and memory-efficient on modern hardware.

❀️9
likes
πŸ”1
retweets
H
HamelHusain
@HamelHusain
πŸ“…
Mar 17, 2026
8h ago
πŸ†”89510900
⭐0.32

Claude Code CLI > Codex CLI Codex Desktop > Claude Code Desktop It’s a jagged UX frontier

H
HaochengXiUCB
@HaochengXiUCB
πŸ“…
Mar 17, 2026
9h ago
πŸ†”24284251
⭐0.38

Thanks for sharing our newest work @_akhaliq ! Classic algorithms like K-Means deserve to be revisited in the era of massive datasets and GPUs. Flash-KMeans rethinks the algorithm from a systems perspective to make exact K-Means fast and memory-efficient on modern hardware.

@_akhaliq β€’ Thu Mar 12 16:44

Flash-KMeans Fast and Memory-Efficient Exact K-Means paper: https://t.co/Yy7V7L12Bn https://t.co/c1mGipQl3f

C
code
@code
πŸ“…
Mar 17, 2026
9h ago
πŸ†”94910880

🌐 Agentic Browser Tools (Experimental) in @code! Agents can now open pages, read content, click elements, and verify changes directly in the integrated browser while building your web app. Enable βš™οΈ workbench.browser.enableChatTools to try it out. Learn mode: https://t.co/kNwugFcbIA

Media 2
πŸ–ΌοΈ Media
N
nvidianewsroom
@nvidianewsroom
πŸ“…
Mar 16, 2026
9h ago
πŸ†”99149451
⭐0.42

#NVIDIAGTC news: NVIDIA Dynamo 1.0 enters production as the broadly adopted inference operating system for AI factories. Dynamo 1.0 boosts Blackwell inference performance by up to 7x. The industry is scaling on NVIDIA. ⬇️https://t.co/Iaq2H2SmhR

P
PyTorch
@PyTorch
πŸ“…
Mar 16, 2026
10h ago
πŸ†”18333110

#ExecuTorch addresses fragmented native deployment for #AI agents as a #PyTorch native platform. It enables voice models across CPU, GPU, and NPU on Android, iOS, Linux, macOS & Windows πŸ”— https://t.co/NeQQyUniL4 https://t.co/O3itnoQFoG

Media 1
πŸ–ΌοΈ Media
G
gabriberton
@gabriberton
πŸ“…
Mar 16, 2026
10h ago
πŸ†”45334177

VisMatch is on pypi! VisMatch is a wrapper for image matching models, like LightGlue, RoMa-v2, MASt3R, LoFTR, and 50+ more! It's literally as simple as: pip install vismatch vismatch-match --inputs img0 img1 --matcher choose_any To run image matching on any 2 images [1/4] https://t.co/dIr2YapWak

Media 1
πŸ–ΌοΈ Media
πŸ”jxnlco retweeted
E
edwin
@edwinarbus
πŸ“…
Mar 16, 2026
14h ago
πŸ†”50334333
⭐0.34

Matt Maher tested frontier models in Cursor v. other harnesses. Cursor boosted model performance by 11% on average: Gemini: 52% β†’ 57% GPT-5.4: 82% β†’ 88% Opus: 77% β†’ 93% His benchmark measures how well models implement a 100-feature PRD. @cursor_ai consistently outperformed. https://t.co/hrjCmWMNKN

❀️176
likes
πŸ”17
retweets
Z
ZGojcic
@ZGojcic
πŸ“…
Mar 16, 2026
11h ago
πŸ†”28929828
⭐0.40

A new generation in AV simulation is here! We are announcing AlpaDreams, a real time interactive generative world model for AV simualtion! Just a year ago it took minutes to generate a few seconds of video, today it is real time and interactive! https://t.co/FbhKu3PMqe

R
RayFernando1337
@RayFernando1337
πŸ“…
Mar 16, 2026
11h ago
πŸ†”10226271
⭐0.34

Nvidia GTC 2026 OpenClaw Setup on DGX Spark IRL https://t.co/zQwwfCF9XP

_
_akhaliq
@_akhaliq
πŸ“…
Mar 16, 2026
11h ago
πŸ†”76800000

Mistral Small 4 is out https://t.co/IdAowSpHpN

Media 1
πŸ–ΌοΈ Media
J
jasteinerman
@jasteinerman
πŸ“…
Mar 16, 2026
12h ago
πŸ†”75976987
⭐0.32

Love this submission from our world models hackathon this weekend - a generative FPS!

@AnshulDhawan001 β€’ Mon Mar 16 21:08

Spent the weekend hacking at the Worlds in Action hackathon at @fdotinc by @SensAIHackademy. It was so much fun playing with the world models by @theworldlabs . I believe generative games are the future where characters, rules and even parts of the world can be generated and ad

A
ArtificialAnlys
@ArtificialAnlys
πŸ“…
Mar 16, 2026
13h ago
πŸ†”52868861

NVIDIA has released Nemotron 3 VoiceChat! A ~12B parameter Speech to Speech model that leads our open weights Conversational Dynamics vs. Speech Reasoning pareto frontier Understanding Speech to Speech model performance is multidimensional - two key and distinct dimensions are raw intelligence and conversational dynamics: how well a model handles the natural rhythms of human conversation such as turn-taking, interruptions. Amongst full duplex open weights models, NVIDIA’s new Nemotron 3 VoiceChat, V1, leads in balancing these dimensions, setting itself apart from other models on the Conversational Dynamics vs. Speech Reasoning pareto frontier. Key benchmarking results: ➀ Conversational Dynamics (Full Duplex Bench): Nemotron 3 VoiceChat (V1) scores 77.8%, second among open weights speech to speech models behind NVIDIA's own PersonaPlex (91.0%) and ahead of FLM-Audio (62.0%), Moshi (61.0%) and Freeze-Omni (58.7%) ➀ Speech Reasoning (Big Bench Audio): Nemotron 3 VoiceChat (V1) scores 29.2%, second among open weights speech to speech models behind Freeze-Omni (33.9%) and well ahead of PersonaPlex (12.6%), FLM-Audio (5.3%) and Moshi (1.7%) ➀ Pareto leader: While Freeze-Omni leads on speech reasoning and PersonaPlex leads on conversational dynamics, Nemotron 3 VoiceChat (V1) is the only open weights model that performs amongst the top 3 on both - making it the clear leader on the pareto frontier between these two critical dimensions ➀ Larger than other open weights models but still relatively small compared to LLMs: Nemotron 3 VoiceChat (V1) has 12B parameters, making it one of the larger open weights speech to speech models, while NVIDIA's PersonaPlex is ~7B. While larger compared to other larger open weights speech to speech models the model still is relatively small compared to leading LLMs ➀ Context vs. proprietary models: While this release materially advances open weights performance, open weights speech to speech models still significantly underperform leading proprietary offerings. For comparison, proprietary models on our Big Bench Audio benchmark score substantially higher - Step-Audio R1.1 at 96%, Grok Voice Agent at 92%, Gemini 2.5 Flash (Thinking) at 92%, and Nova 2.0 Sonic at 87%. The gap between open weights and proprietary remains large in this modality. As the capability and adoption of Speech to Speech models increases, we expect to expand our set of benchmarks to include elements such as tool-calling and multi-turn instruction following. See more details below ⬇️

Media 1
πŸ–ΌοΈ Media
πŸ”jxnlco retweeted
O
OpenAI Developers
@OpenAIDevs
πŸ“…
Mar 16, 2026
13h ago
πŸ†”48174967
⭐0.34

Subagents are now available in Codex. You can accelerate your workflow by spinning up specialized agents to: β€’ Keep your main context window clean β€’ Tackle different parts of a task in parallel β€’ Steer individual agents as work unfolds https://t.co/QJC2ZYtYcA

❀️798
likes
πŸ”74
retweets
πŸ”jeremyphoward retweeted
R
raia hadsell
@RaiaHadsell
πŸ“…
Mar 16, 2026
15h ago
πŸ†”56989392
⭐0.36

It's been about 20 years since I first started working on embeddings with Yann LeCun (siamese networks!), and I've been fascinated ever since. Gemini Embeddings 2 approaches the platonic ideal: native embedding of text, image, video, audio, and docs to a single space.

❀️277
likes
πŸ”24
retweets
H
HuggingPapers
@HuggingPapers
πŸ“…
Mar 16, 2026
13h ago
πŸ†”83694046

OmniForcing unlocks real-time joint audio-visual generation Achieves ~25 FPS with 0.7s latencyβ€”a 35Γ— speedup over offline diffusion modelsβ€”by distilling bidirectional LTX-2 into a causal streaming generator with maintained multi-modal fidelity. https://t.co/UGYGMyTQOs

Media 1
πŸ–ΌοΈ Media
P
PyTorch
@PyTorch
πŸ“…
Mar 16, 2026
13h ago
πŸ†”07617111
⭐0.38

@Nvidiadev πŸ—“οΈ MONDAY @ Booth #338 2PM: Shaping the Future w/ @matthew_d_white 3PM: TensorRT + PyTorch w/ Angela Yi & @narendasan 4PM: DeepSpeed Trillion-Param Training w/ @PKUWZP 5PM: PyTorch Export w/ Angela Yi 6PM: Ray Distributed Computing w/ @robertnishihara #AI #GTC2025

O
OpenAIDevs
@OpenAIDevs
πŸ“…
Mar 16, 2026
13h ago
πŸ†”48174967

Subagents are now available in Codex. You can accelerate your workflow by spinning up specialized agents to: β€’ Keep your main context window clean β€’ Tackle different parts of a task in parallel β€’ Steer individual agents as work unfolds https://t.co/QJC2ZYtYcA

πŸ–ΌοΈ Media
E
edwinarbus
@edwinarbus
πŸ“…
Mar 16, 2026
14h ago
πŸ†”50334333
⭐0.44

Matt Maher tested frontier models in Cursor v. other harnesses. Cursor boosted model performance by 11% on average: Gemini: 52% β†’ 57% GPT-5.4: 82% β†’ 88% Opus: 77% β†’ 93% His benchmark measures how well models implement a 100-feature PRD. @cursor_ai consistently outperformed. https://t.co/hrjCmWMNKN

πŸ”ai_fast_track retweeted
T
David Hendrickson
@TeksEdge
πŸ“…
Mar 14, 2026
2d ago
πŸ†”30554364
⭐0.34

🚨 Want to parse complex PDFs with SOTA accuracy, 100% locally? πŸ“„πŸ” At just 0.9B parameters, you can drop GLM-OCR straight into LM Studio and run it on almost any machine! πŸ₯” 🧠 0.9B total parameters πŸ’Ύ Runs on < 1.5GB VRAM (or ~1GB quantized!) πŸ’Έ Zero API costs πŸ”’ Total data privacy Desktop document AI is officially here. πŸ’»βš‘

❀️2,365
likes
πŸ”218
retweets
πŸ”ai_fast_track retweeted
A
Adina Yakup
@AdinaYakup
πŸ“…
Mar 16, 2026
20h ago
πŸ†”41999406

Covo Audio πŸ”ŠA end-to-end audio language model from @TencentAI_News https://t.co/tic5cH1A39 ✨ 7B ✨ Audio β†’ Audio in one model ✨ Multi-speaker + voice transfer ✨ Real-time full duplex conversations https://t.co/hFrsxQgzkT

Media 1
❀️77
likes
πŸ”11
retweets
πŸ–ΌοΈ Media
A
alex_peys
@alex_peys
πŸ“…
Mar 16, 2026
15h ago
πŸ†”51888850
⭐0.40

this was one of the things i co-led at fair, then fb had ~2b users, embeddings of ~128d made it a 300b-1T parameter model depending on how you count entities (e.g. ad campaigns). at the time, this was big, now it's medium. we trained it purely on distributed cpus

@ylecun β€’ Mon Mar 16 18:09

@RaiaHadsell Universal embeddings FTW 😊 One of the flagship projects at FAIR was to "embed the world" (i.e. represent every entity on Facebook). The name was soon changed to "Filament", deployed internally, and eventually open-sourced as "PyTorch-BigGraph" The techniques were m

πŸ”ylecun retweeted
A
alphaXiv
@askalphaxiv
πŸ“…
Mar 16, 2026
1d ago
πŸ†”49397718
⭐0.36

Yann LeCun is pumping out papers recently β€œTemporal Straightening for Latent Planning” This paper shows that by straightening latent trajectories in a world model, Euclidean distance starts to reflect true reachable progress, so it's closer to geodesic/minimum-step distance. This makes gradient-based planning far more stable and effective without relying as heavily on expensive search.

❀️702
likes
πŸ”115
retweets