Your curated collection of saved posts and media

Showing 20 posts ยท last 7 days ยท quality filtered
S
Sanemavcil
@Sanemavcil
๐Ÿ“…
Mar 15, 2026
9m ago
๐Ÿ†”59686729

The first humanoids in our homes might not look human at all. They might just be vacuum robots that slowly evolve arms, tools, and intelligence. Evolution, but in hardware. via Peter Kappes #Ai #robotics #innovation https://t.co/6AbmG6WPOY

๐Ÿ–ผ๏ธ Media
๐Ÿ”MFordFuture retweeted
M
Martin Ford
@MFordFuture
๐Ÿ“…
Mar 10, 2026
4d ago
๐Ÿ†”93930895
โญ0.32

I'm pleased to announce that a new 2026 edition of my New York Times bestselling book, Rise of the Robots: Technology and the Threat of a Jobless Future, will be available on June 2. I have extensively updated the book to cover the latest advances in generative #AI and robotics and to examine the future economic and job market implications of the unfolding AI disruption. The book focuses on what we can do as individuals, and as a society, to successfully navigate the looming transition into the age of AI. You can pre-order from the link in the reply. @BasicBooks #RiseoftheRobots

โค๏ธ358
likes
๐Ÿ”68
retweets
M
Mid0
@Mid0
๐Ÿ“…
Mar 15, 2026
55m ago
๐Ÿ†”10757761
โญ0.38

@Bhavani_00007 Trick question there is no windsurf. I vote for Augment + Codex + Claude code. I rarely use Gemini CLI but I like it for research, test/mock data creation. All three. VS Code just for reading. Previously it supported other

O
omarsar0
@omarsar0
๐Ÿ“…
Mar 15, 2026
57m ago
๐Ÿ†”66759535

15K stars already!? Great idea. CLIs work amazingly well with coding agents. Worth playing around with. Do run a lot of tests if you are planning to use this to build tools. https://t.co/Aigh3uAI5Y

Media 1
๐Ÿ–ผ๏ธ Media
๐Ÿ”dair_ai retweeted
O
elvis
@omarsar0
๐Ÿ“…
Mar 15, 2026
57m ago
๐Ÿ†”66759535
โญ0.32

15K stars already!? Great idea. CLIs work amazingly well with coding agents. Worth playing around with. Do run a lot of tests if you are planning to use this to build tools. https://t.co/Aigh3uAI5Y

โค๏ธ8
likes
๐Ÿ”2
retweets
H
HuaxiuYaoML
@HuaxiuYaoML
๐Ÿ“…
Mar 15, 2026
12h ago
๐Ÿ†”53405308

Everyone's excited about Karpathy's autoresearch that automates the experiment loop. We automated the whole damn thing. ๐Ÿฆž Meet AutoResearchClaw: one message in, full conference paper out. Real experiments. Real citations. Real code. No human in the loop. One message in โ†’ full paper out. Here's what happens in between: ๐Ÿ“š Raids arXiv & Semantic Scholar, digests 50+ papers in minutes ๐ŸฅŠ Three AI agents FIGHT over the best hypothesis (one swings big, one sanity-checks, one tries to kill every idea) ๐Ÿ’ป Writes experiment code from scratch, adapts to your hardware ๐Ÿ’ฅ Code crashes at 3am? It reads the stack trace, rewrites the fix, keeps going ๐Ÿ”„ Results weak? It pivots to entirely new hypotheses and starts over ๐Ÿ“ Drafts a full paper with citations, every single one verified against live databases No babysitting. No Slack messages. No "hey can you re-run this." Karpathy built the experiment loop. We built the whole lab. Chat an idea. Get a paper. ๐Ÿฆž Try it ๐Ÿ‘‰: https://t.co/KLOcnzFYaD Kudos to the team @JiaqiLiu835914, @richardxp888, @lillianwei423, @StephenQS0710, @Xinyu2ML, @HaoqinT, @zhengop, @cihangxie, @dingmyu, and we are looking for more contributors.

Media 1Media 2
๐Ÿ–ผ๏ธ Media
G
gerardsans
@gerardsans
๐Ÿ“…
Mar 15, 2026
2h ago
๐Ÿ†”52086940
โญ0.36

@eigenron It has the same problems RL ran into, maybe worse. Collapsing branches to win benchmarks doesnโ€™t improve real capability. It mostly compresses output variance; by shifting the weights, it distorts the latent space and hurts performance elsewhere. Careful what you optimize for. Benchmaxxing isnโ€™t the path forward.

S
shao__meng
@shao__meng
๐Ÿ“…
Mar 15, 2026
4h ago
๐Ÿ†”73171514

ไปŠๅคฉๆญฃๅผๅ‡†ๅค‡ไปŽ Claude Code ๅˆ‡ๆขๅˆฐ Codex ไบ† ไน‹ๅ‰็”จ Claude Code ๆ—ถๅ› ไธบๆฒกๆœ‰ Anthropic ๅฎ˜ๆ–น API๏ผŒไธ€็›ดๅœจ็”จ Minimax ๅ’Œ Kimi ็ญ‰ API ๅˆ‡ๆข็€็”จใ€‚ ๆœ€่ฟ‘่‚‰็œผๅฏ่ง @OpenAIDevs ๅœจ Codex ไธŠ็š„ๅ†ณๅฟƒๅ’ŒๅŠจไฝœ่ถŠๆฅ่ถŠๅฏ†้›†๏ผŒOpenClaw ๅˆ›ๅง‹ไบบ @steipeteใ€Instructor ไฝœ่€… @jxnlco ็ญ‰ๅผ€ๆบๅ’Œ AI ๆ•™่‚ฒๅˆ†ไบซ้žๅธธๆดป่ทƒ็š„ๅคงไฝฌๅŠ ๅ…ฅ Codex๏ผŒ่ฟ˜ๆœ‰ไธๅฎšๆœŸ Reset limit ็š„ @thsottiaux ๐Ÿ˜„ ๅ…ˆ่ฎข้˜…ไธช Plus ไผšๅ‘˜ไฝœไธบไธปๅŠ› AI ็”จ่ตทๆฅ๏ผๅฏน Codex ๆŒ‡ไปคไธๅคŸ็†Ÿๆ‚‰๏ผŒๅ…ˆๅšไธช Cheatsheet ็ป™ๅˆšๅˆšไบ†่งฃ Codex ็š„ๆœ‹ๅ‹ไปฌ๏ผŒๅŒ…ๆ‹ฌๆˆ‘่‡ชๅทฑใ€‚

Media 1
๐Ÿ–ผ๏ธ Media
๐Ÿ”jxnlco retweeted
S
meng shao
@shao__meng
๐Ÿ“…
Mar 15, 2026
4h ago
๐Ÿ†”73171514
โญ0.36

ไปŠๅคฉๆญฃๅผๅ‡†ๅค‡ไปŽ Claude Code ๅˆ‡ๆขๅˆฐ Codex ไบ† ไน‹ๅ‰็”จ Claude Code ๆ—ถๅ› ไธบๆฒกๆœ‰ Anthropic ๅฎ˜ๆ–น API๏ผŒไธ€็›ดๅœจ็”จ Minimax ๅ’Œ Kimi ็ญ‰ API ๅˆ‡ๆข็€็”จใ€‚ ๆœ€่ฟ‘่‚‰็œผๅฏ่ง @OpenAIDevs ๅœจ Codex ไธŠ็š„ๅ†ณๅฟƒๅ’ŒๅŠจไฝœ่ถŠๆฅ่ถŠๅฏ†้›†๏ผŒOpenClaw ๅˆ›ๅง‹ไบบ @steipeteใ€Instructor ไฝœ่€… @jxnlco ็ญ‰ๅผ€ๆบๅ’Œ AI ๆ•™่‚ฒๅˆ†ไบซ้žๅธธๆดป่ทƒ็š„ๅคงไฝฌๅŠ ๅ…ฅ Codex๏ผŒ่ฟ˜ๆœ‰ไธๅฎšๆœŸ Reset limit ็š„ @thsottiaux ๐Ÿ˜„ ๅ…ˆ่ฎข้˜…ไธช Plus ไผšๅ‘˜ไฝœไธบไธปๅŠ› AI ็”จ่ตทๆฅ๏ผๅฏน Codex ๆŒ‡ไปคไธๅคŸ็†Ÿๆ‚‰๏ผŒๅ…ˆๅšไธช Cheatsheet ็ป™ๅˆšๅˆšไบ†่งฃ Codex ็š„ๆœ‹ๅ‹ไปฌ๏ผŒๅŒ…ๆ‹ฌๆˆ‘่‡ชๅทฑใ€‚

โค๏ธ48
likes
๐Ÿ”7
retweets
H
HuggingPapers
@HuggingPapers
๐Ÿ“…
Mar 15, 2026
2h ago
๐Ÿ†”83475344

Top AI papers on @huggingface this week: Language feedback for RL, training agents by talking, and fixing LLM story consistency - Bootstrapping Exploration with Group-Level Natural Language Feedback in Reinforcement Learning - Geometry-Guided Reinforcement Learning for Multi-view Consistent 3D Scene Editing - Penguin-VL by Tencent: Exploring the Efficiency Limits of VLM with LLM-based Vision Encoders - OpenClaw-RL: Train Any Agent Simply by Talking - Lost in Stories: Consistency Bugs in Long Story Generation by LLMs - Holi-Spatial: Evolving Video Streams into Holistic 3D Spatial Intelligence - Spatial-TTT: Streaming Visual-based Spatial Intelligence with Test-Time Training - Flash-KMeans: Fast and Memory-Efficient Exact K-Means - Thinking to Recall: How Reasoning Unlocks Parametric Knowledge in LLMs - LoGeR: Long-Context Geometric Reconstruction with Hybrid Memory

Media 1
๐Ÿ–ผ๏ธ Media
๐Ÿ”_akhaliq retweeted
H
DailyPapers
@HuggingPapers
๐Ÿ“…
Mar 15, 2026
2h ago
๐Ÿ†”83475344
โญ0.38

Top AI papers on @huggingface this week: Language feedback for RL, training agents by talking, and fixing LLM story consistency - Bootstrapping Exploration with Group-Level Natural Language Feedback in Reinforcement Learning - Geometry-Guided Reinforcement Learning for Multi-view Consistent 3D Scene Editing - Penguin-VL by Tencent: Exploring the Efficiency Limits of VLM with LLM-based Vision Encoders - OpenClaw-RL: Train Any Agent Simply by Talking - Lost in Stories: Consistency Bugs in Long Story Generation by LLMs - Holi-Spatial: Evolving Video Streams into Holistic 3D Spatial Intelligence - Spatial-TTT: Streaming Visual-based Spatial Intelligence with Test-Time Training - Flash-KMeans: Fast and Memory-Efficient Exact K-Means - Thinking to Recall: How Reasoning Unlocks Parametric Knowledge in LLMs - LoGeR: Long-Context Geometric Reconstruction with Hybrid Memory

โค๏ธ3
likes
๐Ÿ”3
retweets
Z
zhuokaiz
@zhuokaiz
๐Ÿ“…
Mar 15, 2026
11h ago
๐Ÿ†”07654255
โญ0.46

Latent world models learn differentiable dynamics in a learned representation space, which should make planning as simple as gradient descent. But it almost never works. What I mean is, at test time, you can treat the action sequence as learnable parameters, roll out the frozen world model, measure how far the predicted final state is from the goal, and backprop through the entire unrolled chain to optimize actions directly. Yet many of the systems that work (Dreamer, TD-MPC2, DINO-WM) abandon this and fall back to sampling-based search instead. That's why I really like this new paper by @yingwww_, @ylecun, and @mengyer, which gives a clean diagnosis of why, and a principled fix. The reason everyone abandons gradient descent on actions is that the planning objective is highly non-convex in the learned latent space. So instead most systems use CEM (cross-entropy method) or MPPI (model predictive path integral), both derivative-free. CEM samples batches of action sequences, evaluates them by rolling out the world model, keeps the top-k, and refits the sampling distribution. MPPI does something similar but weights trajectories by exponentiated negative cost instead of hard elite selection. These work when gradients are unreliable but the compute cost is substantial โ€” hundreds of candidate rollouts per planning step vs a single forward-backward pass. This paper asks what exactly makes the latent planning landscape so hostile to gradients and what you can do about it. The diagnosis. Their baseline is DINO-WM, a JEPA-style world model with a ViT predictor planning in frozen DINOv2 feature space, minimizing terminal MSE between predicted and goal embeddings. The problem is that DINOv2 latent trajectories are highly curved (when you use MSE as the planning cost you're implicitly assuming euclidean distance approximates geodesic distance along feasible transitions). For curved trajectories this breaks badly, gradient-based planners get trapped and straight-line distances in embedding space misrepresent actual reachability. The fix draws from the perceptual straightening hypothesis in neuroscience โ€” the idea that biological visual systems transform complex video into internally straighter representations. So they add a curvature regularizer during world model training. Given consecutive encoded states z_t, z_{t+1}, z_{t+2}, define velocity vectors as v_t = z_{t+1} - z_t measure curvature as the cosine similarity between consecutive velocities, and minimize L_curv = 1 - cos(v_t, v_{t+1}). Total loss is then L_pred + ฮป * L_curv with stop-gradient on the target branch to prevent collapse. The theory backs this up cleanly โ€” they prove that reducing curvature directly bounds how well-conditioned the planning optimization is โ€” straighter latent trajectories guarantee faster convergence of gradient descent over longer horizons. Worth noting that even without the curvature loss, training the encoder with a prediction objective alone produces some "implicit straightening" โ€” the JEPA loss naturally favors representations whose temporal evolution is predictable. Explicit regularization simply pushes this much further. Empirical results across four 2D goal-reaching environments are consistently strong. Open-loop success improves by 20-50%, and the GD with straightening matches or beats CEM at a fraction of the compute. The most convincing evidence is the distance heatmaps: after straightening, latent Euclidean distance closely matches the shortest distance between states, even though the model was trained only on suboptimal random trajectories. What I find interesting beyond the specific method is that the planning algorithm didn't change. The dynamics model didn't change. A single regularization term on the embedding geometry turned gradient descent from unreliable to competitive with sampling methods. The field has largely treated representation learning and planning as separate concerns โ€” learn good features, then figure out how to plan in them. This paper makes a concrete case that the representation geometry is itself the bottleneck. This connects to a broader pattern in ML. When optimization fails, the instinct is to fix the optimizer (better search, more samples, adaptive schedules). But often the real lever is the shape of the space you're optimizing in. Same principle shows up in RL post-training where reward landscape shaping matters as much as the algorithm itself. Shape the space so simple optimization works, rather than building complex optimization to handle a bad space. Their paper: https://t.co/NLPGxqbP2x

๐Ÿ”ylecun retweeted
Z
Zhuokai Zhao
@zhuokaiz
๐Ÿ“…
Mar 15, 2026
11h ago
๐Ÿ†”07654255
โญ0.36

Latent world models learn differentiable dynamics in a learned representation space, which should make planning as simple as gradient descent. But it almost never works. What I mean is, at test time, you can treat the action sequence as learnable parameters, roll out the frozen world model, measure how far the predicted final state is from the goal, and backprop through the entire unrolled chain to optimize actions directly. Yet many of the systems that work (Dreamer, TD-MPC2, DINO-WM) abandon this and fall back to sampling-based search instead. That's why I really like this new paper by @yingwww_, @ylecun, and @mengyer, which gives a clean diagnosis of why, and a principled fix. The reason everyone abandons gradient descent on actions is that the planning objective is highly non-convex in the learned latent space. So instead most systems use CEM (cross-entropy method) or MPPI (model predictive path integral), both derivative-free. CEM samples batches of action sequences, evaluates them by rolling out the world model, keeps the top-k, and refits the sampling distribution. MPPI does something similar but weights trajectories by exponentiated negative cost instead of hard elite selection. These work when gradients are unreliable but the compute cost is substantial โ€” hundreds of candidate rollouts per planning step vs a single forward-backward pass. This paper asks what exactly makes the latent planning landscape so hostile to gradients and what you can do about it. The diagnosis. Their baseline is DINO-WM, a JEPA-style world model with a ViT predictor planning in frozen DINOv2 feature space, minimizing terminal MSE between predicted and goal embeddings. The problem is that DINOv2 latent trajectories are highly curved (when you use MSE as the planning cost you're implicitly assuming euclidean distance approximates geodesic distance along feasible transitions). For curved trajectories this breaks badly, gradient-based planners get trapped and straight-line distances in embedding space misrepresent actual reachability. The fix draws from the perceptual straightening hypothesis in neuroscience โ€” the idea that biological visual systems transform complex video into internally straighter representations. So they add a curvature regularizer during world model training. Given consecutive encoded states z_t, z_{t+1}, z_{t+2}, define velocity vectors as v_t = z_{t+1} - z_t measure curvature as the cosine similarity between consecutive velocities, and minimize L_curv = 1 - cos(v_t, v_{t+1}). Total loss is then L_pred + ฮป * L_curv with stop-gradient on the target branch to prevent collapse. The theory backs this up cleanly โ€” they prove that reducing curvature directly bounds how well-conditioned the planning optimization is โ€” straighter latent trajectories guarantee faster convergence of gradient descent over longer horizons. Worth noting that even without the curvature loss, training the encoder with a prediction objective alone produces some "implicit straightening" โ€” the JEPA loss naturally favors representations whose temporal evolution is predictable. Explicit regularization simply pushes this much further. Empirical results across four 2D goal-reaching environments are consistently strong. Open-loop success improves by 20-50%, and the GD with straightening matches or beats CEM at a fraction of the compute. The most convincing evidence is the distance heatmaps: after straightening, latent Euclidean distance closely matches the shortest distance between states, even though the model was trained only on suboptimal random trajectories. What I find interesting beyond the specific method is that the planning algorithm didn't change. The dynamics model didn't change. A single regularization term on the embedding geometry turned gradient descent from unreliable to competitive with sampling methods. The field has largely treated representation learning and planning as separate concerns โ€” learn good features, then figure out how to plan in them. This paper makes a concrete case that the representation geometry is itself the bottleneck. This connects to a broader pattern in ML. When optimization fails, the instinct is to fix the optimizer (better search, more samples, adaptive schedules). But often the real lever is the shape of the space you're optimizing in. Same principle shows up in RL post-training where reward landscape shaping matters as much as the algorithm itself. Shape the space so simple optimization works, rather than building complex optimization to handle a bad space. Their paper: https://t.co/NLPGxqbP2x

โค๏ธ132
likes
๐Ÿ”15
retweets
Y
ylecun
@ylecun
๐Ÿ“…
Mar 15, 2026
3h ago
๐Ÿ†”45027353
โญ0.32

@zhuokaiz Nice summary ๐Ÿ˜Š

๐Ÿ”omarsar0 retweeted
D
DAIR.AI
@dair_ai
๐Ÿ“…
Mar 15, 2026
3h ago
๐Ÿ†”04105379
โญ0.38

The Top AI Papers of the Week (March 9 - March 15) - KARL - OpenDev - SkillNet - Memex(RL) - AutoHarness - FlashAttention-4 - The Spike, the Sparse, and the Sink Read on for more:

โค๏ธ7
likes
๐Ÿ”1
retweets
D
dair_ai
@dair_ai
๐Ÿ“…
Mar 15, 2026
3h ago
๐Ÿ†”37608283
โญ0.34

https://t.co/0lQ8NXM4M3

D
dair_ai
@dair_ai
๐Ÿ“…
Mar 15, 2026
3h ago
๐Ÿ†”04105379
โญ0.38

The Top AI Papers of the Week (March 9 - March 15) - KARL - OpenDev - SkillNet - Memex(RL) - AutoHarness - FlashAttention-4 - The Spike, the Sparse, and the Sink Read on for more:

S
strickvl
@strickvl
๐Ÿ“…
Mar 15, 2026
4h ago
๐Ÿ†”70282353
โญ0.42

I've been building panlabel โ€” a fast Rust CLI that converts between dataset annotation formats โ€” and I'm a few releases behind on sharing updates. Here's a quick catch-up. v0.3.0 added Hugging Face ImageFolder support, including remote Hub import via --hf-repo. You can point it at a HF dataset repo and it figures out the layout (metadata.jsonl, parquet shards, even zip-style splits that contain YOLO or COCO inside). v0.4.0 overhauled auto-detection so it gives you concrete evidence when format detection is ambiguous ("found YOLO labels/ but missing images/") instead of a generic error. Also added Docker images. v0.5.0 brought split-aware YOLO reading for Roboflow/Ultralytics Hub exports and conversion report explainability โ€” every adapter now explains its deterministic policies so you know exactly what happens to your data. v0.6.0 is the big one. Five new format adapters: โ†’ LabelMe JSON (per-image, with polygon-to-bbox envelope) โ†’ Apple CreateML JSON (center-based coords) โ†’ KITTI (autonomous driving standard โ€” 15 fields per line) โ†’ VGG Image Annotator (VIA) JSON โ†’ RetinaNet Keras CSV That brings panlabel to 13 supported formats with full read, write, and auto-detection. Also in v0.6.0: YOLO confidence token support, dry-run mode for previewing conversions, and content-based CSV detection. Single binary, no Python dependencies. Install via pip, brew, cargo, or grab a pre-built binary from GitHub releases. This is the kind of project I enjoy just steadily plodding away at โ€” ticking off one format at a time until every common object detection annotation format is covered. Still sticking with detection bboxes for now, but the format list keeps growing. #ObjectDetection #Rust #MachineLearning #ComputerVision #OpenSource

๐Ÿ”dicksonneoh7 retweeted
S
Alex Strick van Linschoten
@strickvl
๐Ÿ“…
Mar 15, 2026
4h ago
๐Ÿ†”70282353
โญ0.34

I've been building panlabel โ€” a fast Rust CLI that converts between dataset annotation formats โ€” and I'm a few releases behind on sharing updates. Here's a quick catch-up. v0.3.0 added Hugging Face ImageFolder support, including remote Hub import via --hf-repo. You can point it at a HF dataset repo and it figures out the layout (metadata.jsonl, parquet shards, even zip-style splits that contain YOLO or COCO inside). v0.4.0 overhauled auto-detection so it gives you concrete evidence when format detection is ambiguous ("found YOLO labels/ but missing images/") instead of a generic error. Also added Docker images. v0.5.0 brought split-aware YOLO reading for Roboflow/Ultralytics Hub exports and conversion report explainability โ€” every adapter now explains its deterministic policies so you know exactly what happens to your data. v0.6.0 is the big one. Five new format adapters: โ†’ LabelMe JSON (per-image, with polygon-to-bbox envelope) โ†’ Apple CreateML JSON (center-based coords) โ†’ KITTI (autonomous driving standard โ€” 15 fields per line) โ†’ VGG Image Annotator (VIA) JSON โ†’ RetinaNet Keras CSV That brings panlabel to 13 supported formats with full read, write, and auto-detection. Also in v0.6.0: YOLO confidence token support, dry-run mode for previewing conversions, and content-based CSV detection. Single binary, no Python dependencies. Install via pip, brew, cargo, or grab a pre-built binary from GitHub releases. This is the kind of project I enjoy just steadily plodding away at โ€” ticking off one format at a time until every common object detection annotation format is covered. Still sticking with detection bboxes for now, but the format list keeps growing. #ObjectDetection #Rust #MachineLearning #ComputerVision #OpenSource

โค๏ธ1
likes
๐Ÿ”1
retweets
R
rasbt
@rasbt
๐Ÿ“…
Mar 15, 2026
4h ago
๐Ÿ†”02210058

I (finally) put together a new LLM Architecture Gallery that collects the architecture figures all in one place! https://t.co/NO7z6XSRHS https://t.co/X41FrK4i94

Media 1
๐Ÿ–ผ๏ธ Media