Your curated collection of saved posts and media

Showing 10 posts Β· last 14 days Β· by score
βž• Add New Post
R
RedHat_AI
@RedHat_AI
πŸ“…
Apr 10, 2026
13d ago
πŸ†”97110649

Speculative decoding for Gemma 4 31B (EAGLE-3) A 2B draft model predicts tokens ahead; the 31B verifier validates them. Same output, faster inference. Early release. vLLM main branch support is in progress (PR #39450). Reasoning support coming soon. https://t.co/PoK8zbA7li

Media 1
πŸ–ΌοΈ Media
πŸ”jeremyphoward retweeted
G
Google for Developers
@googledevs
πŸ“…
Apr 07, 2026
16d ago
πŸ†”60211334
⭐0.32

A new PyTorch-native backend is coming to unlock the power of Google TPUs: ✨ Run existing PyTorch with minimal code changes. ✨ Get a 50-100%+ performance boost with Fused Eager mode. Read the engineering deep dive here: https://t.co/GQPRYaKz7E #TorchTPU #PyTorch #MLOps #AI https://t.co/HiIdXVw6Oh

❀️762
likes
πŸ”118
retweets
T
thsottiaux
@thsottiaux
πŸ“…
Apr 16, 2026
7d ago
πŸ†”73879269

Codex just got a lot more powerful. Computer use, in-app browser, image generation and editing, 90+ new plugins to connect to everything, multi-terminal, SSH into devboxes, thread automations, rich document editing. Learns from experience and proactively suggestions work. And a ton more.

Media 1
πŸ–ΌοΈ Media
_
_lewtun
@_lewtun
πŸ“…
Apr 12, 2026
11d ago
πŸ†”52359638

We've rebuilt TRL's on-policy distillation trainer from the ground up to: 🐳 support huge teachers with 100B+ params ⚑️ train >40x faster thanks to some nifty buffer and payload optimisations This means you can now distill models in the Llama, Qwen and Gemma families across any scale! Technical deep dive with all the optimisations and pretty animations ⬇️

πŸ–ΌοΈ Media
πŸ”jxnlco retweeted
A
Akshat Bubna
@akshat_b
πŸ“…
Apr 15, 2026
8d ago
πŸ†”11880169
⭐0.38

To show off what you can do with @OpenAI Agent SDK + @modal, we built an ML research agent (inspired by @karpathy). It can: - Spin up GPU sandboxes of any shape - Run a pool of subagents - Persist memory - Snapshot state for fork/resume Here it is playing Parameter Golf: https://t.co/r7QhvNmdEq

❀️14
likes
πŸ”2
retweets
Y
yazins
@yazins
πŸ“…
Apr 15, 2026
8d ago
πŸ†”20724190
⭐0.38

ok i've barely used cursor the past 18 months.. but today, i pulled it up and asked it to implement one of the more ambitious projects i had in mind .. something that really pushes it, y'know? the project was a nextjs web app, with a chrome extension + tauri desktop app (mac/win/unix) spent 10 mins working on a spec for the work, and left the thing to run -- checking in at key points to make sure everything worked. all told, it worked for like 3 hours.. and in the end, i had a suite of fully working apps! and the crazy bit is ... i don't even know which model it used?! i just picked "Premium" > "Max mode" (i have $60k in credits to burn and 6 months to do so) needless to say, i'm going to be using cursor alot more from now..

@cursor_ai β€’ Thu Apr 02 18:14

We’re introducing Cursor 3. It is simpler, more powerful, and built for a world where all code is written by agents, while keeping the depth of a development environment. https://t.co/rXR9vaZDnO

A
arena
@arena
πŸ“…
Apr 15, 2026
9d ago
πŸ†”17667644
⭐0.36

New video model HappyHorse-1.0 by Alibaba-ATH debuts at #1 in Video Edit Arena. It scores 1299, leading Grok Image Video by +42 points and Kling o3 Pro by +48 points. Video editing is an emerging frontier capability for video models, and only a small number of models support it today. Huge congrats to the Alibaba-ATH team on this incredible milestone!

@HappyHorseATH β€’ Wed Apr 15 03:25

HappyHorse-1.0 is now live on Arena! πŸš€ Early evals show exceptional performance in Video Edit. We are now in the final optimization sprint for the official launch in 2 weeks. We invite the community to get early access and test our capabilities at https://t.co/iiyfgPtib5. 🐎✨

πŸ”Sanemavcil retweeted
O
OpenClaw🦞
@openclaw
πŸ“…
Apr 11, 2026
13d ago
πŸ†”58742012
⭐0.34

OpenClaw 2026.4.10 🦞 🧠 Active Memory plugin πŸŽ™οΈ local MLX Talk mode πŸ€– Codex app-server harness plugin 🧾 Teams pins/reactions/read actions πŸ›‘οΈ SSRF hardening + launchd fixes stability, but with attitude🦞 https://t.co/PW7WDumTf1

❀️1,943
likes
πŸ”203
retweets
B
Bayang_BM
@Bayang_BM
πŸ“…
Apr 10, 2026
13d ago
πŸ†”01391043

Now the masked_token_weighted is learning. We ablated the inpainting task, swapped MSE for SmoothL1Loss (more robust to outliers), and per-dim normalized the reconstruction targets, significantly reducing curvature-dim dominance. ref: https://t.co/FL5X61xpbQ https://t.co/0j03IXFXR2

@Bayang_BM β€’ Thu Apr 09 17:11

Training a 10M params foundation model on 8xH100s. The regime is self-supervised pretraining on 29GB of CAD and engineering meshes with masked token modeling, contrastive consistency, and spatial inpainting. You could guess what it is for. https://t.co/y9c9v5d8fN

Media 1Media 2
πŸ–ΌοΈ Media
W
winglian
@winglian
πŸ“…
Apr 17, 2026
6d ago
πŸ†”16104719
⭐0.38

@reza_byt Doesn't tokenwise looped transformer have issues with pretraining since each token has a different depth and also has to learn the recursion depth?