Your curated collection of saved posts and media

Showing 31 posts ยท last 14 days ยท by score
V
vercel_dev
@vercel_dev
๐Ÿ“…
Apr 15, 2026
8d ago
๐Ÿ†”73960733
โญ0.36

Use Vercel Sandbox with the OpenAI agents SDK as an official extension. Build agents that can run code, read files, and analyze data safely inside isolated microVMs. Control the compute and data flow from your secure cloud environment.

@OpenAIDevs โ€ข Wed Apr 15 17:23

Build long-running agents with more control over agent execution. New capabilities in the Agents SDK: โ€ข Run agents in controlled sandboxes โ€ข Inspect and customize the open-source harness โ€ข Control when memories are created and where theyโ€™re stored https://t.co/zPyuLup6b6

R
rawat_ritvik
@rawat_ritvik
๐Ÿ“…
Apr 15, 2026
8d ago
๐Ÿ†”42603056
โญ0.36

We are looking for excellent people to help build our vertically integrated AI stack. Numerics, quantization, HW simulators, compiler, runtime, kernel performance, RTL, verification, emulation, DFT, physical design, post Si bringup. Join us at Tesla!

@ โ€ข

C
codestantine
@codestantine
๐Ÿ“…
Apr 15, 2026
8d ago
๐Ÿ†”22495977

Excited about the Agents SDK updates we just launched. Check out my cookbook on using it with sandboxes for code migration: https://t.co/Fz7cknz64d

Media 1
๐Ÿ–ผ๏ธ Media
E
e2b
@e2b
๐Ÿ“…
Apr 15, 2026
8d ago
๐Ÿ†”67416751

OpenAI x E2B: build your agents with the new OpenAI Agents SDK, powered by E2B sandboxes. We're excited to support OpenAI as a launch partner! The new @OpenAI Agents SDK will now get dedicated sandboxes - perfect for persistent, long-running agents. With E2B, you'll get a custom environment with resource isolation and security boundaries, with no infrastructure setup required. Your agents will be able to: - Edit files and run shell commands in isolated environments - Maintain temporary workspace state across steps - Produce artifacts you can review before publishing - Run multiple sandboxes in parallel for concurrent workloads - Generate frontend output with live preview URLs ... and more, with a few lines of code! Learn more and see the end-to-end example in the thread:

Media 1
๐Ÿ–ผ๏ธ Media
๐Ÿ”tri_dao retweeted
S
Monishwaran Maheswaran
@sudomonish
๐Ÿ“…
Apr 10, 2026
13d ago
๐Ÿ†”67173352
โญ0.32

Super excited to Introduce our latest work: Squeeze Evolve. We unify test-time scaling methods into one evolutionary framework โ€” then orchestrate many models across it. 3x lower cost. 10x throughput. 97.5%(SoTA) on ARC-AGI-V2. No verifier required. Framework: https://t.co/5hmOyZvSKU

โค๏ธ51
likes
๐Ÿ”12
retweets
M
mlejva
@mlejva
๐Ÿ“…
Apr 15, 2026
8d ago
๐Ÿ†”73523661

OpenAI x e2b: Build your agents with the new OpenAI Agents SDK, powered by @E2B Sandboxes. Excited to support @OpenAI as a launch partner! https://t.co/RsSw1HsF86

Media 1
๐Ÿ–ผ๏ธ Media
L
LeeChiJung
@LeeChiJung
๐Ÿ“…
Apr 12, 2026
11d ago
๐Ÿ†”79538485

No cameras. No extra sensors. Your smartwatch already has everything it needs to track your hand. โŒš๏ธโœŒ๐Ÿป Monday at #CHI2026, @jiwan_hci and I are presenting WatchHand, a continuous 3D hand pose tracking system that uses just the speaker and mic in your smartwatch. https://t.co/8bXMI2Mux4

๐Ÿ–ผ๏ธ Media
M
mntruell
@mntruell
๐Ÿ“…
Apr 13, 2026
10d ago
๐Ÿ†”68555717
โญ0.38

@jasonkneen We A/B test many parts of Cursor: model checkpoints, UX, and the agent harness. In this case, we tested less than 1% of traffic to compare how Claude behaves with the CC harness versus our default harness (something we do often with offline evals). Our team does lots to improve the speed, feel, and accuracy of our harness on the queries users ask and care about. Hope to share more about this work soon.

C
cocktailpeanut
@cocktailpeanut
๐Ÿ“…
Apr 16, 2026
7d ago
๐Ÿ†”57964024

Generate FULLY CONTROLLABLE 3D assets from a SINGLE image, locally on your PC. Made a 1-click launcher for the official Anigen Gradio app, and a dedicated viewer. Crazy this is now possible. What you're seeing here came from one image. Requires: NVIDIA GPU 6GB VRAM

@yanpei_cao โ€ข Tue Apr 14 16:45

Static 3D generation isn't enough. We need assets ready for animation. Our new #SIGGRAPH work, AniGen, takes a single image and generates the 3D shape, skeleton, and skinning weights all at once. Code is fully open-sourced! Kudos to @KyrieIr31012755 and @VastAIResearch ๐Ÿงต(1/4) h

๐Ÿ–ผ๏ธ Media
V
victormustar
@victormustar
๐Ÿ“…
Apr 13, 2026
10d ago
๐Ÿ†”19359033

GLM-5.1 > Claude Code (Opus 4.6)? I'm tripping or CC has become very bad but built a Three.js racing game to eval and it's extremely impressive. Thoughts: - One-shot car physics with real drift mechanics (this is hard) - My fav part: Awesome at self iterating (with no vision!) created 20+ Bun.WebView debugging tools to drive the car programmatically and read game state. Proved a winding bug with vector math without ever seeing the screen - 531-line racing AI in a single write: 4 personalities, curvature map, racing lines, tactical drifting. Built telemetry tools to compare player vs AI speed curves and data-tuned parameters - All assets from scratch: 3D models, procedural textures, sky shader, engine sounds, spatial AI audio! - Can do hard math: proved road normals pointed DOWN via vector cross products, computed track curvature normalized by arc length to tune AI cornering speed You are going to hear about this model a lot in the next months - open source let's go ๐Ÿš€๐Ÿš€

๐Ÿ–ผ๏ธ Media
T
theo
@theo
๐Ÿ“…
Apr 17, 2026
6d ago
๐Ÿ†”03919748
โญ0.34

I have feelings about Opus 4.7. https://t.co/km54XbnDMk

๐Ÿ”SpirosMargaris retweeted
C
Claude
@claudeai
๐Ÿ“…
Apr 14, 2026
9d ago
๐Ÿ†”60309790
โญ0.38

Now in research preview: routines in Claude Code. Configure a routine once (a prompt, a repo, and your connectors), and it can run on a schedule, from an API call, or in response to an event. Routines run on our web infrastructure, so you don't have to keep your laptop open. https://t.co/m2XJWYqkf8

โค๏ธ13,953
likes
๐Ÿ”1,079
retweets
M
mervenoyann
@mervenoyann
๐Ÿ“…
Apr 16, 2026
7d ago
๐Ÿ†”16340765

this is the year of 3D World models ๐Ÿ”ฅ > Lyra 2.0 by NVIDIA: image โ†’ 3D world with Gaussians, 14B params, built on WAN-14B > HY-World 2.0 by Tencent: text/image/video 3D โ†’ editable world (meshes + Gaussians) drop-in to Blender/Unity/Unreal weights on the next one โžก๏ธ https://t.co/MTSdHmRE2P

Media 1
๐Ÿ–ผ๏ธ Media
H
Heyhassan
@Heyhassan
๐Ÿ“…
Apr 16, 2026
7d ago
๐Ÿ†”45131317

This is crazy good Grok Code built a full e-commerce website in less than an hour. Here is how i do this full tutorial + prompts: โ†“ https://t.co/bAmlxqEoOv

๐Ÿ–ผ๏ธ Media
A
arcinstitute
@arcinstitute
๐Ÿ“…
Apr 13, 2026
10d ago
๐Ÿ†”19823515

Most genomic AI models use fixed rules to process DNA into chunks, imposing arbitrary boundaries on a sequence with its own biological structure. @arnavshah0, @victor_ljz, and team developed dnaHNet, a tokenizer-free foundation model that learns its own segmentation from scratch, supervised by @_albertgu, @genophoria, and @BoWang87.

Media 1
๐Ÿ–ผ๏ธ Media
A
Azaliamirh
@Azaliamirh
๐Ÿ“…
Apr 13, 2026
10d ago
๐Ÿ†”90192893

Turns out we can get SOTA on agentic benchmarks with a simple test-time method! Excited to introduce LLM-as-a-Verifier. Test-time scaling is effective, but picking the "winner" among many candidates is the bottleneck. We introduce a way to extract a cleaner signal from the model: 1๏ธโƒฃ Ask the LLM to rank results on a scale of 1-k 2๏ธโƒฃ Use the log-probs of those rank tokens to calculate an expected score You can get a verification score in a single sampling pass per candidate pair. Blog: https://t.co/jYPZUgncLe Code: https://t.co/caBpzd3Xkx Led by @jackyk02 and in collaboration with a great team: @shululi256, @pranav_atreya, @liu_yuejiang, @drmapavone, @istoica05

Media 1Media 2
๐Ÿ–ผ๏ธ Media
๐Ÿ”unknown_user retweeted
U
unknown_user
@unknown_user
๐Ÿ“…
Apr 17, 2026
6d ago
๐Ÿ†”20431022
โญ0.36

This paper makes a strong case for open-world evaluations as a complement to traditional benchmarks, particularly for realistic, long-horizon, open-ended settings! Glad the AISI SoE team could contribute to this effort.

โค๏ธ18
likes
๐Ÿ”5
retweets
H
hardmaru
@hardmaru
๐Ÿ“…
Apr 13, 2026
10d ago
๐Ÿ†”81204232

ไธ€ใคใฎใƒ‹ใƒฅใƒผใƒฉใƒซใƒใƒƒใƒˆใซ็ฌฆๅทใจ่จ˜ๅทใฏๅ‰ต็™บใ—ใ†ใ‚‹ใ‹๏ผŸ ใ€ŒNeural Computersใ€่ซ–ๆ–‡ใ‹ใ‚‰่€ƒใˆใ‚‹ @rmaruy https://t.co/JZPE0bdCeK

Media 1
๐Ÿ–ผ๏ธ Media
A
Adam_Fish
@Adam_Fish
๐Ÿ“…
Apr 17, 2026
6d ago
๐Ÿ†”01580794
โญ0.38

Webflow's CMS API can't publish code blocks. Tables aren't in the API at all!? So I built a Playwright robot that clicks buttons in the Designer for us. In 2026, your API is your product. https://t.co/sA7Csu21uO

S
sparkjsdev
@sparkjsdev
๐Ÿ“…
Apr 14, 2026
9d ago
๐Ÿ†”82816449

Spark 2.0 is here! ๐Ÿš€ Weโ€™re redefining whatโ€™s possible on the web with a streamable LoD system for 3D Gaussian Splatting. Built on Three.js, you can now stream massive 100M+ splat worlds to any device from mobile to VR using WebGL2. All open-source. Dive into the tech ๐Ÿ‘‡ https://t.co/VOd6V0Wz1s

๐Ÿ–ผ๏ธ Media
Y
YinghaoXu1
@YinghaoXu1
๐Ÿ“…
Apr 16, 2026
7d ago
๐Ÿ†”53047028
โญ0.42

๐ŸŽ‰ After one year of teamwork, we are excited to release our 3D foundation model โ€” LingBot-Map! Unlike DA3/VGGT, LingBot-Map is a purely autoregressive model for streaming 3D reconstruction โšก It achieves ~20 FPS on 518ร—378 resolution over sequences exceeding 10,000 frames โ€” and beyond ๐Ÿš€ Two key insights behind LingBot-Map: ๐Ÿ”‘ Keep SLAM's structural wisdom: build Geometric Context Attention with long-context modeling while maintaining a compact streaming state ๐Ÿ”‘ Make everything end-to-end learnable โ€” no optimization, no post-processing Let's check out our demos ๐Ÿ‘‡

A
AkwyZ
@AkwyZ
๐Ÿ“…
Apr 15, 2026
8d ago
๐Ÿ†”87065813
โญ0.44

The Strange Origin of AIโ€™s โ€˜Reasoningโ€™ Abilities https://t.co/lXyZw8U4u4 #TechNews @ArturHabant @elaniazito @IanLJones98 @CurieuxExplorer @Shi4Tech @enilev @Fabriziobustama @mvollmer1 @AnthonyRochand @JolaBurnett @lyakovet @debashis_dutta @3itcom @ahier @Analytics_699 @antgrasso @CathCervoni @chidambara09 @DigitalColmer @dinisguarda @DimitriHommel @EvanKirstel @FrRonconi @GlenGilmore @gvalan @HeinzVHoenen @ipfconline1 @jeancayeux @jorgecunha @kalydeoo @nafisalam @Nicochan33 @pierrepinna @PawlowskiMario @puneetsinghal22 @ralph_ohr @RLDI_Lamy @rshevlin @sarbjeetjohal @SpirosMargaris @StefanoDeCupis @tewoz @thomas_dettling @Ym78200 @aure79lien @jblefevre60

T
techdevnotes
@techdevnotes
๐Ÿ“…
Apr 17, 2026
6d ago
๐Ÿ†”17790610

Grok 4.3 can Take in Video and Extract Audio files https://t.co/5uprx2dM85

Media 1
๐Ÿ–ผ๏ธ Media
A
AWSstartups
@AWSstartups
๐Ÿ“…
Apr 10, 2026
13d ago
๐Ÿ†”25105855

AI is transforming software development, but more code means more pull requests, more edge cases, and more QA pressure on engineering teams. Tusk, a Y Combinator startup, catches bugs that slip past both AI agents and humans using AI-enabled tests based on real production traffic. Built on Amazon Bedrock, Tusk flags issues before code merge so teams can focus on building great products.

๐Ÿ–ผ๏ธ Media
H
hardmaru
@hardmaru
๐Ÿ“…
Apr 11, 2026
13d ago
๐Ÿ†”40265641

A "Neural Computer" is built by adapting video generation architectures to train a World Model of an actual computer that can directly simulate a computer interface. Instead of interacting with a real operating system, these models can take in user actions like keystrokes and mouse clicks alongside previous screen pixels to predict and generate the next video frames. Trained solely on recorded input and output traces, it successfully learned to render readable text and control a cursor, proving that a neural network can run as its own visual computing environment without a traditional operating system. https://t.co/roTpqsdrEE Cool work by @MingchenZhuge @SchmidhuberAI et al.!

@MingchenZhuge โ€ข Fri Apr 10 14:15

๐Ÿซฑ Introducing ๐๐ž๐ฎ๐ซ๐š๐ฅ ๐‚๐จ๐ฆ๐ฉ๐ฎ๐ญ๐ž๐ซs: ๐ฐ๐ก๐š๐ญ ๐ข๐Ÿ ๐€๐ˆ ๐๐จ๐ž๐ฌ ๐ง๐จ๐ญ ๐ฃ๐ฎ๐ฌ๐ญ ๐ฎ๐ฌ๐ž ๐œ๐จ๐ฆ๐ฉ๐ฎ๐ญ๐ž๐ซ๐ฌ ๐›๐ž๐ญ๐ญ๐ž๐ซ, ๐›๐ฎ๐ญ ๐›๐ž๐ ๐ข๐ง๐ฌ ๐ญ๐จ ๐›๐ž๐œ๐จ๐ฆ๐ž ๐ญ๐ก๐ž ๐ซ๐ฎ๐ง๐ง๐ข๐ง๐  ๐œ๐จ๐ฆ๐ฉ๐ฎ๐ญ๐ž๐ซ ๐ข๐ญ๐ฌ๐ž๐ฅ๐Ÿ? Beyond today's conventional computers, agents, and world models, Neural Computers (NCs) are new frontiers where computation, memory, and I/O move in

Media 1
๐Ÿ–ผ๏ธ Media
S
stevibe
@stevibe
๐Ÿ“…
Apr 17, 2026
6d ago
๐Ÿ†”94658539

Introducing HermesAgent-20, a new Bench Pack for BenchLocal. 20 scenarios extracted straight from the Hermes Agent source code, run against a REAL Hermes instance. The actual workload you'd put your model through. Why I built BenchLocal in the first place: most benchmarks are too abstract. We use local LLMs for practical work, and finding the right model for YOUR task efficiently is the single most important thing, especially when you're constrained to what fits on your machine. BenchLocal is a framework: providers, models, side-by-side comparison, all in one UI. Bench Packs are the unit of testing: ToolCall-15 and BugFind-15 shipped first, and when I launched the BenchLocal 0.1.0, added StructOutput, ReasonMath, InstructFollow, DataExtract. Now, HermesAgent-20 is the newest. Bench Packs install like VS Code extensions. The SDK is open, write your own, share it, grow the ecosystem. Here's the goal: a community-built, practical evaluation layer for the local LLM space. Early numbers on HermesAgent-20: > GLM 5.1 โ€” 85 > Gemma4 31B โ€” 83 > Qwen3.5 27B โ€” 79 > MiniMax M2.7 โ€” 76 Upgrade to the latest BenchLocal to install HermesAgent-20 (SDK update required).

๐Ÿ–ผ๏ธ Media
A
ashen_one
@ashen_one
๐Ÿ“…
Apr 17, 2026
6d ago
๐Ÿ†”12381770

okok we officially have GLM 5.1 running on a 256gb mac studio with hermes agent next is linking it to hermes to see how good it is ๐Ÿ—ฃ๏ธ https://t.co/BQTlLiL3jm

Media 1
๐Ÿ–ผ๏ธ Media
U
UnslothAI
@UnslothAI
๐Ÿ“…
Apr 10, 2026
13d ago
๐Ÿ†”60796991

Google DeepMind is hosting a Gemma 4 hackathon with a $10,000 Unsloth prize! ๐Ÿฆฅ Show off your best fine-tuned Gemma 4 model built with Unsloth. There's $200,000 total prizes to be won. Challenge info + Notebook: https://t.co/HndHPaXICT https://t.co/cBnNro1fVI

Media 1
๐Ÿ–ผ๏ธ Media
L
LeRobotHF
@LeRobotHF
๐Ÿ“…
Apr 09, 2026
14d ago
๐Ÿ†”01212658

This is the full video of the hardest version of the task: t-shirt folding from unstructured initial states. This setting really requires at least some strategy, since the robot first has to spread the shirt before it can complete the fold. Full details on data collection strategies in the blog below. ๐Ÿ‘‡

@LeRobotHF โ€ข Tue Apr 07 15:45

Releasing the Unfolding Robotics blog! Time to unfold robotics: we trained a robot to fold clothes using 8 bimanual setups, 100+ hours of demonstrations, and 5k+ GPU hours. Flashy robot demos are everywhere. But you rarely see the real story: the data, the failures, the enginee

๐Ÿ–ผ๏ธ Media
S
SandeepRoutra11
@SandeepRoutra11
๐Ÿ“…
Apr 17, 2026
6d ago
๐Ÿ†”74808506

๐Ÿš€ Excited to share ViPRA: Video Prediction for Robot Actions ๐Ÿ“ Accepted to #ICLR2026 @iclr_conf ๐Ÿ† Best Paper โ€” #NeurIPS2025 Embodied World Models Workshop Robot learning today still needs millions of action labeled videos. Yet videos are abundant โ€” from humans and the web โ€” but lack action labels. Meanwhile, pretrained video models already learn rich dynamics. ViPRA is a recipe for turning pretrained video models into robot policies while enabling robot learning to scale with actionless videos. ๐Ÿงต Thread โ†“

๐Ÿ–ผ๏ธ Media
A
Agent_Crafter
@Agent_Crafter
๐Ÿ“…
Apr 17, 2026
6d ago
๐Ÿ†”94748866

Trying to write a tutorial for hermes, anyone interested? https://t.co/eZDnEgtZGy

Media 1
๐Ÿ–ผ๏ธ Media