Your curated collection of saved posts and media
Helios Real Real-Time Long Video Generation Model paper: https://t.co/ae0ZH4zPzn https://t.co/kCnNfF3ImI
Heterogeneous Agent Collaborative Reinforcement Learning https://t.co/ASb1VwtCeK
Proact-VL A Proactive VideoLLM for Real-Time AI Companions https://t.co/GkHdSKxSvi
CubeComposer Spatio-Temporal Autoregressive 4K 360ยฐ Video Generation from Perspective Video paper: https://t.co/mnDM1VrYn7 https://t.co/iHtlZJCo1w
LTX-2.3 is out on Hugging Face model: https://t.co/te5nwPL1LE https://t.co/biO7szxFGz
Tencent released HY-WU on Hugging Face An Extensible Functional Neural Memory Framework and An Instantiation in Text-Guided Image Editing model: https://t.co/jAnic8Z9i1 https://t.co/LsLpyjMVQT
New model updates from iquestlab. If you're trying to find an inference model that you can run offline, this is probably the one you're looking for. - 7B and 14B coding models - Optimized for tool use, CLI agents and HTML generation - 128k context length - Explicit and detailed prompting works best - MiT license with requirement of display logo - available on @huggingface
With the help of @huggingface we (/w @RisingSayak) are building a ML Club India ๐ฎ๐ณ What we want to do: 1. Online talks 2. IST compatible timing 2. Open to all More to come in this week! Watch this space. ๐ค Special thanks to @LysandreJik who motivated me to keep working on this. ๐ฅ
Thanks @AnthropicAI. Thanks @huggingface for letting me work on Diffusers and other open-source projects across the fleet. https://t.co/R2qCNHk5lb
๐ฅ Learn how to build your own tool-calling agent with @huggingface TRL + @Alibaba_Qwen Qwen3.5 on @Azure Machine Learning! - @NousResearch hermes-function-calling-v1, 500 single-turn samples - SFT with TRL on Qwen3.5 2B (released today!) on a single NVIDIA H100 - Everything on Azure, from Container Registry to Machine Learning! Step-by-step in the thread ๐งต
agentic RL hackathon this weekend! mentors from @PyTorch, @huggingface , and @UnslothAI will guide you to build agentic environments to win from a pool of $100K prizes ๐ + free compute and token credits just for attending! lock in mar 7-8 in SF. https://t.co/erZRAJrgrA
There is no best VLM OCR model - rankings can flip completely by document type. I built ocr-bench: run open OCR models on YOUR documents, get a per-collection leaderboard. VLM-as-judge with Bradley-Terry ELO, all running on @huggingface. No local GPU needed. https://t.co/qZOwI0Wbes
Diffusers 0.37.0 is out ๐ฅ New models, including LTX-2, Helios, GLM-Image, and more. We're proud to be shipping the wild hot RAEs in this release, too! New CP backends, caching methods, etc., are in too! Check out the release notes for more details ๐งจ https://t.co/fzwmRDgk80
New course: Build and Train an LLM with JAX, built in partnership with @Google and taught by @chrisachard. JAX is the open-source library behind Google's Gemini, Veo, and other advanced models. This short course teaches you to build and train a 20-million parameter language model from scratch using JAX and its ecosystem of tools. You'll implement a complete MiniGPT-style architecture from scratch, train it, and chat with your finished model through a graphical interface. Skills you'll gain: - Learn JAX's core primitives: automatic differentiation, JIT compilation, and vectorized execution - Build a MiniGPT-style LLM using Flax/NNX, implementing embedding and transformer blocks - Load a pretrained MiniGPT model and run inference through a chat interface Come learn this important software layer for building LLMs! https://t.co/wm6NZOGIKC
A small Qwen3.5 from-scratch reimplementation for edu purposes: https://t.co/OnupgeE55l (probably the best "small" LLM today for on-device tinkering) https://t.co/LwyF8x6sle

@DnuLkjkjh This one doesn't have MoE; but I have the larger Qwen3's with MoE if you are interested: https://t.co/IcyLHmP4dz
@vo_d_p @DnuLkjkjh Actually currently working on a sequel :) https://t.co/xJobdXHGyd
@vo_d_p @DnuLkjkjh Actually currently working on a sequel :) https://t.co/xJobdXHGyd
What's actually nice about Gated DeltaNet modules is that they don't grow the KV cache size. So with that 3:1 ratio, Qwen3.5 is much more memory friendly than the previous Qwen3 models. https://t.co/AhegasegIB
๐ธ๐ฌ Glad to help bring @aiDotEngineer to Singapore! Speaker applications open now! https://t.co/sglbpN0bZe apply if you always wanted to go and hang out with the top AI engineers in the Wakanda of Asia. https://t.co/9h6PlFZCfl
r/stims is the most insane fucking place on the internet https://t.co/YKvIgh4ADi
r/stims is the most insane fucking place on the internet https://t.co/YKvIgh4ADi
@youwouldntpost @TheNewThoreau My Alma mater, BCU. The baddest gym in the SWAC. https://t.co/AV5FYJNFqB
GPT-5.4 Thinking and GPT-5.4 Pro are rolling out now in ChatGPT. GPT-5.4 is also now available in the API and Codex. GPT-5.4 brings our advances in reasoning, coding, and agentic workflows into one frontier model. https://t.co/1hy6xXLAmJ