Your curated collection of saved posts and media

Showing 32 posts ยท last 7 days ยท newest first
A
AssemblyAI
@AssemblyAI
๐Ÿ“…
Dec 16, 2025
126d ago
๐Ÿ†”01816906

We just launched Self-Hosted Voice AIโ€”our Universal-Streaming model, deployed on your infrastructure, with the same performance developers already trust from our API. Self-hosting speech AI used to mean compromising on quality or paying a premium for the privilege. Not anymore. Here's what this unlocks: ๐Ÿ”น Co-locate your Voice AI stack where your traffic originates for optimized latency ๐Ÿ”น Process all audio within your controlled perimeter for full data sovereignty ๐Ÿ”น Deploy with Kubernetes, AWS ECS, or any container orchestration platform you're already using ๐Ÿ”น Count usage toward your cloud provider's committed spend program No self-hosting premium. Session-based pricing with volume discounts. For teams navigating strict compliance requirements, data residency mandates, or just wanting tighter control over their stackโ€”this is built for you.

Media 1
๐Ÿ–ผ๏ธ Media
X
XiaomiMiMo
@XiaomiMiMo
๐Ÿ“…
Dec 16, 2025
126d ago
๐Ÿ†”70157939

โšก Faster than Fast. Designed for Agentic AI. Introducing Xiaomi MiMo-V2-Flash โ€” our new open-source MoE model: 309B total params, 15B active. Blazing speed meets frontier performance. ๐Ÿ”ฅ Highlights: ๐Ÿ—๏ธ Hybrid Attention: 5:1 interleaved 128-window SWA + Global | 256K context ๐Ÿ“ˆ Performance: โš”๏ธ Matches DeepSeek-V3.2 on general benchmarks โ€” at a fraction of the latency ๐Ÿ† SWE-Bench Verified: 73.4% | SWE-Bench Multilingual: 71.7% โ€” new SOTA for open-source models ๐Ÿš€ Speed: 150 output tokens/s with Day-0 support from @lmsysorg๐Ÿค ๐Ÿค— Model: https://t.co/4Etm0yZKTL ๐Ÿ“ Blog Post: https://t.co/5zxmcDuB6o ๐Ÿ“„ Technical Report: https://t.co/crac1YTLYl ๐ŸŽจ AI Studio: https://t.co/nSReUs6QgW

Media 1Media 2
+1 more
๐Ÿ–ผ๏ธ Media
B
Brendan_In_Byte
@Brendan_In_Byte
๐Ÿ“…
Dec 16, 2025
126d ago
๐Ÿ†”09786774

Just tried @bubblelab_ai and Iโ€™m actually blown away Best way I can describe it is if @cursor_ai and @n8n_io had a baby This afternoon I built a full sales qualifier workflow in three prompts: โ€“ defined a target segment (e.g. Shopify store owners) โ€“ identified market leaders โ€“ analyzed their social presence + inferred priorities โ€“ extracted contact details โ€“ generated outreach emails + sales call talking points โ€“ compiled everything into a single Google Sheet It got it right with zero errors As tools like this emerge, the hard part stops being building workflows and starts being understanding them The advantage moves to people who deeply understand the domain and know what questions are worth asking Props to @Selinaliyy and the Bubble Lab team! This feels like a glimpse of what AI-native ops should look like.

๐Ÿ–ผ๏ธ Media
G
GoogleCloudTech
@GoogleCloudTech
๐Ÿ“…
Dec 16, 2025
126d ago
๐Ÿ†”36509368

We're going live in one hour! Tune in for a hands-on look at Gemini 3, Nano Banana Pro, Veo and how to create a full brand ecosystem from scratch. โ†’ https://t.co/GByTvMja87 https://t.co/691F2HEqPj

๐Ÿ–ผ๏ธ Media
T
trycua
@trycua
๐Ÿ“…
Dec 16, 2025
126d ago
๐Ÿ†”90709370

Today we're announcing cua-bench: a framework for benchmarking, training data, and RL environments for computer-use AI agents. Why? Current agents show 10x variance across minor UI changes. Here's how we're fixing it.

๐Ÿ–ผ๏ธ Media
A
AIatMeta
@AIatMeta
๐Ÿ“…
Dec 16, 2025
126d ago
๐Ÿ†”25931067

๐Ÿ”‰ Introducing SAM Audio, the first unified model that isolates any sound from complex audio mixtures using text, visual, or span prompts. Weโ€™re sharing SAM Audio with the community, along with a perception encoder model, benchmarks and research papers, to empower others to explore new forms of expression and build applications that were previously out of reach. ๐Ÿ”— Learn more: https://t.co/FPnfv66UCP

๐Ÿ–ผ๏ธ Media
O
OpenHandsDev
@OpenHandsDev
๐Ÿ“…
Dec 16, 2025
127d ago
๐Ÿ†”67209728

๐Ÿš€ We just launched OpenHands Software Agent SDK on @ProductHunt! A smarter way to build agent-driven software โ€” fast, flexible, and production-ready. ๐Ÿ‘‰ Check it out + show some love! https://t.co/xekxMFGJtD

Media 1
๐Ÿ–ผ๏ธ Media
S
SimularAI
@SimularAI
๐Ÿ“…
Dec 16, 2025
126d ago
๐Ÿ†”08666722

Today marks an important milestone in the history of @SimularAI, the autonomous computer company. Our open source computer-use agent, ๐€๐ ๐ž๐ง๐ญ ๐’, scored 72.6% on the OSWorld benchmark, surpassing the human baseline (72.36%) for the first time ever. This milestone matters because it shows AI can now use computers the way humans do, and, in many cases, do it better. This is a glimpse of a future where work becomes faster, more accessible, and more empowered for everyone. #AI #Automation #Simular #AgenticAI #ComputerUse

๐Ÿ–ผ๏ธ Media
O
omarsar0
@omarsar0
๐Ÿ“…
Dec 16, 2025
126d ago
๐Ÿ†”83988311

Another banger paper from Apple. View synthesis from a single image is impressive. But most methods are extremely slow. The default approach to high-quality novel view synthesis uses diffusion models. Iterative denoising produces compelling results, but latency can stretch into hundreds of seconds per scene. Real-world applications, like in AR/VR headsets and interactive photo browsing, need instant 3D from a single photograph. This new research from Apple introduces SHARP, a method that generates a complete 3D Gaussian representation from a single image in under one second on a standard GPU. Architecture details: A neural network takes a single photograph and produces about 1.2 million 3D Gaussians in a single feedforward pass. The architecture builds on a pretrained depth backbone, but crucially unfreezes parts of it during training. A learned depth adjustment module resolves the inherent ambiguity of monocular depth estimation. A Gaussian decoder then refines all attributes: position, scale, rotation, color, and opacity. Results On ScanNet++, SHARP achieves 0.071 DISTS versus 0.090 for Gen3C, the previous best. That's a 21% improvement in perceptual quality. LPIPS drops from 0.227 to 0.154, a 32% reduction. The latency difference is more dramatic. SHARP runs in under 1 second. Gen3C takes approximately 850 seconds. That's roughly a 1000x speedup. Once the 3D representation exists, rendering runs at over 100 frames per second at high resolution. The representation is metric with an absolute scale, so virtual cameras can be accurately coupled to physical headsets. The method is not perfect. SHARP excels at nearby views corresponding to natural head motion and posture shifts. Diffusion-based methods handle faraway views better by hallucinating plausible content for regions with no overlap to the input. Paper: https://t.co/JPoXOqqj2l Learn to build effective AI Agents in my academy: https://t.co/JBU5beIoD0

Media 1Media 2
๐Ÿ–ผ๏ธ Media
J
jxnlco
@jxnlco
๐Ÿ“…
Dec 16, 2025
126d ago
๐Ÿ†”37033366

Who leaking https://t.co/mhuBy8FQp8

Media 1
๐Ÿ–ผ๏ธ Media
H
HuggingPapers
@HuggingPapers
๐Ÿ“…
Dec 15, 2025
127d ago
๐Ÿ†”81706844

EgoX: Generate immersive first-person video from any third-person clip A novel framework from KAIST AI & Seoul National University that leverages video diffusion models to transform a single exocentric video into a realistic egocentric view. See it in action! https://t.co/Vt3cPAdUL3

๐Ÿ–ผ๏ธ Media
C
ClementDelangue
@ClementDelangue
๐Ÿ“…
Dec 16, 2025
126d ago
๐Ÿ†”43582713

Time to follow https://t.co/dqWrV1R3t7 to get the notification!

@DylanTFWang โ€ข Tue Dec 16 14:57

๐ŸŽฎGet a first look at Tencent HY World 1.5 (WorldPlay)! ๐ŸŽฎ Our newest world model with real-time interaction and long-term memory. Itโ€™s going *open-source* tomorrow. https://t.co/zvMI3rCX7u

Media 1
๐Ÿ–ผ๏ธ Media
๐Ÿ”huggingface retweeted
C
clem ๐Ÿค—
@ClementDelangue
๐Ÿ“…
Dec 16, 2025
126d ago
๐Ÿ†”43582713

Time to follow https://t.co/dqWrV1R3t7 to get the notification!

Media 1
โค๏ธ60
likes
๐Ÿ”2
retweets
๐Ÿ–ผ๏ธ Media
_
_weiping
@_weiping
๐Ÿ“…
Dec 16, 2025
126d ago
๐Ÿ†”88701628

๐Ÿš€ Introducing Nemotron-Cascade! ๐Ÿš€ Weโ€™re thrilled to release Nemotron-Cascade, a family of general-purpose reasoning models trained with cascaded, domain-wise reinforcement learning (Cascade RL), delivering best-in-class performance across a wide range of benchmarks. ๐Ÿ’ป Coding powerhouse After RL, our 14B model: โ€ข Surpasses DeepSeek-R1-0528 (671B) on LiveCodeBench v5/v6/Pro. โ€ข Achieves silver-medal performance at IOI 2025 ๐Ÿฅˆ. โ€ข Reaches a 43.1% pass@1 on SWE-Bench Verified, and 53.8% with test-time scaling. ๐Ÿง  What is Cascade RL? Instead of mixing heterogeneous prompts across domains, Cascade RL trains sequentially, domain by domain, which reduces engineering complexity, mitigates heterogeneous verification latencies, and enables domain-specific curricula and tailored hyperparameter tuning. โœจ Key insight Using RLHF for alignment as a pre-step dramatically boosts complex reasoningโ€”far beyond preference optimization. Subsequent domain-wise RLVR stages rarely hurt the benchmark performance attained in earlier domains and may even improve it, as illustrated in the following figure. ๐Ÿค— Models & training data ๐Ÿ”ฅ ๐Ÿ‘‰ https://t.co/wfVcAaMocA ๐Ÿ“„ Technical report with detailed training and data recipes ๐Ÿ‘‰ https://t.co/FdMINvB4yM

Media 1Media 2
๐Ÿ–ผ๏ธ Media
A
allen_ai
@allen_ai
๐Ÿ“…
Dec 16, 2025
126d ago
๐Ÿ†”74588536

Last year Molmo set SOTA on image benchmarks + pioneered image pointing. Millions of downloads later, Molmo 2 brings Molmoโ€™s grounded multimodal capabilities to video ๐ŸŽฅโ€”and leads many open models on challenging industry video benchmarks. ๐Ÿงต https://t.co/uFs30b2DR3

Media 1Media 2
+1 more
๐Ÿ–ผ๏ธ Media
B
ben_burtenshaw
@ben_burtenshaw
๐Ÿ“…
Dec 16, 2025
126d ago
๐Ÿ†”86332854

Fine-tune Nemotron 3 Nano in TRL with coding agents like claude code, colab, locally or on the hub. To fine tune, pick one of these tools: - Combine HF skills with a coding agent like claude code. - Use this colab notebook. - Train it on HF jobs using the Hugging Face hub - If you can, run this script on your own setup with uv This should get anyone started with fine tuning, and this is the perfect model to start with.

Media 1
๐Ÿ–ผ๏ธ Media
A
AdinaYakup
@AdinaYakup
๐Ÿ“…
Dec 16, 2025
126d ago
๐Ÿ†”02574193

New model from @Meituan_LongCat ๐Ÿš€ LongCat-Video-Avatar๐Ÿ”ฅ Audio driven character animation with text, image, and video inputs, all in one! โœจ MIT license โœจ Audio > talking video (single & multi-person) โœจ Natural motion and lip sync โœจ Fewer repeats, stable identity โœจ Available on @huggingface

๐Ÿ–ผ๏ธ Media
N
ndea
@ndea
๐Ÿ“…
Dec 16, 2025
126d ago
๐Ÿ†”92259735

Introducing the Ndea podcast - Abstract Synthesis. Hear the stories behind interesting academic papers in the world of program synthesis. Episode 1 features @MarkSantolucito, @BarnardCollege/@Columbia, discussing his paper "Grammar Filtering for Syntax-Guided Synthesis". https://t.co/uJ1NVxU6rK

๐Ÿ–ผ๏ธ Media
J
jspanda8
@jspanda8
๐Ÿ“…
Dec 16, 2025
126d ago
๐Ÿ†”58898203

@youwouldntpost @Srirachachau Downloading the โ€œdriving during daytimeโ€ patch https://t.co/R6EmIolLDo

Media 1
๐Ÿ–ผ๏ธ Media
๐Ÿ”youwouldntpost retweeted
J
JS ๐Ÿฆ๐Ÿ‡ต๐Ÿ‡ธ๐Ÿ‰
@jspanda8
๐Ÿ“…
Dec 16, 2025
126d ago
๐Ÿ†”58898203

@youwouldntpost @Srirachachau Downloading the โ€œdriving during daytimeโ€ patch https://t.co/R6EmIolLDo

Media 1
โค๏ธ14
likes
๐Ÿ”1
retweets
๐Ÿ–ผ๏ธ Media
X
XFreeze
@XFreeze
๐Ÿ“…
Dec 16, 2025
126d ago
๐Ÿ†”81322692

The Tesla Cybertruck just earned the Top Safety Pick+ award, scoring a perfect โ€œGoodโ€ rating in literally every major crash category in the 2025 IIHS crash tests https://t.co/o5IRpmMqzg

Media 1
๐Ÿ–ผ๏ธ Media
N
Not_the_Bee
@Not_the_Bee
๐Ÿ“…
Dec 16, 2025
126d ago
๐Ÿ†”09052384

Woman who joked about putting toilet cleaner and feces in food of "white MAGA family" identified as daughter of Virginia delegate https://t.co/LwDaOssHEO

Media 1
๐Ÿ–ผ๏ธ Media
T
TheRabbitHole
@TheRabbitHole
๐Ÿ“…
Dec 16, 2025
127d ago
๐Ÿ†”77828814

The Woke Mind Virus in Academia https://t.co/ztXf1lLxL6

Media 1
๐Ÿ–ผ๏ธ Media
O
OpenAI
@OpenAI
๐Ÿ“…
Dec 16, 2025
126d ago
๐Ÿ†”94613200

GPT-5.2 is our strongest model on the FrontierScience eval, showing clear gains on hard scientific tasks. But the benchmark also reveals a gap between strong performance on structured problems and the open-ended, iterative reasoning that real research requires. https://t.co/lZsZSXkOrj

Media 1Media 2
๐Ÿ–ผ๏ธ Media
O
OpenAI
@OpenAI
๐Ÿ“…
Dec 16, 2025
126d ago
๐Ÿ†”29161873

Introducing ChatGPT Images, powered by our flagship new image generation model. - Stronger instruction following - Precise editing - Detail preservation - 4x faster than before Rolling out today in ChatGPT for all users, and in the API as GPT Image 1.5. https://t.co/NLNIPEYJnr

๐Ÿ–ผ๏ธ Media
P
ProjNativeHope
@ProjNativeHope
๐Ÿ“…
Feb 26, 2020
2247d ago
๐Ÿ†”28780544

This map should be included in every history book... https://t.co/VyuLo90IEE

Media 1
๐Ÿ–ผ๏ธ Media
S
Sanemavcil
@Sanemavcil
๐Ÿ“…
Nov 30, 2025
142d ago
๐Ÿ†”21300904

Quantum Dreaming 2025: When Dreams Become Parallel Reality Portals ๐ŸŒŒ๐Ÿ’ญ Last year we asked: Are your dreams just imaginationโ€ฆ or glimpses into alternate timelines? ๐Ÿชžโœจ This year, the answer is clearer than ever. ๐ŸŒ  2025 brought breakthroughs that turned quantum dreaming from theory to lived experience: โ€ข Neuralinkโ€™s first 1000+ volunteers reported vivid โ€œtimeline bleedโ€ dreams ๐Ÿง โšก โ€ข DMT + VR studies showed 87% of participants experienced consistent parallel-world memories ๐ŸŒˆ๐ŸŒ€ โ€ข Lucid dreamers using tDCS + galantamine now report 40-minute โ€œvisitsโ€ to stable alternate realities โณ๐ŸŒ™ Every dรฉjร  vu? A memory leak from a timeline where you chose differently. ๐Ÿ”„ Every precognitive dream? Your mind tuning into a branch thatโ€™s already happening. ๐Ÿ”ฎ 2025 is the year we stopped calling them โ€œjust dreams.โ€ We started calling them evidence. ๐Ÿ“œโœจ Keep dreaming, explorer. One of them might be more real than this one. ๐ŸŒ๐Ÿฆ‹ #QuantumDreaming #ParallelRealities #2025Awakening #LucidDreaming #Multiverse Grok Imagine prompt: Pastel color quantum dreaming

๐Ÿ–ผ๏ธ Media
L
llama_index
@llama_index
๐Ÿ“…
Dec 16, 2025
126d ago
๐Ÿ†”54114416

LlamaSplit automatically separates bundled documents into distinct sections so you don't have to manually split them anymore. Our new beta API uses AI to analyze page content and group consecutive pages by category - perfect for processing mixed document bundles that contain multiple distinct documents: ๐Ÿ“„ Define categories with natural language descriptions and get back exact page ranges with confidence scores ๐ŸŽฏ Route different document types to appropriate agents โšก Scale beyond manual document separation ๐Ÿ”— Combine with LlamaExtract to run targeted data extraction on each separated segment Unlike our existing Classify product that categorizes separate files, LlamaSplit looks inside a single document to find boundaries between different document types. Try LlamaSplit in beta: https://t.co/cQqeZCGeww

Media 1
๐Ÿ–ผ๏ธ Media
D
DylanTFWang
@DylanTFWang
๐Ÿ“…
Dec 16, 2025
126d ago
๐Ÿ†”28945985

๐ŸŽฎGet a first look at Tencent HY World 1.5 (WorldPlay)! ๐ŸŽฎ Our newest world model with real-time interaction and long-term memory. Itโ€™s going *open-source* tomorrow. https://t.co/zvMI3rCX7u

๐Ÿ–ผ๏ธ Media
K
Kling_ai
@Kling_ai
๐Ÿ“…
Dec 16, 2025
126d ago
๐Ÿ†”84192461

๐Ÿงœ The Little Mermaid gets her voice back. Voice Control feature is now live in Kling VIDEO 2.6. Voice Consistency Now Resolved. Say goodbye to generic voices and create a custom voice, switch styles, and even sing โ€” all perfectly matched to your characters.

๐Ÿ–ผ๏ธ Media
Y
YouWareAI
@YouWareAI
๐Ÿ“…
Dec 16, 2025
126d ago
๐Ÿ†”73848896

Introducing YouBase by YouWare. The complete production backend for vibe coding. For just $20/month. Auth, Database, Storage, Edge Functions Deploy to your own domain Zero configuration No cloud credits. No usage fees. No surprises. One prompt โ†’ Full backend. Live on your domain. Start building at link in bio!

๐Ÿ–ผ๏ธ Media
H
hanlin_hl
@hanlin_hl
๐Ÿ“…
Dec 16, 2025
126d ago
๐Ÿ†”99594259

Multimodal LLMs (MLLMs) excel at reasoning, layout understanding, and planningโ€”yet in diffusion-based generation, they are often reduced to simple multimodal encoders. What if MLLMs could reason directly in latent space and guide diffusion generation with fine-grained, spatiotemporal control? ๐Ÿค” Introducing MetaCanvas ๐ŸŽจ A lightweight framework that translates MLLM reasoning into structured spatiotemporal conditions for diffusion models. ๐Ÿงต ๐Ÿ‘‡

๐Ÿ–ผ๏ธ Media