Your curated collection of saved posts and media

Showing 32 posts Β· last 14 days Β· by score
D
davidsacks47
@davidsacks47
πŸ“…
Sep 05, 2025
233d ago
πŸ†”57000694

Only President Trump could convene so many tech leaders in one place. We had a constructive conversation about how to grow the economy with new infrastructure spending. This will benefit not only software companies but manufacturing, energy, construction, the trades and workers. https://t.co/VK1fdHlxSq

Media 1
πŸ–ΌοΈ Media
W
WhiteHouse
@WhiteHouse
πŸ“…
Sep 05, 2025
233d ago
πŸ†”89673518

"I'm thrilled to announce that the 2026 G20 conference... will be held in one of our country's greatest cities β€” beautiful Miami, Florida." @POTUS πŸ‡ΊπŸ‡Έ https://t.co/fziGJ5aqo4

Media 1
πŸ–ΌοΈ Media
B
briannekimmel
@briannekimmel
πŸ“…
Sep 05, 2025
233d ago
πŸ†”25258657

A great builder night at @elevenlabsio, especially love learning about the psychology of voice. Such a core part of who we are and how we experience the world. https://t.co/lWPW0eOSUP

Media 1
πŸ–ΌοΈ Media
B
briannekimmel
@briannekimmel
πŸ“…
Sep 05, 2025
233d ago
πŸ†”54498020

@maxniederhofer @elevenlabsio @matistanis packed house, also short lol https://t.co/PmoMemOZzD

Media 1
πŸ–ΌοΈ Media
A
anatelorenzen
@anatelorenzen
πŸ“…
Sep 05, 2025
233d ago
πŸ†”00744793

β€œMom how did we get so rich?” β€œYour dad listened to Nikita and posted about plumbing for 180 days.” https://t.co/2JDah5YvhG

Media 1
πŸ–ΌοΈ Media
B
briannekimmel
@briannekimmel
πŸ“…
Sep 06, 2025
232d ago
πŸ†”11589026

@alanaagoyal same as a hair, a little trim and some product and it’ll be great. We the Wild has great products if you like a more waxy look. https://t.co/Cq9oKgBsUT

Media 1
πŸ–ΌοΈ Media
M
mervenoyann
@mervenoyann
πŸ“…
Sep 02, 2025
236d ago
πŸ†”32229797

best small vision LM with reasoning has dropped on @huggingface πŸ”₯ Tencent dropped R-4B, small vision LM that claims sota with Apache 2.0 license πŸ’— the model enables different thinking options and transformers support through custom code! https://t.co/xUCM010flp

Media 1
πŸ–ΌοΈ Media
M
multimodalart
@multimodalart
πŸ“…
Sep 02, 2025
236d ago
πŸ†”76729488

we hacked Wan 2.2 and discovered that it does first and last frame filling, works out of the box on 🧨 diffusers i've built an app for it on @huggingface Spaces (which is powering powering our nano banana video mode too 🍌 🎬) https://t.co/40yRzpWCvN

πŸ–ΌοΈ Media
C
corbtt
@corbtt
πŸ“…
Sep 02, 2025
236d ago
πŸ†”78048297

🚨 We’ve just published a recipe to train a frontier-level deep research agent using RL. With just 30 hours on an H200, any developer can now beat Sonnet-4 on DeepResearch Bench using open-source tools. (Thread 🧡) https://t.co/Ul7htDkmPX

Media 1
πŸ–ΌοΈ Media
H
Hesamation
@Hesamation
πŸ“…
Sep 02, 2025
236d ago
πŸ†”93447178

a senior engineer at google just dropped a 400-page free book on docs for review: agentic design patterns. the table of contents looks like everything you need to know about agents + code: > advanced prompt techniques > multi-agent patterns > tool use and MCP > you name it https://t.co/DIIaDOpdGj

Media 1
πŸ–ΌοΈ Media
T
TheAhmadOsman
@TheAhmadOsman
πŸ“…
Sep 03, 2025
235d ago
πŸ†”97095037

why don't i like ollama & what do i use on my AI server? a thread of blogposts where i go over: > my ai homelab setup > what are inference engines > how local llms actually work > why i don't recommend ollama > what do i use on my AI server > the different use cases https://t.co/c1oqvvAUcU

Media 1Media 2
πŸ–ΌοΈ Media
T
TheAhmadOsman
@TheAhmadOsman
πŸ“…
Sep 03, 2025
235d ago
πŸ†”70053053

COMING SOON: The LLM GPU Build Guide > from 1x to 8x GPUs > inference, training, cpu offload, and rackmounts > budgets from $2K to $15K+ > full parts lists, tradeoffs, where to buy, and everything learned the hard way > PCIe lanes & bandwidth, why 2^n # of GPUs matter & more https://t.co/Zu6D7DfodC

Media 1
πŸ–ΌοΈ Media
L
lusxvr
@lusxvr
πŸ“…
Sep 04, 2025
234d ago
πŸ†”46293448

Today, we are releasing FineVision, a huge open-source dataset for training state-of-the-art Vision-Language Models: > 17.3M images > 24.3M samples > 88.9M turns > 9.5B answer tokens Here are my favourite findings: https://t.co/pfP8OMBvmH

Media 1
πŸ–ΌοΈ Media
J
JinaAI_
@JinaAI_
πŸ“…
Sep 04, 2025
234d ago
πŸ†”39007824

Today we're releasing jina-code-embeddings, a new suite of code embedding models in two sizesβ€”0.5B and 1.5B parametersβ€”along with 1~4bit GGUF quantizations for both. Built on latest code generation LLMs, these models achieve SOTA retrieval performance despite their compact size. They support over 15 programming languages and 5 tasks: nl2code, code2code, code2nl, code2completions and qa.

Media 1
πŸ–ΌοΈ Media
A
Alibaba_Qwen
@Alibaba_Qwen
πŸ“…
Sep 05, 2025
233d ago
πŸ†”40562976

Big news: Introducing Qwen3-Max-Preview (Instruct) β€” our biggest model yet, with over 1 trillion parameters! πŸš€ Now available via Qwen Chat & Alibaba Cloud API. Benchmarks show it beats our previous best, Qwen3-235B-A22B-2507. Internal tests + early user feedback confirm: stronger performance, broader knowledge, better at conversations, agentic tasks & instruction following. Scaling works β€” and the official release will surprise you even more. Stay tuned! Qwen Chat: https://t.co/V7RmqMaVNZ Alibaba Cloud API: https://t.co/zjCKdWee5v

Media 1
πŸ–ΌοΈ Media
K
Kimi_Moonshot
@Kimi_Moonshot
πŸ“…
Sep 05, 2025
234d ago
πŸ†”30947698

Kimi K2-0905 update πŸš€ - Enhanced coding capabilities, esp. front-end & tool-calling - Context length extended to 256k tokens - Improved integration with various agent scaffolds (e.g., Claude Code, Roo Code, etc) πŸ”— Weights & code: https://t.co/SsFKTnWslD πŸ’¬ Chat with new Kimi K2 on: https://t.co/2bLWEHF6az ⚑️ For 60–100 TPS + guaranteed 100% tool-call accuracy, try our turbo API: https://t.co/EOZkbOwCN4

Media 1Media 2
+1 more
πŸ–ΌοΈ Media
S
StanfordOnline
@StanfordOnline
πŸ“…
Sep 05, 2025
233d ago
πŸ†”86590180

New Stanford CS231N Deep Learning for Computer Vision lectures taught by Professor Fei-Fei Li, Assistant Professors Ehsan Adeli and Justin Johnson, and Zane Durante are now available! Watch the complete playlist here: https://t.co/yZZhxheLMa

Media 1
πŸ–ΌοΈ Media
J
jandotai
@jandotai
πŸ“…
Sep 04, 2025
234d ago
πŸ†”69956243

Meet Jan-v1-edge: an experimental 1.7B distilled model for Perplexity-style search. Jan-v1-edge is our lightweight distillation experiment, derived from Jan v1. We're testing how well web search and reasoning can transfer into a smaller 1.7B parameter model that runs on edge devices. Performance - 83% SimpleQA accuracy, close to Jan-nano-128k while being lighter - Outperforms Qwen3-1.7B Thinking on SimpleQA To experiment with it, find the GGUF model on @huggingface, click Use this model and select Jan. To enable search in Jan: go to Settings -> MCP Servers -> enable or add a search-related MCP (SearXNG, Serper, Exa, etc.). - Jan-v1-edge: https://t.co/SgHEcdqpQi - Jan-v1-edge GGUF: https://t.co/rnAyvFBNzA Credit to the @Alibaba_Qwen team for Qwen3-1.7B Thinking and @ggerganov for llama.cpp.

Media 1Media 2
+1 more
πŸ–ΌοΈ Media
M
mark_k
@mark_k
πŸ“…
Sep 03, 2025
235d ago
πŸ†”75933838

Aider leaderboard has been updated with @OpenAI GPT-5 scores https://t.co/VUBSBHqLbQ

Media 1
πŸ–ΌοΈ Media
πŸ”johnrobinsn retweeted
M
Mark Kretschmann
@mark_k
πŸ“…
Sep 03, 2025
235d ago
πŸ†”75933838

Aider leaderboard has been updated with @OpenAI GPT-5 scores https://t.co/VUBSBHqLbQ

Media 1
❀️230
likes
πŸ”23
retweets
πŸ–ΌοΈ Media
P
pcuenq
@pcuenq
πŸ“…
Sep 02, 2025
236d ago
πŸ†”14115963

Ahead-of-time compilation comes to ZeroGPU, resulting in 1.7x speedup for compute-bound tasks (such as diffusion models). It's a perfect time to try and make those H200 go brrr https://t.co/vgCOP0CFnE

Media 1
πŸ–ΌοΈ Media
G
ggerganov
@ggerganov
πŸ“…
Sep 03, 2025
235d ago
πŸ†”73677959

VS Code adds support for custom OAI-compatible endpoints This a big win for local AI as it allows us to use any local model provider without vendor lock-in. Big thanks to the VS Code devs and especially @IsidorN for listening to the community feedback and adding this option! https://t.co/3aFawjtWwM

Media 1
πŸ–ΌοΈ Media
A
alxfazio
@alxfazio
πŸ“…
Sep 02, 2025
236d ago
πŸ†”51915011

oh lord, how did i not think of this before? giving claude ast-grep for code searches and refactors has turned it into an unstoppable coding monster https://t.co/ASsi4CEgGK

Media 1
πŸ–ΌοΈ Media
J
jeremyphoward
@jeremyphoward
πŸ“…
Sep 04, 2025
234d ago
πŸ†”82565384

Guess who was the 1st to point out that Adam can be used for pretty much everything? (Answer: it was @fastdotai back in 2018 -- @GuggerSylvain's first research project in fact!) https://t.co/dzbvL2Qta8 https://t.co/usIRQt1hUg

@percyliang β€’ Thu Sep 04 16:59

We did a very careful study of 10 optimizers with no horse in the race. Despite all the excitement about Muon, Mars, Kron, Soap, etc., at the end of the day, if you tune the hyperparameters rigorously and scale up, the speedup over AdamW diminishes to only 10% :-( Experiments a

Media 1Media 2
πŸ–ΌοΈ Media
_
__tinygrad__
@__tinygrad__
πŸ“…
Sep 04, 2025
234d ago
πŸ†”46818722

Protip: if it's in a golden case, it's hype and not a real product. 4 overpriced crammed GPUs that throttle at 300W vs 4 GPUs running < 80C at 600W in tinybox green v2. https://t.co/ebrMfo9JEJ

Media 1
πŸ–ΌοΈ Media
S
sam_paech
@sam_paech
πŸ“…
Sep 06, 2025
232d ago
πŸ†”22871889

Recent models tested on longform writing: Sonoma-sky-alpha: cloaked model appears to be grok. Qwen3-max: has the same long context degradation issue as qwen3-235b: converges on super short 2-3 word paragraphs. Kimi-k2-0905: slightly worse than k2 but ~within margin of error https://t.co/3zsRuQwc6m

Media 1Media 2
+1 more
πŸ–ΌοΈ Media
H
HamelHusain
@HamelHusain
πŸ“…
Sep 06, 2025
232d ago
πŸ†”07400893

https://t.co/YsOlgQhURn

Media 1
πŸ–ΌοΈ Media
πŸ”johnrobinsn retweeted
H
Hamel Husain
@HamelHusain
πŸ“…
Sep 06, 2025
232d ago
πŸ†”07400893

https://t.co/YsOlgQhURn

Media 1
❀️56
likes
πŸ”3
retweets
πŸ–ΌοΈ Media
X
xunhuang1995
@xunhuang1995
πŸ“…
Jun 09, 2025
321d ago
πŸ†”74275059

Real-time video generation is finally real β€” without sacrificing quality. Introducing Self-Forcing, a new paradigm for training autoregressive diffusion models. The key to high quality? Simulate the inference process during training by unrolling transformers with KV caching.

πŸ–ΌοΈ Media
_
_akhaliq
@_akhaliq
πŸ“…
Mar 13, 2025
409d ago
πŸ†”70586262

Block Diffusion Interpolating Between Autoregressive and Diffusion Language Models https://t.co/CSgo0NlxfT

πŸ–ΌοΈ Media
πŸ”johnrobinsn retweeted
_
AK
@_akhaliq
πŸ“…
Mar 13, 2025
409d ago
πŸ†”70586262

Block Diffusion Interpolating Between Autoregressive and Diffusion Language Models https://t.co/CSgo0NlxfT

❀️1,594
likes
πŸ”251
retweets
πŸ–ΌοΈ Media
G
geneweng
@geneweng
πŸ“…
Sep 05, 2025
233d ago
πŸ†”49107997

GEPA: Reflective Prompt Evolution Can Outperform Reinforcement Learning https://t.co/5fxnQcbRuo via @YouTube

Media 1
πŸ–ΌοΈ Media