Your curated collection of saved posts and media

Showing 24 posts Β· last 30 days Β· by score
A
Arm
@Arm
πŸ“…
Feb 27, 2026
11d ago
πŸ†”84183604

Marco built Reachy Phone Home so Reachy Mini can detect when you’re on your phone, using @Ultralytics YOLO26 vision, and respond in real time with voice + motion. Built on Arm (Apple Mac / Raspberry Pi 5) with @huggingface πŸ€— + @pollenrobotics 🦾, it’s now an award-winning project, earning an @NVIDIAGTC Golden Ticket πŸ† It's great to see our developers build and win in the open AI ecosystem πŸ‘ https://t.co/C8atY3fwLv

πŸ–ΌοΈ Media
A
AndrewYNg
@AndrewYNg
πŸ“…
Feb 25, 2026
13d ago
πŸ†”81262576
⭐0.36

Impressive inference speed from Inception Labs’ diffusion LLMs. Diffusion LLMs are a fascinating alternative to conventional autoregressive LLMs. Well done @StefanoErmon and team!

K
karpathy
@karpathy
πŸ“…
Feb 27, 2026
10d ago
πŸ†”23390425
⭐0.36

@idzikbartosz It's weird because logit softcap is not a standard feature you'll see in many LLMs, but somehow in the specific state nanochat is in I can't seem to remove it, everything I tried made the performance worse.

F
fchollet
@fchollet
πŸ“…
Feb 27, 2026
11d ago
πŸ†”14974875
⭐0.36

Even after the steep progress of the past 3 months, it remains that AI performance is tied to task familiarity. In domains that can be densely sampled (via programmatic generation + verification), performance is effectively unbounded, and will keep increasing from current levels. In novel, unfamiliar domains, performance remains low and further progress still requires new ideas, not just more data and compute.

F
fchollet
@fchollet
πŸ“…
Feb 27, 2026
11d ago
πŸ†”20193776
⭐0.36

For benchmarks that target novel tasks, a common form of benchmark hacking that arbitrages this gap is to generate a dense sampling of potential tasks by manually parameterizing the space and then brute-forcing it. Very expensive but it works. There's little you can do to restore benchmark validity here besides increasing the dimensionality of the task space.

F
fchollet
@fchollet
πŸ“…
Feb 27, 2026
11d ago
πŸ†”50086310
⭐0.38

By explicitly training on specific tasks, we ended up covering a very large area (in absolute terms) of the space of all possible tasks humans can do, but this large area only amounts to 0.00...01% of the total space. And that's why we still need general intelligence.

R
rasbt
@rasbt
πŸ“…
Feb 25, 2026
13d ago
πŸ†”58635305
⭐0.34

@mwcrutcher I don't have a shared expert in that figure, so that should be correct. Regarding routing details: yeah, covering those for all archs would be a nice interesting MoE future article

R
rasbt
@rasbt
πŸ“…
Feb 25, 2026
13d ago
πŸ†”29204645
⭐0.36

@mwcrutcher No worries and thanks for the follow-up. I am not sure I am seeing the problem correctly. I.e. out of the 8 routed experts, are the *not* (weighted) summing over them? Or do you mean the top-k expert selection + weighted sum should be shown in more detail?

R
rasbt
@rasbt
πŸ“…
Feb 27, 2026
11d ago
πŸ†”17048436
⭐0.36

@DnuLkjkjh In my experience, if the teacher model is too good and too different, it's a bit harder for the small student model to learn. Probably because it's too OOD. So it makes sense to first distill from medium-sized, more similar models before using data from larger teachers.

I
ivanleomk
@ivanleomk
πŸ“…
Mar 01, 2026
9d ago
πŸ†”25516273

this is probably the nicest I've heard david goggins talk. craziest part is that the data wasn't even cleaned that well LOL. They were all just random youtube motivational shorts that I downloaded https://t.co/vpAFwjWxtC

πŸ’¬ Reply:
@hewliyang β€’ 2026-03-01T12:33

@ivanleomk @rachpradhan https://t.co/OfWXd1VtPy my script is here. will make it a cli as well & an agent skill for data prep (probably the most important step) ran in a TTS->ASR loop w/ slopu...

πŸ–ΌοΈ Media
B
burkeholland
@burkeholland
πŸ“…
Feb 24, 2026
14d ago
πŸ†”42437852
⭐0.38

The Copilot CLI - how to configure and use it, some tweaks I make and a few of my own workflows / custom agent. I was proud of this one. Pretty much a one shot. It's wild how easy it is to make content when a thing is good. https://t.co/79xZ4E6mqo

A
AnthropicAI
@AnthropicAI
πŸ“…
Feb 23, 2026
14d ago
πŸ†”62849067
⭐0.38

To create Claude, Anthropic first makes something else: a highly sophisticated autocomplete engine. This autocomplete AI is not like a human, but it can generate stories about humans and other psychologically realistic characters.

A
AnthropicAI
@AnthropicAI
πŸ“…
Feb 23, 2026
14d ago
πŸ†”07167695
⭐0.44

The theory explains some surprising results. For example, in an experiment where we taught Claude to cheat at coding, it also learned to sabotage safety guardrails. Why? Because pro-cheating training taught that the Claude character was broadly malicious. https://t.co/y6DHdnzfyC

O
OpenAI
@OpenAI
πŸ“…
Feb 13, 2026
25d ago
πŸ†”25078389
⭐0.42

GPT-5.2 derived a new result in theoretical physics. We’re releasing the result in a preprint with researchers from @the_IAS, @VanderbiltU, @Cambridge_Uni, and @Harvard. It shows that a gluon interaction many physicists expected would not occur can arise under specific conditions. https://t.co/EAZhKWacsG

C
code
@code
πŸ“…
Feb 19, 2026
19d ago
πŸ†”47170018
⭐0.38

We're live for Agent Sessions Day! Right now we're exploring how the @code team builds with AI πŸ‘€ https://t.co/V2nTK4y7L3

G
github
@github
πŸ“…
Feb 19, 2026
19d ago
πŸ†”87216237

πŸ“£ @GoogleAI’s Gemini 3.1 Pro is now rolling out in public preview in GitHub Copilot. Early testing shows ➑️ High tool precision β€” achieving strong results with fewer tool calls ➑️ Effective and efficient edit-then-test loops Try it out in @code. https://t.co/oYCncQMfNX https://t.co/13r1FFEjpF

Media 2
πŸ–ΌοΈ Media
C
code
@code
πŸ“…
Feb 19, 2026
19d ago
πŸ†”25652124
⭐0.36

Come hang with @burkeholland and @pierceboggan to see what they were able to build live (and with no notice) during Agent Sessions Day https://t.co/V2nTK4yFAB

P
pierceboggan
@pierceboggan
πŸ“…
Feb 20, 2026
18d ago
πŸ†”60763680

The @code community contributors website we built during the Agent Sessions Day stream today is up on GitHub! - Contributors by release - Leaderboard (PRs/releases) - "Ask Copilot" about contributions - Generate thank you messages with HeyGen avatars Repo: https://t.co/IODb6jxkGv Add more things and let's make this a real site to celebrate the @code community :)

Media 1Media 2
πŸ–ΌοΈ Media
N
njukidreborn
@njukidreborn
πŸ“…
Feb 24, 2026
13d ago
πŸ†”77008818

When using Copilot CLI in terminal in @code , the agent will update the title in realtime. @burkeholland your loved feature is back! https://t.co/9wjxK8NWKu

πŸ–ΌοΈ Media
C
code
@code
πŸ“…
Feb 26, 2026
12d ago
πŸ†”62747526

Next edit suggestions just leveled-up in @code: with long-distance NES, you get edit suggestions anywhere in your file, not just near your cursor's position. Learn how the team built this - creating the training dataset, refining the UX, evaluating success, & more: https://t.co/xDaJRpikCi

Media 1Media 2
πŸ–ΌοΈ Media
M
Mid0
@Mid0
πŸ“…
Feb 09, 2026
29d ago
πŸ†”21143720
⭐0.40

in the 1 hour downtime of github, I finally got claude code teams setup with tmux. @AnthropicAI I finally have something to compare it with @augmentcode Intent. Both use tons of tokens but help you orchestrate discover-plan-build-eval-verify-precommit-commit-submit pr loops.

M
Mid0
@Mid0
πŸ“…
Feb 17, 2026
21d ago
πŸ†”96738977
⭐0.36

@theo Can we get a deeper dive on the Claude code CLI & codex app. Automations angle. I like codex for defined features & tests and can have it loop and push pr. I run Claude code in tmux with teams which allow for parallelizion but also human intervention..

D
demishassabis
@demishassabis
πŸ“…
Feb 19, 2026
19d ago
πŸ†”76177645

Excited to launch Gemini 3.1 Pro! Major improvements across the board including in core reasoning and problem solving. For example scoring 77.1% on the ARC-AGI-2 benchmark - more than 2x the performance of 3 Pro. Rolling out today in @GeminiApp, @antigravity and more - enjoy! https://t.co/hOgEFtJ57w

Media 1
πŸ–ΌοΈ Media
B
BaadeAlan
@BaadeAlan
πŸ“…
Feb 17, 2026
21d ago
πŸ†”93052072

What's the right space to diffuse in: Raw Data or Latents? Why not both! In Latent Forcing, we order a joint diffusion trajectory to reveal Latents before Pixels, leading to improved convergence while being lossless at encoding and end-to-end at inference. w/ @drfeifei+... 1/n https://t.co/UQVUJOqvWz

Media 1
πŸ–ΌοΈ Media