Your curated collection of saved posts and media

Showing 24 posts ยท last 30 days ยท by score
S
steipete
@steipete
๐Ÿ“…
Mar 05, 2026
5d ago
๐Ÿ†”99042332

TIL: There's a whole bunch of interesting skills in the oss codex repo: https://t.co/gNFHV3MD2j $skill-installer playwright-interactive (also /fast is sweeeeet, 1.5x codex makes a huge diff!) https://t.co/XTENPuZ9Ie

Media 1Media 2
๐Ÿ–ผ๏ธ Media
L
LiorOnAI
@LiorOnAI
๐Ÿ“…
Mar 02, 2026
8d ago
๐Ÿ†”52031145
โญ0.42

Someone just bypassed Apple's Neural Engine to train models. The Neural Engine inside every M-series Mac was designed for inference. Run models, don't train them. No public API, no documentation, and certainly no backpropagation. A researcher reverse-engineered the private APIs anyway and built a transformer training loop that runs forward and backward passes directly on the ANE hardware. The method bypasses CoreML entirely. Instead of using Apple's official tools, the project constructs programs in MIL (Model Intermediate Language), compiles them in-memory using undocumented `_ANEClient` APIs, and feeds data through IOSurface shared memory buffers. Weights get baked into the compiled programs as constants. E ach training step dispatches six custom kernels: attention forward, feedforward forward, then four backward passes that compute gradients with respect to inputs. Weight gradients still run on the CPU using Accelerate's matrix libraries, but the heavy lifting (matrix multiplies, softmax, activation functions) happens on the ANE. This makes three things possible that weren't before: 1. Training small models locally without burning through your battery 2. Fine-tuning on-device without sending data to a server or spinning up the GPU 3. Research into what the ANE hardware can actually do when you ignore Apple's guardrails If this approach scales, the next wave of on-device AI stops being about running someone else's frozen model.

O
omarsar0
@omarsar0
๐Ÿ“…
Mar 03, 2026
7d ago
๐Ÿ†”96343923

Can AI agents agree? Communication is one of the biggest challenges in multi-agent systems. New research tests LLM-based agents on Byzantine consensus games, scenarios where agents must agree on a value even when some participants behave adversarially. The main finding: valid agreement is unreliable even in fully benign settings, and degrades further as group size grows. Most failures come from convergence stalls and timeouts, not subtle value corruption. Why does it matter? Multi-agent systems are being deployed in high-stakes coordination tasks. This paper is an early signal that reliable consensus is not an emergent property you can assume. It needs to be designed explicitly. Paper: https://t.co/3fllhchiKX Learn to build effective AI agents in our academy: https://t.co/1e8RZKs4uX

Media 1
๐Ÿ–ผ๏ธ Media
O
omarsar0
@omarsar0
๐Ÿ“…
Mar 03, 2026
7d ago
๐Ÿ†”22674842

MCP is dead? What are your thoughts? I mostly use Skills and CLI lately. I still use a few MCP tools for orchestrating agents more efficiently. https://t.co/o6saSxNQ9s

Media 1
๐Ÿ–ผ๏ธ Media
๐Ÿ”dair_ai retweeted
O
elvis
@omarsar0
๐Ÿ“…
Mar 03, 2026
7d ago
๐Ÿ†”22674842
โญ0.32

MCP is dead? What are your thoughts? I mostly use Skills and CLI lately. I still use a few MCP tools for orchestrating agents more efficiently. https://t.co/o6saSxNQ9s

โค๏ธ239
likes
๐Ÿ”17
retweets
P
ph_singer
@ph_singer
๐Ÿ“…
Feb 26, 2026
12d ago
๐Ÿ†”83063300
โญ0.38

@alex_prompter Without opening the paper, how did they gather the ground truth? My naive assumption is if they are able to gather the ground truth, it is somewhere out there.

D
deedydas
@deedydas
๐Ÿ“…
Mar 02, 2026
8d ago
๐Ÿ†”31435114

Narrative violation. Cursor goes $1B to $2B in 3mos. Claude Code went $0 to $2.5B in 8mos. Everyone in the tech/X bubble think people are wholesale ditching Cursor, but enterprise diffusion is glacial. Most of the world just got a hold of it. https://t.co/7RBU7mvosz

Media 1
๐Ÿ–ผ๏ธ Media
๐Ÿ”random_walker retweeted
D
Deedy
@deedydas
๐Ÿ“…
Mar 02, 2026
8d ago
๐Ÿ†”31435114
โญ0.34

Narrative violation. Cursor goes $1B to $2B in 3mos. Claude Code went $0 to $2.5B in 8mos. Everyone in the tech/X bubble think people are wholesale ditching Cursor, but enterprise diffusion is glacial. Most of the world just got a hold of it. https://t.co/7RBU7mvosz

โค๏ธ1,118
likes
๐Ÿ”54
retweets
L
llama_index
@llama_index
๐Ÿ“…
Mar 04, 2026
6d ago
๐Ÿ†”06983152

If you need to split complex or composite documents into structured categories or sections, LlamaSplit is built for the job โœ‚๏ธ With the intuitive UI, you can: โ€ขDefine a custom configuration for how your documents should be categorized โ€ขSpecify the exact sections or impact types you want extracted โ€ขRun the job and explore the results through an interactive interface ๐Ÿ” In this walkthrough, @itsclelia demonstrates how to configure LlamaSplit to break down Environmental Impact Reports into clearly defined impact categories ๐ŸŒณ ๐ŸŽฅ Watch the full video here: ๐Ÿ“˜ Or get started right away with the docs (UI + code examples): https://t.co/kAMUqwOCDW

๐Ÿ–ผ๏ธ Media
M
Modular
@Modular
๐Ÿ“…
Mar 02, 2026
8d ago
๐Ÿ†”96696317

Building for the AI era means rethinking the stack from the ground up. Modular co-founder and CEO @clattner_llvm joined @shanselman on @Hanselminutes to talk about Mojo ๐Ÿ”ฅ, heterogeneous compute, and why AI infrastructure demands new abstractions. Watch hereโ†“ https://t.co/AKCJQEoKNJ

Media 1
๐Ÿ–ผ๏ธ Media
B
braingridai
@braingridai
๐Ÿ“…
Mar 03, 2026
7d ago
๐Ÿ†”97137307

We just shipped Designs. Here's the problem it solves: most UI work fails because you don't know what it should look like until after your coding agent already built it wrong. You describe a dashboard. The agent builds it. You realize the layout doesn't work. You prompt again. The agent rebuilds. Something else breaks. Three iterations later you're debugging CSS instead of shipping features. Designs puts the iteration where it belongs, before a single line of code gets written. BrainGrid now generates actual UI designs for your requirements. You can iterate on them with the agent, annotate what needs to change, select specific elements to tweak. Once you lock it in, that design becomes part of the requirement that gets handed to your coding tool. No more building the wrong UI three times because you couldn't visualize it from a text prompt. It works with new apps and existing ones. If you're adding a feature to something you've already built, BrainGrid matches your existing app's look and feel so the new design doesn't feel bolted on. The designs get included in your Requirements doc when you fetch from CLI or MCP. Your coding agent knows exactly what to build. This is the part most builders skip, and it's why UI work takes twice as long as it should. Now you can see it, fix it, and lock it before the agent touches your codebase.

Media 1
๐Ÿ–ผ๏ธ Media
A
acossta
@acossta
๐Ÿ“…
Mar 06, 2026
5d ago
๐Ÿ†”52968508
โญ0.34

Tackling a really gnarly issue. Played Codex 5.4 and Opus 4.6 side by side. Codex came back relatively fast with a diagnostic/solution that seemed feasible Opus took a while, a long one, and came back with the correct root cause.

A
AskPerplexity
@AskPerplexity
๐Ÿ“…
Mar 06, 2026
5d ago
๐Ÿ†”84600043

You can now run three frontier models at once and select your orchestrator model directly inside Perplexity Computer. Model Council automatically runs GPT-5.4, Claude Opus 4.6 and Gemini 3.1 Pro simultaneously. Three frontier models. One workflow. Best answer wins. https://t.co/40rPcXpr6s

๐Ÿ–ผ๏ธ Media
E
emollick
@emollick
๐Ÿ“…
Mar 05, 2026
5d ago
๐Ÿ†”30916497
โญ0.38

It is increasingly clear that the constraint is going to be compute, just as the AI labs warned. The token requirements for agentic work are high, making using frontier agents only cost efficient for high value tasks. There are tons of other opportunities waiting for lower costs

E
emollick
@emollick
๐Ÿ“…
Mar 05, 2026
5d ago
๐Ÿ†”03018817

Had early access to GPT-5.4 and Pro. They are very good. One fun illustration of progress, this is the same prompt I used in GPT-4 below (making a 3D space inspired by Piranesi) now in GPT-5.4 Pro. There were no errors, made in a single prompt plus one to "make it better." https://t.co/7Vgmc60SKc

๐Ÿ–ผ๏ธ Media
E
emollick
@emollick
๐Ÿ“…
Mar 05, 2026
5d ago
๐Ÿ†”51220362
โญ0.40

To clarify: Gemini Deep Think is a really smart model, but it doesn't have access to the same tools as Claude or ChatGPT - it can't download files, cannot consistently run code on its own, cannot produce downloadable files, does not clearly show when it does web search, etc

B
BoWang87
@BoWang87
๐Ÿ“…
Mar 05, 2026
5d ago
๐Ÿ†”78072654

Two major AI releases this week: โ€ข Qwen3.5 โ€” new open-source small models โ€ข GPT-5.4 โ€” newest frontier closed model Most benchmarks compare math and coding. But the real test for frontier AI should be biology and healthcare. Thatโ€™s where mistakes actually matter. So our team at @UHN ran them on EURORAD โ€” 207 expert-validated radiology differential diagnosis cases. Results: GPT-5.4: 92.2% Qwen3.5-27B: 85% Gemini 3.1 Pro: ~79% A 27B open model that runs on a laptop is only 7 points behind the most powerful AI model on earth โ€” and already beating Gemini on this benchmark. That gap is much smaller than people expected. And it matters. For years hospitals faced an impossible tradeoff: Frontier models โ†’ patient data leaves the hospital Local models โ†’ not good enough That tradeoff may finally be ending. Qwen3.5-27B runs fully local. No API. No cloud. No patient data leaving the building. HIPAA / PHIPA compliance becomes architecture, not paperwork. Interesting detail: 27B and 122B score almost identically here. Scaling bigger didnโ€™t help much. One caveat: with web-scale training, itโ€™s hard to completely rule out that frontier models like GPT-5.4 may have seen parts of evaluation datasets. Still, the signal is clear: Small models are getting good enough for real clinical AI. And if we want to measure real AI progress, biology and healthcare should be the benchmark. Huge credit to the team @alifmunim @AlhusainAbdalla @JunMa_AI4Health @Omar_Ibr12 @oliviaamwei

Media 1
๐Ÿ–ผ๏ธ Media
๐Ÿ”iScienceLuvr retweeted
B
Bo Wang
@BoWang87
๐Ÿ“…
Mar 05, 2026
5d ago
๐Ÿ†”78072654
โญ0.38

Two major AI releases this week: โ€ข Qwen3.5 โ€” new open-source small models โ€ข GPT-5.4 โ€” newest frontier closed model Most benchmarks compare math and coding. But the real test for frontier AI should be biology and healthcare. Thatโ€™s where mistakes actually matter. So our team at @UHN ran them on EURORAD โ€” 207 expert-validated radiology differential diagnosis cases. Results: GPT-5.4: 92.2% Qwen3.5-27B: 85% Gemini 3.1 Pro: ~79% A 27B open model that runs on a laptop is only 7 points behind the most powerful AI model on earth โ€” and already beating Gemini on this benchmark. That gap is much smaller than people expected. And it matters. For years hospitals faced an impossible tradeoff: Frontier models โ†’ patient data leaves the hospital Local models โ†’ not good enough That tradeoff may finally be ending. Qwen3.5-27B runs fully local. No API. No cloud. No patient data leaving the building. HIPAA / PHIPA compliance becomes architecture, not paperwork. Interesting detail: 27B and 122B score almost identically here. Scaling bigger didnโ€™t help much. One caveat: with web-scale training, itโ€™s hard to completely rule out that frontier models like GPT-5.4 may have seen parts of evaluation datasets. Still, the signal is clear: Small models are getting good enough for real clinical AI. And if we want to measure real AI progress, biology and healthcare should be the benchmark. Huge credit to the team @alifmunim @AlhusainAbdalla @JunMa_AI4Health @Omar_Ibr12 @oliviaamwei

โค๏ธ222
likes
๐Ÿ”30
retweets
B
BlancheMinerva
@BlancheMinerva
๐Ÿ“…
Mar 02, 2026
9d ago
๐Ÿ†”73443869
โญ0.38

@CFGeek @xeophon Third party audits detected an irregularity in a Pythia model that had gone undetected in 2 years. https://t.co/pXefsdzIkm

B
BlancheMinerva
@BlancheMinerva
๐Ÿ“…
Mar 02, 2026
9d ago
๐Ÿ†”52851348
โญ0.36

@CFGeek @xeophon Actually it was their second paper: https://t.co/7FO6BwP9Mz

T
therealoliulv
@therealoliulv
๐Ÿ“…
Mar 02, 2026
9d ago
๐Ÿ†”78066456

WE WON THE @MistralAI LONDON HACKATHON ๐Ÿ‡ฌ๐Ÿ‡ง๐Ÿ‡ซ๐Ÿ‡ท We made Mistralverse, here's our demo vid. @HarryStebbings who says the UK isn't shipping?? https://t.co/lVWr43XkNj

๐Ÿ–ผ๏ธ Media
๐Ÿ”Scobleizer retweeted
T
Oliver Ulvebne
@therealoliulv
๐Ÿ“…
Mar 02, 2026
9d ago
๐Ÿ†”78066456
โญ0.34

WE WON THE @MistralAI LONDON HACKATHON ๐Ÿ‡ฌ๐Ÿ‡ง๐Ÿ‡ซ๐Ÿ‡ท We made Mistralverse, here's our demo vid. @HarryStebbings who says the UK isn't shipping?? https://t.co/lVWr43XkNj

โค๏ธ94
likes
๐Ÿ”8
retweets
V
Vassivasss
@Vassivasss
๐Ÿ“…
Mar 02, 2026
8d ago
๐Ÿ†”73523954

What if AI could see the world the way we do? Thatโ€™s the idea we bet our weekend on at the Mistral Worldwide Hackathon. With @haaspierre_ and Arman Artola-Zanganeh, we built ๐—ฃ๐—ผ๐—ฟ๐˜:๐—ช๐—ผ๐—ฟ๐—น๐—ฑ๐ŸŒ, an open-source framework that lets anyone connect their Meta glasses to any AI system. Let me take you back to saturday morning. So before knowing it could work we needed the hardware. So I ran to Rue de Rivoli and bought โ‚ฌ500 Meta glasses on the spot. If thatโ€™s not commitment, I donโ€™t know what is (a true bet). We then built non-stop for 36 hours to make it usable. End-to-end. The glasses stream what you see โ†’ the AI makes sense of it โ†’ it answers back through the glassesโ€™ speaker. And suddenly when we understood that it was going to work, the question changed. It was no longer โ€œ๐—œ๐˜€ ๐˜๐—ต๐—ถ๐˜€ ๐—ฑ๐—ผ๐—ฎ๐—ฏ๐—น๐—ฒ?โ€ It became โ€œ๐—ช๐—ต๐—ฎ๐˜ ๐—ฐ๐—ฎ๐—ป ๐—ฝ๐—ฒ๐—ผ๐—ฝ๐—น๐—ฒ ๐—ฏ๐˜‚๐—ถ๐—น๐—ฑ ๐˜„๐—ถ๐˜๐—ต ๐˜๐—ต๐—ถ๐˜€?โ€ - A plumber getting live assistance while repairing something. - A technician repairing industrial machinery. - A traveler exploring a new country. - A visually impaired person navigating space. At first, we were looking for the โ€œrightโ€ use case. Then we realized something more interesting. If AI can share your perspective, continuously, the use cases are not ours to decide. Thatโ€™s why ๐—ฃ๐—ผ๐—ฟ๐˜:๐—ช๐—ผ๐—ฟ๐—น๐—ฑ๐ŸŒ is fully open source. If you want to connect your Meta glasses, plug in your own models, customize with your own prompts, your own MCP, your Openclawโ€ฆ you can. Link to the open source repo (you can contribute and give it a little star โค๏ธ): https://t.co/UueLnkMZpM Link to the demo video: https://t.co/qcTDqKGvax Huge thanks to the organizing team of the hackathon, it was truly great. @Jthmas404

Media 1Media 2
+1 more
๐Ÿ–ผ๏ธ Media
๐Ÿ”Scobleizer retweeted
V
Vassili de Rosen
@Vassivasss
๐Ÿ“…
Mar 02, 2026
8d ago
๐Ÿ†”73523954
โญ0.34

What if AI could see the world the way we do? Thatโ€™s the idea we bet our weekend on at the Mistral Worldwide Hackathon. With @haaspierre_ and Arman Artola-Zanganeh, we built ๐—ฃ๐—ผ๐—ฟ๐˜:๐—ช๐—ผ๐—ฟ๐—น๐—ฑ๐ŸŒ, an open-source framework that lets anyone connect their Meta glasses to any AI system. Let me take you back to saturday morning. So before knowing it could work we needed the hardware. So I ran to Rue de Rivoli and bought โ‚ฌ500 Meta glasses on the spot. If thatโ€™s not commitment, I donโ€™t know what is (a true bet). We then built non-stop for 36 hours to make it usable. End-to-end. The glasses stream what you see โ†’ the AI makes sense of it โ†’ it answers back through the glassesโ€™ speaker. And suddenly when we understood that it was going to work, the question changed. It was no longer โ€œ๐—œ๐˜€ ๐˜๐—ต๐—ถ๐˜€ ๐—ฑ๐—ผ๐—ฎ๐—ฏ๐—น๐—ฒ?โ€ It became โ€œ๐—ช๐—ต๐—ฎ๐˜ ๐—ฐ๐—ฎ๐—ป ๐—ฝ๐—ฒ๐—ผ๐—ฝ๐—น๐—ฒ ๐—ฏ๐˜‚๐—ถ๐—น๐—ฑ ๐˜„๐—ถ๐˜๐—ต ๐˜๐—ต๐—ถ๐˜€?โ€ - A plumber getting live assistance while repairing something. - A technician repairing industrial machinery. - A traveler exploring a new country. - A visually impaired person navigating space. At first, we were looking for the โ€œrightโ€ use case. Then we realized something more interesting. If AI can share your perspective, continuously, the use cases are not ours to decide. Thatโ€™s why ๐—ฃ๐—ผ๐—ฟ๐˜:๐—ช๐—ผ๐—ฟ๐—น๐—ฑ๐ŸŒ is fully open source. If you want to connect your Meta glasses, plug in your own models, customize with your own prompts, your own MCP, your Openclawโ€ฆ you can. Link to the open source repo (you can contribute and give it a little star โค๏ธ): https://t.co/UueLnkMZpM Link to the demo video: https://t.co/qcTDqKGvax Huge thanks to the organizing team of the hackathon, it was truly great. @Jthmas404

โค๏ธ10
likes
๐Ÿ”5
retweets