Your curated collection of saved posts and media

Showing 30 posts ยท last 7 days ยท newest first
D
dair_ai
@dair_ai
๐Ÿ“…
Dec 15, 2025
107d ago
๐Ÿ†”55325806

Adaptive retrieval is the way to go! And this RouteRAG paper shows why. Let's talk about it: RAG systems have a retrieval problem. The default approach to multi-hop reasoning today relies on fixed retrieval pipelines. It typically involves fetching text + maybe graph data, and hope everything is retrieved in one-shot. But the reality is that complex questions and real-world tasks require adaptive retrieval. Sometimes you need text. Sometimes you need relational structure from a graph. Sometimes it will be great to use both. And it's not secrete that graph retrieval is expensive, so retrieving it unnecessarily wastes compute. This new research introduces RouteRAG, an RL-based framework that teaches LLMs to make adaptive retrieval decisions during reasoning. When to retrieve, what source to retrieve from, and when to stop. The model learns a unified generation policy through two-stage training. > Stage 1 optimizes for answer correctness, establishing reasoning capability. > Stage 2 adds an efficiency reward that discourages unnecessary retrieval, teaching the model to balance accuracy against computational cost. The action space includes three retrieval modes: passage-only, graph-only, or hybrid. The model dynamically selects based on evolving query needs. Text retrieval works well for simple questions. Graph retrieval shines for multi-hop reasoning. The policy learns when each is appropriate. Results across five QA benchmarks: RouteRAG-7B achieves 60.6 average F1, outperforming Search-R1 (56.8 F1) despite being trained on only 10k examples versus 170k. On multi-hop datasets like 2Wiki, it reaches 64.6 F1 compared to 58.9 for Search-R1. The efficiency gains are also substantial. RouteRAG-7B reduces average retrieval turns by 20% compared to training without the efficiency reward, while actually improving accuracy by 1.1 F1 points. So we get best of both worlds: fewer retrieval calls and better answers. And here is something exciting: Small models also approach large model performance. RouteRAG with Qwen2.5-3B surpasses several graph-based RAG systems built on GPT-4o-mini, suggesting that improving the retrieval policy can be as impactful as scaling the backbone. Teaching models when and what to retrieve through RL yields more efficient and accurate multi-hop reasoning than scaling training data or model size alone. Paper: https://t.co/a4J6oAX0GC Learn to build RAG and effective AI Agents in our academy: https://t.co/zQXQt0PMbG

Media 1Media 2
๐Ÿ–ผ๏ธ Media
N
nummanthinks
@nummanthinks
๐Ÿ“…
Dec 14, 2025
107d ago
๐Ÿ†”59562907

OpenSkills v1.3.0 is out ๐Ÿš€ The Universal Skills loader for AI Coding Agents Now you can: โ€ข Use Symlinks with your skills โ€ข Install skills from local paths & private git repos โ€ข Sync to any .md file (--output flag) โ€ข Run fully headless in CI/CD (--yes) npm i -g openskills https://t.co/Rt0le0Akxy

Media 1
๐Ÿ–ผ๏ธ Media
S
SpirosMargaris
@SpirosMargaris
๐Ÿ“…
Dec 15, 2025
107d ago
๐Ÿ†”87072141

Jamie Dimon says soft skills like emotional intelligence and communication are vital as AI eliminates roles https://t.co/XBm2hITjVp @jpmorgan @fortunemagazine

Media 1
๐Ÿ–ผ๏ธ Media
S
SpirosMargaris
@SpirosMargaris
๐Ÿ“…
Dec 15, 2025
107d ago
๐Ÿ†”65762837

At this small buyout firm, talking about AI for cost-cutting is off-limits https://t.co/tOUS8W7n7y @nicollsanddimes @businessinsider

Media 1
๐Ÿ–ผ๏ธ Media
S
SpirosMargaris
@SpirosMargaris
๐Ÿ“…
Dec 15, 2025
107d ago
๐Ÿ†”16335977

After 3 tech layoffs, I knew I had to lean into being a founder https://t.co/Bp0QhYDrKu @TimSParadis @businessinsider

Media 1
๐Ÿ–ผ๏ธ Media
O
omarsar0
@omarsar0
๐Ÿ“…
Dec 15, 2025
107d ago
๐Ÿ†”13850055

DeepCode: Open Agentic Coding AI coding agents still can't reliably turn research papers into working code. The best LLM agents achieve only 42% replication scores on scientific papers, while human PhD experts hit 72%. But the problem isn't model capability. This new paper suggests it might have to do with information management. It introduces DeepCode, an open agentic coding framework that treats repository synthesis as a channel optimization problem, maximizing task-relevant signals under finite context budgets. How does this work? Scientific papers are high-entropy specifications with scattered multimodal constraints, equations, pseudocode, and hyperparameters. Naive approaches that concatenate raw documents with growing code history cause channel saturation, where redundant tokens mask critical algorithmic details and signal-to-noise ratio collapses. DeepCode addresses this through four orchestrated information operations: 1) Source compression via blueprint distillation: A planning agent transforms unstructured papers into structured implementation blueprints with file hierarchies, component specifications, and verification protocols. 2) Structured indexing using stateful code memory: As files are generated, the system maintains compact memory entries of the evolving codebase to preserve cross-file consistency without context saturation. 3) Conditional knowledge injection via RAG: The system bridges implicit specification gaps by pulling standard implementation patterns from external knowledge bases. 4) Closed-loop error correction: A validation agent treats runtime execution feedback as corrective signals to identify and fix bugs iteratively. The results on OpenAI's PaperBench benchmark are impressive. DeepCode achieves 73.5% replication score, a 70% relative improvement over the best LLM agent baseline (o1 at 43.3%). It decisively outperforms commercial agents: Cursor at 58.4%, Claude Code at 58.7%, and Codex at 40.0%. Most notably, DeepCode surpasses human experts. On a 3-paper subset evaluated by ML PhD students from Berkeley, Cambridge, and Carnegie Mellon, humans scored 72.4%. DeepCode scored 75.9%. Principled information-flow management yields significantly larger performance gains than merely scaling model size or context length. The framework is fully open source. Paper: https://t.co/LXVKsxOXfi Learn to build effective AI Agents here: https://t.co/JBU5beIoD0

Media 1Media 2
๐Ÿ–ผ๏ธ Media
S
sailorsocialist
@sailorsocialist
๐Ÿ“…
Dec 14, 2025
107d ago
๐Ÿ†”46171134

Sign if you know/remember https://t.co/bfJo4lplht

Media 1
๐Ÿ–ผ๏ธ Media
๐Ÿ”youwouldntpost retweeted
S
knife wife
@sailorsocialist
๐Ÿ“…
Dec 14, 2025
107d ago
๐Ÿ†”46171134

Sign if you know/remember https://t.co/bfJo4lplht

Media 1
โค๏ธ28
likes
๐Ÿ–ผ๏ธ Media
C
cb_doge
@cb_doge
๐Ÿ“…
Dec 14, 2025
107d ago
๐Ÿ†”69835302

Make sure to set up the all new ๐• and Grok widgets on your lock screen for instant access. https://t.co/tkkryFdh3t

Media 1
๐Ÿ–ผ๏ธ Media
B
BehizyTweets
@BehizyTweets
๐Ÿ“…
Dec 15, 2025
107d ago
๐Ÿ†”05237748

WOW! An absolute ocean of Chileans have flooded the streets to celebrate the end of socialist rule in their country. A capitalist revolution is ensuing all over the Americas. https://t.co/B5RGsIY2AM

๐Ÿ–ผ๏ธ Media
U
UniqueKhaddy_0x
@UniqueKhaddy_0x
๐Ÿ“…
Dec 09, 2025
112d ago
๐Ÿ†”64269828

@Wizarab10 Ohh!! Oh babyyyyy Sir Dickson you like Native very well oo Do you wear Jeans and shorts? https://t.co/YoCGEmfj1O

Media 1
๐Ÿ–ผ๏ธ Media
S
SpirosMargaris
@SpirosMargaris
๐Ÿ“…
Dec 15, 2025
107d ago
๐Ÿ†”04429715

Amazon pulls AI recap from Fallout TV show after it made several mistakes https://t.co/PYvJIJuqZW @liv_mcmahon @bbctech @bbc

Media 1
๐Ÿ–ผ๏ธ Media
S
SpirosMargaris
@SpirosMargaris
๐Ÿ“…
Dec 15, 2025
107d ago
๐Ÿ†”04120228

With AI, @MIT researchers teach a robot to build furniture by just asking https://t.co/YVPz0y1ILE @therobotreport https://t.co/Pgnft1Unj4

๐Ÿ–ผ๏ธ Media
H
HitFactoryPod
@HitFactoryPod
๐Ÿ“…
Dec 14, 2025
107d ago
๐Ÿ†”84804485

@youwouldntpost it begins... https://t.co/jcrQiTbVbS

Media 1
๐Ÿ–ผ๏ธ Media
F
fadouce
@fadouce
๐Ÿ“…
Dec 14, 2025
107d ago
๐Ÿ†”15260498

๐Ÿ‡ช๐Ÿ‡บOn vote pour cette couvโ€™๐Ÿ™Œ๐Ÿผ!! Merci @roxannevarza @ClaraChappaz et ๐Ÿ‘๐Ÿผ@itsrwanman pour ce matching! Story ยซย Thankfully we have @demishassabis in the photo.ย ยป & @ClementDelangue @julien_c @Thom_Wolf @arthurmensh @ptrkprz @ylecun @antonosika @FabianHedin @ASMLcompany @Lovable https://t.co/u8npbbj1qG

Media 1
๐Ÿ–ผ๏ธ Media
P
PioneerWorks_
@PioneerWorks_
๐Ÿ“…
Dec 12, 2025
109d ago
๐Ÿ†”48381064

Do LLMs really understand? To cut through the noise around AI, we brought in two experts shaping the field: Yann LeCun (@ylecun) and Adam Brown of @GoogleDeepMind. Watch them in conversation with @JannaLevin about the true intelligence of artificial minds. https://t.co/LnzQGPUIwA https://t.co/f6gbwJSELF

Media 2
๐Ÿ–ผ๏ธ Media
A
anilananth
@anilananth
๐Ÿ“…
Dec 13, 2025
108d ago
๐Ÿ†”71713437

It was a joy to watch this substantive discussion. @JannaLevin for @PioneerWorks_ talking to @ylecun and Adam Brown of Google DeepMind, about many things AI! And a surprise plant in the audience toward the end shouts out questions about machine consciousness! https://t.co/gIWX6FtPv8

Media 1
๐Ÿ–ผ๏ธ Media
R
random_walker
@random_walker
๐Ÿ“…
Dec 15, 2025
107d ago
๐Ÿ†”56473841

I've never had a reason to really try Grok so I've missed out on all the insanity. It turns out that @grok thinks I'm @emollick. I'm flattered to be confused with Ethan, but also... what? How does a bug like this even arise?! https://t.co/pp6WcNXDi7

Media 1Media 2
๐Ÿ–ผ๏ธ Media
R
random_walker
@random_walker
๐Ÿ“…
Dec 15, 2025
107d ago
๐Ÿ†”52558833

Update: I asked it what my username is and answered correctly (after looking at 19 webpages?!) and now I can't reproduce the bug even in a fresh session. Oh well. You know what they say, everyone can be @emollick for 15 minutes. https://t.co/rSflSk5zT3

Media 1
๐Ÿ–ผ๏ธ Media
A
acossta
@acossta
๐Ÿ“…
Dec 15, 2025
107d ago
๐Ÿ†”15863815

Embarking on a massive redesign for @braingridai with the new brand. @braingridai crafted a plan, broke down the tasks, and provided detailed prompts for each one. Now, Claude code is pulling tasks one by one and building them ๐Ÿ’ช๐Ÿผ https://t.co/XzlSxsRCu3

Media 1
๐Ÿ–ผ๏ธ Media
L
LBC
@LBC
๐Ÿ“…
Dec 10, 2025
111d ago
๐Ÿ†”48586677

UK charities brace for ยฃ300m Christmas donation slump, with childrenโ€™s and homelessness charities among worst affected https://t.co/BB4KQZnF0J

Media 1
๐Ÿ–ผ๏ธ Media
T
thenativehare
@thenativehare
๐Ÿ“…
Dec 10, 2025
111d ago
๐Ÿ†”26554685

Let them suffer. Provide directly to your own communities, these charlatans just redirect your money without your knowing. Follow the trail, charity starts at home. https://t.co/CaQVsdqFJi

@LBC โ€ข Wed Dec 10 13:23

UK charities brace for ยฃ300m Christmas donation slump, with childrenโ€™s and homelessness charities among worst affected https://t.co/BB4KQZnF0J

Media 1Media 2
+1 more
๐Ÿ–ผ๏ธ Media
L
lucrafund
@lucrafund
๐Ÿ“…
Dec 12, 2025
109d ago
๐Ÿ†”03087206

@nathan_ga19 me and the boys banning Indians from X app https://t.co/FK4naa2SZX

Media 1
๐Ÿ–ผ๏ธ Media
L
Lucredfinance
@Lucredfinance
๐Ÿ“…
Dec 12, 2025
109d ago
๐Ÿ†”56497414

Your weekend just got lighter. ๐Ÿ’™ Upgrade what you need today and split the payments with zero stress. Lucredโ€™s got you covered. #LucredFinance #StressFreeShopping #BNPL #NigeriaFintech https://t.co/GJTLwXVfIA

Media 1
๐Ÿ–ผ๏ธ Media
N
NaderLikeLadder
@NaderLikeLadder
๐Ÿ“…
Dec 14, 2025
107d ago
๐Ÿ†”15852721

Weโ€™re hiring 2 summer interns on the NVIDIA Brev team! Apply here: https://t.co/h7iswNxJCU https://t.co/F33d8DBq1U

๐Ÿ–ผ๏ธ Media
M
MLStreetTalk
@MLStreetTalk
๐Ÿ“…
Dec 10, 2025
111d ago
๐Ÿ†”24996624

@Tim_Dettmers "in other words, AGI โ€” then that intelligence can improve itself, leading to a runaway effect. This idea comes from Oxford-based philosophers who brought these concepts to the Bay Area." Amen. By the way, David Krakauer imo understands intelligence very well (see images and our recent coverage of him). Most of the ideology around "AGI" conceives of intelligence as an abstract monomaniacal algorithm with a single goal - something which no serious cognitive science (other than connectionists, LOL) would give any credence to. As @MelMitchell1 says, it's multi scale, multi objective and multi domain. I just finished reading this book - https://t.co/9fLwfAy5Hu by Adam Becker and you would really enjoy reading chapter 2 "Machines of Loving Grace"

Media 1Media 2
+1 more
๐Ÿ–ผ๏ธ Media
J
JessePeltan
@JessePeltan
๐Ÿ“…
Dec 14, 2025
107d ago
๐Ÿ†”46607284

Exports are only 20% of China's GDP. The U.S. is 15% of that (3%). The Chinese domestic market is way bigger than people realize.

Media 1
๐Ÿ–ผ๏ธ Media
C
cb_doge
@cb_doge
๐Ÿ“…
Dec 14, 2025
107d ago
๐Ÿ†”64670882

This picture was taken on Mars this week. https://t.co/Dd0KM8z4yC

Media 1
๐Ÿ–ผ๏ธ Media
S
SakanaAILabs
@SakanaAILabs
๐Ÿ“…
Dec 14, 2025
107d ago
๐Ÿ†”77079325

็”ŸๆˆAIใฏใ€Œไบบใฎๆˆ้•ทใ‚’ๅŠ ้€Ÿใ•ใ›ใ‚‹ใƒ‘ใƒผใƒˆใƒŠใƒผใ€ใซใชใ‚Œใ‚‹ใ‹ Google Cloudๆง˜ๆไพ›ใฎใ€Žๆ—ฅ็ตŒใƒ“ใ‚ธใƒใ‚นใ€่จ˜ไบ‹ใซใฆใ€Sakana AI CEO David Ha๏ผˆ@hardmaru๏ผ‰ใŒๅ‚ๅŠ ใ—ใŸใ‚คใƒณใ‚ฟใƒ“ใƒฅใƒผใŒๅ…ฌ้–‹ใ•ใ‚Œใพใ—ใŸใ€‚ๆฅญๅ‹™ใซ้ฉใ—ใŸAIใƒขใƒ‡ใƒซใฎ็ต„ใฟๅˆใ‚ใ›ๆ–นใ‚„ใ€ๆ—ฅๆœฌไผๆฅญใฎๅผทใฟใ‚’ๆดปใ‹ใ™AIๅฐŽๅ…ฅใซใคใ„ใฆ่ชžใ‚Šใพใ—ใŸใ€‚ ็ต„็น”ๆ–‡ๅŒ–ใซๅฏ„ใ‚Šๆทปใ†AIใ‚’ ใ€Œๆ—ฅๆœฌใฎๅคงๆ‰‹ไผๆฅญใฎไปŠใฎใ‚ใ‚Šๆ–นใฏใ€ไฝ•ๅๅนดใ‚‚ใฎๆ™‚้–“ใ‚’ใ‹ใ‘ใฆๅ‡บๆฅไธŠใŒใฃใฆใใŸใ‚‚ใฎใงใ™ใ€‚ใใฎ็ต„็น”ใฎๅฝขใ‚„ๅƒใๆ–นใซใชใฃใŸใ“ใจใซใฏใ€ใใ‚Œใชใ‚Šใฎ็†็”ฑใŒใ‚ใ‚Šใพใ™ใ€‚ใใ‚Œใซๅฏพใ—ใฆใ€Œ็ต„็น”ใ‚’ใƒ•ใƒฉใƒƒใƒˆใซใ€ใจ่ฆๆฑ‚ใ™ใ‚‹ใฎใฏ้žๅˆ็†็š„ใงใ™ใ€‚ใ‚€ใ—ใ‚ใ€AIใŒใใ“ใ‚’ๆ”ฏๆดใ—ใฆไธ€็ท’ใซๅƒใใ“ใจใŒๅคงๅˆ‡ใงใ™ใ€‚ใ€ ไบบใฎๆˆ้•ทใ‚’ๅŠ ้€Ÿใ•ใ›ใ‚‹ใƒ‘ใƒผใƒˆใƒŠใƒผใ‚’็›ฎๆŒ‡ใ—ใฆ ใ€Œ็”ŸๆˆAIใฎ็›ฎๆจ™ใฎไธ€ใคใฏใ€ไบบใจใ‚ณใƒณใƒ”ใƒฅใƒผใ‚ฟใƒผใ‚’ๆฉ‹ๆธกใ—ใ™ใ‚‹Companionใซใชใ‚‹ใ“ใจใงใ™ใ€‚โ€ฆๆทฑใ„ๆฅญๅ‹™ใฎ็†่งฃใ‚’AIใŒๆŒใกใ€ไบบ้–“ใจๅŒใ˜ๆ–นๅ‘ใ‚’ๅ‘ใ„ใฆๅ•้กŒใ‚’่งฃๆฑบใ—ใฆใ„ใใ‚ˆใ†ใซใชใฃใฆใ„ใใ“ใจใŒๆฌกใฎๆŒ‘ๆˆฆใ ใจๆ€ใ„ใพใ™ใ€‚ใ€ Sakana AIใฎใƒŸใƒƒใ‚ทใƒงใƒณ ใ€ŒSakana AIใฏAIใƒขใƒ‡ใƒซใ‚„AIใ‚จใƒผใ‚ธใ‚งใƒณใƒˆใฎ็ ”็ฉถ้–‹็™บใซๅ–ใ‚Š็ต„ใ‚€ใ‚นใ‚ฟใƒผใƒˆใ‚ขใƒƒใƒ—ใงใ™ใ€‚ใ„ใพๅ–ใ‚Š็ต„ใ‚“ใงใ„ใ‚‹ใ“ใจใฏ2ใคใ‚ใ‚Šใพใ™ใ€‚ใ‚จใƒใƒซใ‚ฎใƒผๅŠน็އใŒ้ซ˜ใ„ๆ—ฅๆœฌ็™บใฎAIใƒขใƒ‡ใƒซใฎ้–‹็™บใจใ€ใ„ใใคใ‚‚ใฎAIใƒขใƒ‡ใƒซใ‚’ใ†ใพใไฝฟใ„ๅˆ†ใ‘ใฆ้ซ˜ๅบฆใชใƒ—ใƒญใ‚ปใ‚นใ‚’่‡ชๅ‹•ๅฎŸ่กŒใงใใ‚‹ใ‚ˆใ†ใชAIใ‚จใƒผใ‚ธใ‚งใƒณใƒˆใฎ้–‹็™บใงใ™ใ€‚ใ€ AIใƒขใƒ‡ใƒซใ‚’ไฝฟใ„ๅˆ†ใ‘ใ€็ต„ใฟๅˆใ‚ใ›ใ‚‹ ใ€ŒAIใƒขใƒ‡ใƒซใซใ‚ˆใฃใฆๅผทใ„ใจใ“ใ‚ใจๅผฑใ„ใจใ“ใ‚ใŒ้•ใฃใฆใ„ใพใ™ใญใ€‚ใ ใ‹ใ‚‰็งใŸใกใฏใ€่ค‡ๆ•ฐใฎAIใ‚’็ต„ใฟๅˆใ‚ใ›ใฆๆฑ‚ใ‚ใ‚‹ๆฉŸ่ƒฝใ‚’ๅฎŸ็พใ™ใ‚‹ใƒžใƒซใƒใƒ•ใ‚กใ‚ฆใƒณใƒ‡ใƒผใ‚ทใƒงใƒณใƒขใƒ‡ใƒซใ‚’ๆๅ”ฑใ—ใฆใ„ใพใ™ใ€‚โ€ฆไผๆฅญใฎไบ‹ๆฅญใƒ‰ใƒกใ‚คใƒณใ‚„้ฉ็”จใ™ใ‚‹ๆฅญๅ‹™ใซใ‚ˆใฃใฆ้ธใถในใ็”ŸๆˆAIใฏๅค‰ใ‚ใ‚Šใพใ™ใ€‚ๅˆฉ็”จ่€…ใฎๅฟ—ๅ‘ใ‚„ๆฅญๅ‹™ใƒŽใ‚ฆใƒใ‚ฆใ‚’็†่งฃใ—ใŸใ†ใˆใงใ€AIใƒขใƒ‡ใƒซใ‚’้ธใณใ€ใใ—ใฆใƒขใƒ‡ใƒซใŒๅ‡บใ—ใฆใใ‚‹ๅ›ž็ญ”ใฎ้ฉๅˆ‡ใ•ใ‚’่ฉ•ไพกใ™ใ‚‹ๅฟ…่ฆใŒใ‚ใ‚Šใพใ™ใ€‚ใใฎใƒ—ใƒญใ‚ปใ‚นใ“ใใŒใ€AIๅฐŽๅ…ฅใฎๆˆๅŠŸใ‚’ๅทฆๅณใ™ใ‚‹้ตใซใชใ‚Šใพใ™ใ€‚ใ€ ๅ…จๆ–‡ใฏใ“ใกใ‚‰๏ผšhttps://t.co/RSSy8bFbLa

Media 1Media 2
๐Ÿ–ผ๏ธ Media
H
hardmaru
@hardmaru
๐Ÿ“…
Dec 14, 2025
107d ago
๐Ÿ†”20660298

Applied Research Engineer https://t.co/FuEoI2xrzS

Media 1
๐Ÿ–ผ๏ธ Media