Your curated collection of saved posts and media

Showing 32 posts ยท last 14 days ยท by score
_
_akhaliq
@_akhaliq
๐Ÿ“…
Dec 18, 2025
101d ago
๐Ÿ†”62049784

discuss: https://t.co/g84qyZGEbl

Media 1
๐Ÿ–ผ๏ธ Media
_
_akhaliq
@_akhaliq
๐Ÿ“…
Dec 18, 2025
101d ago
๐Ÿ†”13158081

SAGE Training Smart Any-Horizon Agents for Long Video Reasoning with Reinforcement Learning https://t.co/jTjfgywf8N

Media 1
๐Ÿ–ผ๏ธ Media
_
_akhaliq
@_akhaliq
๐Ÿ“…
Dec 18, 2025
101d ago
๐Ÿ†”99668090

discuss: https://t.co/GzOUznS5mS

Media 1
๐Ÿ–ผ๏ธ Media
_
_akhaliq
@_akhaliq
๐Ÿ“…
Dec 18, 2025
101d ago
๐Ÿ†”42901556

End-to-End Training for Autoregressive Video Diffusion via Self-Resampling https://t.co/LoohHXFgTH

Media 1
๐Ÿ–ผ๏ธ Media
_
_akhaliq
@_akhaliq
๐Ÿ“…
Dec 18, 2025
101d ago
๐Ÿ†”03555166

IC-Effect Precise and Efficient Video Effects Editing via In-Context Learning https://t.co/sRRIxNJcaX

Media 1
๐Ÿ–ผ๏ธ Media
_
_akhaliq
@_akhaliq
๐Ÿ“…
Dec 18, 2025
101d ago
๐Ÿ†”32834645

Fast and Accurate Causal Parallel Decoding using Jacobi Forcing https://t.co/MDLUIfsxUn

Media 1
๐Ÿ–ผ๏ธ Media
L
lhoestq
@lhoestq
๐Ÿ“…
Dec 18, 2025
101d ago
๐Ÿ†”59687264

Wow there are now 600,000 public datasets on @huggingface ! They were 600 exactly five years ago ๐Ÿ˜ณ (@mervenoyann @abhi1thakur can testify) So, to summarize: - 2020>2025 : x1000 - 2025>2030: x1000 too ??? ๐Ÿ˜‚ The open source community is crazy ! https://t.co/UHe4nfrlbs

Media 1
๐Ÿ–ผ๏ธ Media
F
fofrAI
@fofrAI
๐Ÿ“…
Dec 18, 2025
101d ago
๐Ÿ†”80240614

It arrived ๐Ÿค— https://t.co/glGN3Pwaov

Media 1
๐Ÿ–ผ๏ธ Media
๐Ÿ”_akhaliq retweeted
F
fofr
@fofrAI
๐Ÿ“…
Dec 18, 2025
101d ago
๐Ÿ†”80240614

It arrived ๐Ÿค— https://t.co/glGN3Pwaov

Media 1
โค๏ธ66
likes
๐Ÿ–ผ๏ธ Media
V
vanstriendaniel
@vanstriendaniel
๐Ÿ“…
Dec 18, 2025
101d ago
๐Ÿ†”28282035

๐Ÿธ x ๐Ÿค— Let's go! $10 free @huggingface Inference Provider credits to try open models with Toad! https://t.co/ISkm0YfbcH

@willmcgugan โ€ข Thu Dec 18 15:18

Alrighty. The Toad is out of the bag. ๐Ÿ‘œ๐Ÿธ Install toad to work with a variety of #AI coding agents with one beautiful terminal interface. Check out the blog post for more information... https://t.co/KpQu5cYZzR I've been told I'm very authentic on camera. You just can't fake th

Media 1
๐Ÿ–ผ๏ธ Media
C
ClementDelangue
@ClementDelangue
๐Ÿ“…
Dec 17, 2025
102d ago
๐Ÿ†”94221105

Let's create a new generation of AI builders, not just AI users! https://t.co/tedfrUHUUt

Media 1
๐Ÿ–ผ๏ธ Media
๐Ÿ”huggingface retweeted
C
clem ๐Ÿค—
@ClementDelangue
๐Ÿ“…
Dec 17, 2025
102d ago
๐Ÿ†”94221105

Let's create a new generation of AI builders, not just AI users! https://t.co/tedfrUHUUt

Media 1
โค๏ธ121
likes
๐Ÿ”9
retweets
๐Ÿ–ผ๏ธ Media
C
code
@code
๐Ÿ“…
Dec 18, 2025
101d ago
๐Ÿ†”75911650

Is software engineering still in demand? With Sajjaad Khader ๐Ÿš€ https://t.co/aaW1pLJwI7

๐Ÿ–ผ๏ธ Media
E
emollick
@emollick
๐Ÿ“…
Dec 18, 2025
101d ago
๐Ÿ†”29977943

I think everyone, even the most cynical & informed among us, is going to fall for at least one AI-faked story, photo, or post this coming year & likely many more. (You will also likely believe a real thing was AI) This has bad implications (but wouldnโ€™t blame those taken in) https://t.co/Y5YDLWuaTN

Media 1
๐Ÿ–ผ๏ธ Media
E
emollick
@emollick
๐Ÿ“…
Dec 18, 2025
101d ago
๐Ÿ†”17143895

It would be ironic if my post was in response to fake AI content (they are likely deleted but no way to know) https://t.co/DWBBH2rVNS

Media 1
๐Ÿ–ผ๏ธ Media
H
HistoryGPT
@HistoryGPT
๐Ÿ“…
Dec 17, 2025
102d ago
๐Ÿ†”77375297

Gemini 3 flash is as good at reading handwriting as the average human (pro is expert human level). It is much better than both GPT-5.2 and Opus 4.5 with character level error rates of 1.43% and word level error rates of 2.74%. This is a 47-63% improvement over 2.5 Flash, the same leap we saw with pro. At a fraction of a cent per page, this is a big deal. Read more about the Gemini models on handwriting: https://t.co/6HYnyWu2A9

Media 1
๐Ÿ–ผ๏ธ Media
S
SakanaAILabs
@SakanaAILabs
๐Ÿ“…
Dec 18, 2025
101d ago
๐Ÿ†”16366557

ใ€้˜ฒ่ก›ใƒปใ‚คใƒณใƒ†ใƒชใ‚ธใ‚งใƒณใ‚น้ ˜ๅŸŸใฎใ‚จใƒณใ‚ธใƒ‹ใ‚ขๅ‹Ÿ้›†ใ€‘ ๅ…ˆๆ—ฅใฎใ‚ทใƒชใƒผใ‚บB่ชฟ้”ใ‚’็ตŒใฆใ€Sakana AIใงใฏ็พๅœจใ€ใ€Œ้˜ฒ่ก›ใ€ใŠใ‚ˆใณใ€Œใ‚คใƒณใƒ†ใƒชใ‚ธใ‚งใƒณใ‚นใ€ๅˆ†้‡ŽใซใŠใ‘ใ‚‹ใ‚จใƒณใ‚ธใƒ‹ใ‚ขๆŽก็”จใ‚’ๅŠ ้€Ÿใ•ใ›ใฆใ„ใพใ™ใ€‚ โ–ผใ€Œ้˜ฒ่ก›ใ€ใŠใ‚ˆใณใ€Œใ‚คใƒณใƒ†ใƒชใ‚ธใ‚งใƒณใ‚นใ€ๅˆ†้‡Žใงใฎๅ–ใ‚Š็ต„ใฟใฎไพ‹๏ผˆๆ—ฅ็ตŒใ‚ฏใƒญใ‚นใƒ†ใƒƒใ‚ฏ๏ผ‰ https://t.co/YpI6iakFgn ใ“ใฎๆœช่ธใฎ้ ˜ๅŸŸใ‚’ๅ…ฑใซๅˆ‡ใ‚Šๆ‹“ใไปฒ้–“ใ‚’ๅฟƒใ‹ใ‚‰ๆฑ‚ใ‚ใฆใ„ใพใ™ใ€‚ ใƒ“ใ‚ธใƒงใƒณใ‚„ไบ‹ๆฅญๅ†…ๅฎนใซๅฐ‘ใ—ใงใ‚‚่ˆˆๅ‘ณใ‚’ๆŒใฃใฆใ„ใŸใ ใ„ใŸๆ–นใฏใ€ใœใฒๆŽก็”จใƒšใƒผใ‚ธใ‚’ใ”่ฆงใใ ใ•ใ„ใ€‚ โ€ข Applied Research Engineer https://t.co/KCa6GgvA3z โ€ข Software Engineer https://t.co/Kl2IXgRo7h โ–ผAppliedใƒใƒผใƒ ใฎ็ดนไป‹ https://t.co/2rdYAUlVer

Media 1Media 2
+3 more
๐Ÿ–ผ๏ธ Media
A
antonosika
@antonosika
๐Ÿ“…
Dec 18, 2025
101d ago
๐Ÿ†”25830182

Lovable just raised $330M at a $6.6B valuation. It's been an iconic, insane, wonderful journey so far. Thank you to everyone who made this possible: (thread) https://t.co/EJLOSzpAw3

Media 1
๐Ÿ–ผ๏ธ Media
L
Lovable
@Lovable
๐Ÿ“…
Dec 18, 2025
101d ago
๐Ÿ†”03746647

Announcing Lovable raised $330M at a $6.6B valuation. We launched Lovable to empower the 99%, the people with ideas who donโ€™t code. Now everyone is a builder: founders, teachers, artists, and teams inside the world's largest companies. Meet some of those incredible builders: https://t.co/WJtil2YIE2

๐Ÿ–ผ๏ธ Media
J
jxnlco
@jxnlco
๐Ÿ“…
Dec 18, 2025
101d ago
๐Ÿ†”55524663

Look at this guy, The guy is afraid of asking for what he wants https://t.co/lZ5RdQhCQP

@tenobrus โ€ข Wed Dec 17 01:16

@viemccoy custom instructions ? i would never lobotomize a model like that

Media 1
๐Ÿ–ผ๏ธ Media
_
_akhaliq
@_akhaliq
๐Ÿ“…
Dec 18, 2025
101d ago
๐Ÿ†”15429975

Qwen-Image-Layered Towards Inherent Editability via Layer Decomposition https://t.co/s5sWJsMjpX

Media 1
๐Ÿ–ผ๏ธ Media
_
_akhaliq
@_akhaliq
๐Ÿ“…
Dec 18, 2025
101d ago
๐Ÿ†”28668640

discuss: https://t.co/M8oBAuiwgR

Media 1
๐Ÿ–ผ๏ธ Media
H
HyperNation8
@HyperNation8
๐Ÿ“…
Sep 24, 2022
1282d ago
๐Ÿ†”63096578

Here, a nation where we ๐Ÿ™๏ธoverthrow capitalist society ๐Ÿ™๏ธlive with sincerity, trust & consensus More from Mr H๐Ÿ‘‰https://t.co/3T3qqseYre #hypernation #defi #dao #blockchain #defiproject #daoverse #blockchain #crypto #btc

Media 1
๐Ÿ–ผ๏ธ Media
W
Winterice10
@Winterice10
๐Ÿ“…
Dec 15, 2025
103d ago
๐Ÿ†”70767771

TurboDiffusion: 100โ€“205ร— faster video generation on a single RTX 5090 ๐Ÿš€ Only takes 1.8s to generate a high-quality 5-second video. The key to both high speed and high quality? ๐Ÿ˜SageAttention + Sparse-Linear Attention (SLA) + rCM Github: https://t.co/vT3nfax8H9 Technical Report: https://t.co/LEgLyhdPXh

Media 2
๐Ÿ–ผ๏ธ Media
D
dair_ai
@dair_ai
๐Ÿ“…
Dec 18, 2025
101d ago
๐Ÿ†”40029344

Scaling Laws and Symmetry The common belief is that scaling outperforms inductive biases. Give the model enough data and compute, and it will learn the structure on its own. But this new research finds the opposite. Researchers conducted comprehensive scaling experiments on neural network interatomic potentials, comparing architectures that encode rotational and permutation symmetry to varying degrees. The main finding is that equivariant architectures don't just have lower loss at any given scale. They have better scaling exponents. The performance gap grows as you add more compute. For geometric tasks with known symmetries, the scaling laws favor architectures that encode those symmetries. "Contrary to common belief, we should not leave it to the model to discover fundamental inductive biases such as symmetry, especially as we scale, because they change the inherent difficulty of the task and its scaling laws." Paper: https://t.co/HR5cNwFyaL Learn to build effective AI Agents in our academy: https://t.co/zQXQt0PMbG

Media 1
๐Ÿ–ผ๏ธ Media
O
omarsar0
@omarsar0
๐Ÿ“…
Dec 18, 2025
101d ago
๐Ÿ†”85164341

This is a fascinating paper. It's well known that long-horizon agents have a memory problem. The standard approach is to append everything to the context. Every past observation, action, and thought gets added to the prompt. This creates three compounding issues: O(N) memory growth, degraded reasoning on out-of-distribution context lengths, and attention dilution that causes the model to forget key details even when they're technically in the prompt. This new research unifies memory and reasoning into a single process. It introduces MEM1, an RL framework that trains agents to maintain constant memory across arbitrarily long multi-turn tasks. At each turn, the model updates a compact internal state that simultaneously consolidates prior information and reasons about next actions. After each turn, all previous observations, actions, and states are discarded. Only the most recent internal state remains. Inference-time reasoning serves two purposes. While reasoning about the current query, the model also extracts and stores exactly what it needs for future turns. Memory consolidation becomes part of the reasoning process itself, not a separate module. Training uses PPO with a masked trajectory technique. Because MEM1 dynamically prunes context, standard policy optimization breaks since tokens don't belong to a single continuous trajectory. The authors solve this by stitching sub-trajectories together and applying 2D attention masks that restrict each token's attention to only what was visible when it was generated. The results show dramatic efficiency gains: On 16-objective multi-hop QA, MEM1-7B improves performance by 3.5x while reducing memory usage by 3.7x compared to Qwen2.5-14B-Instruct. Peak token usage stays nearly constant as task complexity increases, while baseline methods scale linearly. On WebShop navigation, MEM1-7B outperforms AgentLM-13B (twice the parameters) with 2.8x less peak token usage. Notably, the agent trained on 2-objective tasks generalizes to 16-objective tasks. Performance actually improves relative to baselines as horizon length increases, because baseline models degrade on out-of-distribution context lengths while MEM1 maintains constant context. Emergent behaviors appear in the trained agents: maintaining structured memory for multiple concurrent questions, shifting focus when one objective stalls, and interleaving reasoning with selective memory updates. External memory modules require separate training and engineering overhead. Full-context approaches don't scale. MEM1 shows that end-to-end RL can train models to consolidate memory as part of reasoning, achieving both efficiency and performance without architectural changes. Paper: https://t.co/q9pEIxBpit Learn to build effective AI Agents in my academy: https://t.co/JBU5beIoD0

Media 1
๐Ÿ–ผ๏ธ Media
E
eliebakouch
@eliebakouch
๐Ÿ“…
Dec 17, 2025
102d ago
๐Ÿ†”72687005

distillation might be one of the most impactful technology of the llm era, really impressive scores https://t.co/8NHdHOtf1b

@ โ€ข

Media 1
๐Ÿ–ผ๏ธ Media
A
AWSstartups
@AWSstartups
๐Ÿ“…
Dec 18, 2025
101d ago
๐Ÿ†”60536085

๐Ÿ”Ž๐Ÿš€@turbopufferโ€™s solution for scalable, reliable & cost-effective search has become a key ingredient for high-performing LLMs. Find out how the startup is working with #AWS to support leading #AI enterprises. ๐Ÿ‘‰ https://t.co/UeVXbkeqbJ https://t.co/BgEX85SuHU

๐Ÿ–ผ๏ธ Media
S
SpirosMargaris
@SpirosMargaris
๐Ÿ“…
Dec 18, 2025
101d ago
๐Ÿ†”56710026

Coursera and Udemy enter a merger agreement valued at around $2.5B https://t.co/LcThQ9cMPA @LaurenForristal @techcrunch

Media 1
๐Ÿ–ผ๏ธ Media
S
SpirosMargaris
@SpirosMargaris
๐Ÿ“…
Dec 18, 2025
101d ago
๐Ÿ†”97502121

Is XRP crashing? The sustained break below $2 signals trouble https://t.co/n94cPbGD5a @godbole17 @coindesk

Media 1
๐Ÿ–ผ๏ธ Media
S
SpirosMargaris
@SpirosMargaris
๐Ÿ“…
Dec 18, 2025
101d ago
๐Ÿ†”14837219

A โ€œscientific sandboxโ€ lets researchers explore the evolution of vision systems https://t.co/eHJlLqS23B @mit

Media 1
๐Ÿ–ผ๏ธ Media
S
SpirosMargaris
@SpirosMargaris
๐Ÿ“…
Dec 18, 2025
101d ago
๐Ÿ†”37156716

Good read Two views of AI and Big Tech https://t.co/G9PN9iqfUp @ft

Media 1
๐Ÿ–ผ๏ธ Media