Your curated collection of saved posts and media

Showing 32 posts Β· last 14 days Β· by score
G
Gradio
@Gradio
πŸ“…
Aug 26, 2025
245d ago
πŸ†”63016486

πŸŽ™οΈ VibeVoice Podcasting πŸ”₯ πŸ™Œ Thanks to @broadfield_dev You Can Now Generate Long-form Multi-speaker AI Podcast with ZeroGPU on @huggingface https://t.co/UKr5JlGPcg

Media 1
πŸ–ΌοΈ Media
_
_akhaliq
@_akhaliq
πŸ“…
Aug 27, 2025
245d ago
πŸ†”29804447

app: https://t.co/esPDyHE1YC check image to image to use in your apps https://t.co/ieYOnUI8Nv

Media 1Media 2
πŸ–ΌοΈ Media
_
_akhaliq
@_akhaliq
πŸ“…
Aug 27, 2025
245d ago
πŸ†”20389432

MV-RAG Retrieval Augmented Multiview Diffusion https://t.co/7Z3RIPGy8M

Media 1
πŸ–ΌοΈ Media
_
_akhaliq
@_akhaliq
πŸ“…
Aug 27, 2025
245d ago
πŸ†”11837716

discuss with author: https://t.co/4hZyFkb4kn

Media 1
πŸ–ΌοΈ Media
_
_akhaliq
@_akhaliq
πŸ“…
Aug 27, 2025
245d ago
πŸ†”34497030

Hermes 4 Technical Report https://t.co/01n2jfk3D5

Media 1
πŸ–ΌοΈ Media
_
_akhaliq
@_akhaliq
πŸ“…
Aug 27, 2025
245d ago
πŸ†”05459212

discuss with author: https://t.co/yXCD4uwFcT

Media 1
πŸ–ΌοΈ Media
_
_akhaliq
@_akhaliq
πŸ“…
Aug 27, 2025
245d ago
πŸ†”39731295

nano banana is now available in anycoder for vibe coding use cases https://t.co/U9AQF3kfcC

Media 1
πŸ–ΌοΈ Media
_
_akhaliq
@_akhaliq
πŸ“…
Aug 27, 2025
245d ago
πŸ†”34771307

nano banana text to image generations in your vibe coded apps is now supported as well https://t.co/o8rkvWAfmz

@_akhaliq β€’ Wed Aug 27 00:12

nano banana is now available in anycoder for vibe coding use cases https://t.co/U9AQF3kfcC

Media 1
πŸ–ΌοΈ Media
_
_akhaliq
@_akhaliq
πŸ“…
Aug 27, 2025
245d ago
πŸ†”62569299

app: https://t.co/esPDyHDu94

Media 1
πŸ–ΌοΈ Media
A
Ali_TongyiLab
@Ali_TongyiLab
πŸ“…
Aug 27, 2025
244d ago
πŸ†”73465095

Hugging Face Paper:https://t.co/d0ZhJe2EE8

Media 1
πŸ–ΌοΈ Media
πŸ”_akhaliq retweeted
A
Tongyi Lab
@Ali_TongyiLab
πŸ“…
Aug 27, 2025
244d ago
πŸ†”73465095

Hugging Face Paper:https://t.co/d0ZhJe2EE8

Media 1
❀️15
likes
πŸ”2
retweets
πŸ–ΌοΈ Media
R
R_Dimm
@R_Dimm
πŸ“…
Aug 25, 2025
246d ago
πŸ†”87638430

I took the Solveit course by @jeremyphoward and @johnowhitaker. Main insight: we can't expect one-shot AI solutions because we can't even ask the right question on the first try. It's no wonder most AI tools feel like a self-driving car hell-bent on driving off a cliff. 🧡 https://t.co/0Pdv6U8JdV

Media 1
πŸ–ΌοΈ Media
R
R_Dimm
@R_Dimm
πŸ“…
Aug 25, 2025
246d ago
πŸ†”60399560

The fix? Work WITH LLM properties, not against them: β€’ RLHF makes them over-eager β†’ Work in small steps, ask clarifying questions β€’ Autoregression causes drift β†’ Edit AI responses, use examples to guide direction β€’ Flawed training data β†’ curate relevant context manually https://t.co/eOGN1PLvnZ

Media 1
πŸ–ΌοΈ Media
J
jeremyphoward
@jeremyphoward
πŸ“…
Aug 25, 2025
246d ago
πŸ†”57074266

@steve2Seattle @SmileyGnome @iwasnevrhere_ @CommunityNotes Seems they actually worked pretty well though? https://t.co/kacXyzLfJv

Media 1
πŸ–ΌοΈ Media
J
jxmnop
@jxmnop
πŸ“…
Aug 26, 2025
245d ago
πŸ†”15528627

first i thought scaling laws originated in OpenAI (2020) then i thought they came from Baidu (2017) now i am enlightened: Scaling Laws were first explored at Bell Labs (1993) https://t.co/CAZPgrxGCX

Media 1Media 2
πŸ–ΌοΈ Media
C
crystalsssup
@crystalsssup
πŸ“…
Aug 27, 2025
244d ago
πŸ†”93762018

Kimi's founder, Zhilin Yang's interview is out. Again, you can let Kimi translate for you: ) lots of insights there. https://t.co/nCEb1Cyq5b Several takes: 1/ Base Model Focus: K2 aims to be a solid base model. We've found that high-quality data growth is slow, and multi-modal data doesn't significantly boost textual "IQ." So, we focus on maximizing every data token's value β€” token efficiency. 2/ Data Rephrasing: With 30T tokens, only a small portion is high-quality data (billions of tokens). We rephrase these to make them more efficient for the model, improving generalization. 3/ Agentic Ability: We aim to enhance generalization. The biggest challenge is making the model generalize well beyond specific tasks. RL improves this over supervised fine-tuning (SFT). 4/ AI-Native Training: We're exploring more AI-native ways to train models. If AI can do good alignment research, it'll generalize better, beyond single-task optimization. 5/ RL vs SFT: RL's generalization is better, as it learns from on-policy samples, but it has its limits. RL helps improve specific tasks, but it's hard to generalize to all scenarios without tailored tasks. 6/ Long Contexts: Context length is crucial, we need millions. The challenge is balancing model size and context length for optimal performance, as some architectures improve with long context but worsen with short ones.

Media 1Media 2
πŸ–ΌοΈ Media
J
JessicaSacher
@JessicaSacher
πŸ“…
Aug 27, 2025
244d ago
πŸ†”80686017

maybe antibiotic resistance would have funding if we didn't prohibit investors from coming to our conferences https://t.co/x0RRAs8nOa

Media 1
πŸ–ΌοΈ Media
A
abidlabs
@abidlabs
πŸ“…
Aug 24, 2025
247d ago
πŸ†”82474556

Follow https://t.co/fl3Mvguo1y to stay up to date. https://t.co/wcYqjR1mqB

@elonmusk β€’ Sat Aug 23 22:16

The @xAI Grok 2.5 model, which was our best model last year, is now open source. Grok 3 will be made open source in about 6 months. https://t.co/TXM0wyJKOh

Media 1Media 2
πŸ–ΌοΈ Media
πŸ”huggingface retweeted
A
πŸ‰ Abubakar Abid
@abidlabs
πŸ“…
Aug 24, 2025
247d ago
πŸ†”82474556

Follow https://t.co/fl3Mvguo1y to stay up to date. https://t.co/wcYqjR1mqB

Media 1Media 2
❀️32
likes
πŸ”5
retweets
πŸ–ΌοΈ Media
L
liran_tal
@liran_tal
πŸ“…
Aug 24, 2025
247d ago
πŸ†”56124378

hugging face is the new github https://t.co/yBo3I7ztEK

Media 1
πŸ–ΌοΈ Media
πŸ”huggingface retweeted
L
Liran Tal | πŸ€– Hacking MCP Servers
@liran_tal
πŸ“…
Aug 24, 2025
247d ago
πŸ†”56124378

hugging face is the new github https://t.co/yBo3I7ztEK

Media 1
❀️110
likes
πŸ”11
retweets
πŸ–ΌοΈ Media
Q
QuanquanGu
@QuanquanGu
πŸ“…
Aug 23, 2025
248d ago
πŸ†”43080770

So many multipliers! Great to see that Grok2 was trained using ΞΌP. https://t.co/mURbaZFkCw https://t.co/li7P9OJCr4

Media 1Media 2
πŸ–ΌοΈ Media
πŸ”huggingface retweeted
Q
Quanquan Gu
@QuanquanGu
πŸ“…
Aug 23, 2025
248d ago
πŸ†”43080770

So many multipliers! Great to see that Grok2 was trained using ΞΌP. https://t.co/mURbaZFkCw https://t.co/li7P9OJCr4

Media 1Media 2
❀️183
likes
πŸ”23
retweets
πŸ–ΌοΈ Media
E
eliebakouch
@eliebakouch
πŸ“…
Aug 24, 2025
247d ago
πŸ†”22536611

Wow, pretty cool that they also open sourced a FSDP2 compatible Muon and PolyNorm working with @huggingface kernels! https://t.co/Gqw7Hpj1v3

@eliebakouch β€’ Sun Aug 24 12:47

Motif 2.6B tech report is pretty insane, first time i see a model with differential attention and polynorm trained at scale! > It's trained on 2.5T of token, with a "data mixture schedule" to continuously adjust the mixture over training. > They use WSD with a "Simple moving ave

Media 1
πŸ–ΌοΈ Media
H
heyshrutimishra
@heyshrutimishra
πŸ“…
Aug 24, 2025
247d ago
πŸ†”45489322

Hugging Face quietly dropped FREE courses with certification It cover everything from LLMs to diffusion models. Here are the best ones you should bookmark today πŸ§΅πŸ‘‡ https://t.co/QvLywX0lZ5

Media 1
πŸ–ΌοΈ Media
S
scaling01
@scaling01
πŸ“…
Aug 23, 2025
248d ago
πŸ†”72407338

Grok-2 got open-sourced same arch as grok-1 https://t.co/eOdmj6zKaK https://t.co/KHb59ymyQ2

Media 1Media 2
+1 more
πŸ–ΌοΈ Media
πŸ”huggingface retweeted
S
Lisan al Gaib
@scaling01
πŸ“…
Aug 23, 2025
248d ago
πŸ†”72407338

Grok-2 got open-sourced same arch as grok-1 https://t.co/eOdmj6zKaK https://t.co/KHb59ymyQ2

Media 1Media 2
+1 more
❀️417
likes
πŸ”32
retweets
πŸ–ΌοΈ Media
H
HaihaoShen
@HaihaoShen
πŸ“…
Aug 25, 2025
246d ago
πŸ†”96211547

πŸ€”A more aggressive INT4 model for DeepSeek-V3.1: https://t.co/mELIFdbpNP #intel #autoround #huggingface @deepseek_ai

Media 1
πŸ–ΌοΈ Media
πŸ”huggingface retweeted
H
Haihao Shen
@HaihaoShen
πŸ“…
Aug 25, 2025
246d ago
πŸ†”96211547

πŸ€”A more aggressive INT4 model for DeepSeek-V3.1: https://t.co/mELIFdbpNP #intel #autoround #huggingface @deepseek_ai

Media 1
❀️279
likes
πŸ”34
retweets
πŸ–ΌοΈ Media
R
reach_vb
@reach_vb
πŸ“…
Aug 25, 2025
246d ago
πŸ†”78417826

Microsoft just released VibeVoice - 1.5B SoTA Text to Speech model - MIT Licensed πŸ”₯ > It can generate up 90 minutes of audio > Supports simultaneous generation of > 4 speakers > Streaming and larger 7B model in-coming > Capable of cross-lingual and singing synthesis Love the expressiveness and the emotion control on the model! Kudos to Microsoft πŸ€—

πŸ–ΌοΈ Media
M
multimodalart
@multimodalart
πŸ“…
Aug 26, 2025
245d ago
πŸ†”21151325

Nano Banana is now available on @huggingface Spaces for free for PRO users! πŸ€— 🀝🍌 https://t.co/rwJ7NOJ8Zd

Media 1
πŸ–ΌοΈ Media
πŸ”huggingface retweeted
M
apolinario 🌐
@multimodalart
πŸ“…
Aug 26, 2025
245d ago
πŸ†”21151325

Nano Banana is now available on @huggingface Spaces for free for PRO users! πŸ€— 🀝🍌 https://t.co/rwJ7NOJ8Zd

Media 1
❀️112
likes
πŸ”24
retweets
πŸ–ΌοΈ Media