Your curated collection of saved posts and media
New whale ๐ https://t.co/2WANNgjN5Q
Bro disappeared like he never existed. https://t.co/Bn88feWU2c
This Thanksgiving, Iโm thankful for the quiet, everyday blessings as much as the big milestones. Nine months into my startup journey, I canโt help but reflect on the growth, the challenges, and the gratitude I feel for everyone whoโs been part of it. I am glad we've already achieved several milestones, like our seed round and release state-of-the-art models. First, I want to thank my mom @4katluvrs and dad @Cornellian1988 for all their love and support. I am truly blessed to have such wonderful parents who share their wisdom and invaluable guidance. And thanks for the surprise gift, the delicious launch-day cake! A special shoutout to my gifted and talented sister @sopranotiara, whose support has been invaluable. Sheโs the one who named both @MedARC and @SophontAI and even stepped in as the โbouncerโ at our first company social at ICML Vancouver. Always grateful for her help whenever she can jump in. This startup would also not exist without my awesome cofounder @humanscotti, who I have worked with for 3 years now, I think we make a pretty good team! I am also grateful for people like @EMostaque, @jeremyphoward, @multiply_matrix, etc. who took a chance on me and supported me as I started my AI journey. I'm filled with gratitude to all my investors @stevejang @km @Shaughnessy119 @PonderingDurian @kevinyzhang @TaiPanich_ @JeffDean @OfficialLoganK @ClementDelangue @l2k @MattHartman @headwatermatt @waikit who believe in our vision and decided to support it. Grateful to the incredible Sophont research team for working hard to build models that will hopefully save lives! You guys are rockstars! A big thank-you to @LukeVerswey and his team for guiding us through everything on the legal front. Also appreciate the amazing MedARC community for their contributions to our mission. There are soooooo many more people, from friends to collaborators who have played a role in my startup journey. Finally I am thankful for the encouragement and support from everyone on this platform! Building a startup is the hardest thing I have done in my life, and I am only able to do it with this support Above all, Iโm grateful to God for His grace and for all the blessings that have carried me through this year๐ Happy Thanksgiving to you and your family ๐ฆ๐๐๐ฅ

Feeling like proper Portland locals now that we have our own mushroom spot to return to each year, and traditions building around enjoying them with family and neighbors ๐ https://t.co/lDcbPeqDDq

@natolambert My childhood was good training for this... https://t.co/i1qYg78zVQ
We're all making our signature scents. I went for the smell of heavy industry being reclaimed by nature. Great fun. @gwern your post on avant garde perfumes is partly to blame :) https://t.co/gyaVIZjmtP
@ZebinWilson @Jianlin_S Chrome's auto translate has me wondering with this heading haha https://t.co/mJ3zw9XBO3
I'm going to try sharing some silly science-y shorts to clear out my side-project backlog! Starting with this dizzying duckweed :) Some shorts like this one come with a longer expainer sharing the backstory - links in next tweet. https://t.co/zFLWlenEy2
Short: https://t.co/KXqdN23yBM Backstory: https://t.co/3seV2uLZdy

@matthen2 Thanks :) And good point! I can tweak things and get some spinning, but modelling them as points in a grid definitely misses the reality of them being leaf pairs with holes in different places, momentum and all that other annoying reality stuff :D https://t.co/o5nCcTk7Eu
I made some splats of weird museum spaces (Piranesi-esque) with @theworldlabs and recorded some explorations - 1) a fly thru the space stacked end to end: https://t.co/KJTEH2X5Bs
Hey, SF! Iโll be at the launch of the 2025 @stateofaireport today. Looking forward to meet everyone, make sure to say hi if youโre around. ๐ https://t.co/j9lUbSFDyF
Keep thinking about how spot on (and too early) @navigator by @gentry + team was. The concept of a robot teammate seems obvious today but definitely wasnโt in 2019. Even the brand still looks fresh. https://t.co/7pZonh7zOa
Very excited about OpenAIโs new AgentKit. Visual agent builders are a game changer for iterating on and shipping agents. https://t.co/kErPjxt1DC

Agentic Context Engineering Great paper on agentic context engineering. The recipe: Treat your system prompts and agent memory as a living playbook. Log trajectories, reflect to extract actionable bullets (strategies, tool schemas, failure modes), then merge as append-only deltas with periodic semantic de-dupe. Use execution signals and unit tests as supervision. Start offline to warm up a seed playbook, then continue online to self-improve. On AppWorld, ACE consistently beats strong baselines in both offline and online adaptation. Example: ReAct+ACE (offline) lifts average score to 59.4% vs 46.0โ46.4% for ICL/GEPA. Online, ReAct+ACE reaches 59.5% vs 51.9% for Dynamic Cheatsheet. Paper: https://t.co/AZRZe0axlI
Great recap of security risks associated with LLM-based agents. The literature keeps growing, but these are key papers worth reading. Analysis of 150+ papers finds that there is a shift from monolithic to planner-executor and multi-agent architectures. Multi-agent security is a widely underexplored space for devs. Issues range from LLM-to-LLM prompt infection, spoofing, trust delegation, and collusion.
One Tuesday night, as my wife and I sat down for dinner, a sheriffโs deputy knocked on the door to serve me a subpoena from OpenAI. I held back on talking about it because I didn't want to distract from SB 53, but Newsom just signed the bill so... here's what happened: ๐งต https://t.co/VnYCJYg2DH
LlamaParse now supports @AnthropicAI's Claude Sonnet 4.5, plus exciting new features for enhanced document processing! We've integrated Sonnet 4.5 into our parsing capabilities, giving you access to Anthropic's latest model for even better document understanding and parsing. Stay tuned for detailed benchmarks comparing the new sonnet model with our existing parsing options. Sign-up to LlamaCloud to start parsing: https://t.co/YDn6qsUnZK
Low-code is nice, but if I had to bet on a future, itโs code-based orchestration + coding agents to let anyone bridge that gap. OpenAIโs AgentKit (left img) lets you get started building various flows, like comparing docs, or a basic assistant. Once you need to encode more domain-specific logic/fetch from a data source/create a longer-running agent, youโll need to export to code and maintain your own workflow. Iโm bullish on building advanced agents over your data that live natively on top of a code-based orchestration framework (right two images). Weโve built core tools in @llama_index to help enable building code-based agentic workflows and then deploying them. You can easily get started through a vibe-coding tool or through our templates, but then you get the full flexibility to add whatever you want on top. You get the underlying benefits of agentic orchestration: state management, checkpointing, human-in-the-loop. Weโve also been super deep in coding tools like Claude Code/Cursor/Codex to make sure youโre able to build these automations super easily. With our latest alpha release of LlamaAgents, you can build whatever workflow you want in code and deploy it as an e2e agent on LlamaCloud! Come check it out ๐ https://t.co/o3nMGoQRjS

๐ Parse vs. Extract: Two Fundamental Approaches to Document Processing Building document agents? Knowing when to parse versus when to extract is fundamental to getting your architecture right. In this deep dive, @tuanacelik breaks down: ๐ Parsing - Converting unstructured documents into structured markdown while preserving layout, tables, and formatting. Perfect for applications where you need full document context. โ๏ธ Extraction - Using LLMs to pull specific fields and entities with schema validation. Ideal for extracting specific info for downstream tasks. Read the full technical guide ๐https://t.co/Zx9Xmnni0K

EmbeddingGemma, a compact 308M parameter multilingual embedding model, perfect for on-device RAG applications - and we've made it super easy to integrate with LlamaIndex! ๐ ๏ธ Ready-to-use integration with LlamaIndex's @huggingface Embedding class - just specify the query and document prompts The model achieves top rankings on the Massive Text Embedding Benchmark while being small enough for mobile devices. Plus, it's easily fine-tunable - the blog shows how fine-tuning on medical data created a model that outperforms much larger alternatives. We love seeing efficient models like this that make powerful embeddings accessible everywhere, especially for edge deployments where every MB counts. See the full technical deep-dive and integration examples: https://t.co/AmMtCEkgKD

Classify documents automatically with LlamaClassify - no more manual sorting through contracts and legal documents. Our new classification service lets you build intelligent document sorting systems that understand content and provide reasoning for their decisions: ๐ Define custom classification rules with simple descriptions instead of complex ML models ๐ค Get both classification results AND detailed reasoning explaining why each document was categorized ๐ฏ Works great for legal documents, contracts, and any structured document types In this example, we show you how to distinguish between affiliate agreements and co-branding contracts. The system doesn't just tell you "this is an affiliate agreement" - it explains exactly why, citing specific document language and structural elements. Learn how to classify contract types: https://t.co/wk7cAwqZt1
๐ฅThrilled to share our new work Reinforce-Ada, which fixes signal collapse in GRPO ๐ฅณNo more blind oversampling or dead updates. Just sharper gradients, faster convergence, and stronger models. โ๏ธ One-line drop-in. Real gains. https://t.co/kJTeVek1S3 https://t.co/7qLywG2KWR https://t.co/4BGowcLpl5

๐ ๐๐ซ๐๐ฌ๐๐ง๐ญ๐ข๐ง๐ ๐ ๐ฉ๐๐ฉ๐๐ซ ๐ข๐ฌ ๐๐ง ๐๐ซ๐ญ!๐ค ๐ค Ever felt that most presentation tools lack flexibility and creativity? ๐๐ฆ๐ณ๐ฆ๐ญ๐บ ๐ฆ๐น๐ต๐ณ๐ข๐ค๐ต๐ช๐ฏ๐จ ๐ค๐ฐ๐ฏ๐ต๐ฆ๐ฏ๐ต, ๐ง๐ฐ๐ณ๐ค๐ช๐ฏ๐จ ๐ณ๐ช๐จ๐ช๐ฅ ๐ฅ๐ฆ๐ด๐ช๐จ๐ฏ๐ด, ๐ข๐ฏ๐ฅ ๐ฅ๐ฆ๐ฎ๐ข๐ฏ๐ฅ๐ช๐ฏ๐จ ๐ฎ๐ข๐ฏ๐ถ๐ข๐ญ ๐ต๐ธ๐ฆ๐ข๐ฌ๐ด. ๐๐ฏ๐จ๐๐ซ๐๐ฌ๐๐ง๐ญ changes all of that! โจ EvoPresent is a self-optimizing framework that unites storytelling, design, and feedback to create effortless, engaging presentation videos. ๐ฅ ๐ก ๐๐๐ฒ ๐๐ข๐ ๐ก๐ฅ๐ข๐ ๐ก๐ญ๐ฌ: ๐ ๐๐ซ๐๐ฌ๐๐๐ฌ๐ญ๐ก, the core multi-task RL model, continuously refines both content and design โ ensuring slides that are impactful and visually captivating. ๐ ๐๐ฏ๐จ๐๐ซ๐๐ฌ๐๐ง๐ญ ๐๐๐ง๐๐ก๐ฆ๐๐ซ๐ค is a comprehensive evaluation suite: 650+ top AI papers & diverse formats to assess content and design, and 2000+ slide pairs for aesthetic scoring, defect correction, and design comparison. ๐ฏ ๐ง ๐๐ช๐จ๐ฉ-๐ฒ๐ถ๐ข๐ญ๐ช๐ต๐บ ๐ง๐ฆ๐ฆ๐ฅ๐ฃ๐ข๐ค๐ฌ ๐ฑ๐ฐ๐ธ๐ฆ๐ณ๐ด ๐ค๐ฐ๐ฏ๐ต๐ช๐ฏ๐ถ๐ฐ๐ถ๐ด ๐ด๐ฆ๐ญ๐ง-๐ช๐ฎ๐ฑ๐ณ๐ฐ๐ท๐ฆ๐ฎ๐ฆ๐ฏ๐ต. โ๏ธ ๐๐ข๐ญ๐ข๐ฏ๐ค๐ฆ ๐ฃ๐ฆ๐ต๐ธ๐ฆ๐ฆ๐ฏ ๐ค๐ฐ๐ฏ๐ต๐ฆ๐ฏ๐ต & ๐ฅ๐ฆ๐ด๐ช๐จ๐ฏ is the secret to presentation excellence. ๐ ๐๐ฎ๐ฅ๐ญ๐ข-๐ญ๐๐ฌ๐ค ๐๐ training boosts generalization in aesthetic awareness tasks. ๐๐ข๐ฌ๐๐ฅ๐๐ข๐ฆ๐๐ซ: the demo video was completely generated by EvoPresent, no human refinement.
Transfusion combines autoregressive with diffusion to train a single transformer, but what if we combine Flow with Flow? ๐ค ๐OneFlow๐ the first non-autoregressive model to generate text and images concurrently using a single transformerโunifying Edit Flow (text) with Flow Matching (images). Performance boost while unlocking new capabilities: ๐ฅ Scales better than Transfusion (AR), 50% fewer FLOPS than Transfusion for similar performance โก Mixed-modal training boosts both generation & understanding ๐ Mask diffusion: fixed-length + extra FLOPS for masked tokens. Transfusion: sequential generation โ OneFlow: variable-length via token deletion + concurrent mixed-modal generation with fewer FLOPS ๐งต1/n
Introducing RND1, the most powerful base diffusion language model (DLM) to date. RND1 (Radical Numerics Diffusion) is an experimental DLM with 30B params (3B active) with a sparse MoE architecture. We are making it open source, releasing weights, training details, and code to catalyze further research on DLM inference and post-training. We are researchers and engineers (DeepMind, Meta, Liquid, Stanford) building the engine for recursive self-improvement (RSI) โ and using it to accelerate our own work. Our goal is to let AI design AI. We are hiring.
Join us Monday, Oct 13, for our Modular Community Meeting! Learn about Modular 25.6, our latest release unifying NVIDIA, AMD, and Apple GPUs, and see community projects like โA generic FFT in Mojoโ and โA MAX backend for PyTorch.โ https://t.co/kshPmT8Oeg
When you're AI coding, a lot of times this is not your full time job. Our new dashboard helps keep track of your progress and consistency, helps you keep shipping, keep building. https://t.co/OHeBuWqPuv
One of the hardest things about AI coding is knowing if the code that was just written does what you need it to do. Code reviews are good and necessary, but they don't tell you if the code does what you *intended*. That's why we built Acceptance Reviews into the BrainGrid MCP - an easy way to know if the agent is done. The new BrainGrid MCP tool validates pull requests against a requirements document to determine if the code meets the acceptance criteria.
๐ docs โ https://t.co/3DIYpV2DHr
If you want this workflow on rails, ๐ โ https://t.co/LwhBxGgzAZ
Love using @upstash Workflows for all async operations in @BrainGridAI. Instead of managing queues + workers, we write workflows like functions: โข Type-safe triggers via WorkflowService โข Auto-retries with exponential backoff โข Built-in step persistence Zero infrastructure management. Just deploy Next.js.