Your curated collection of saved posts and media

Showing 32 posts Β· last 14 days Β· by score
O
omarsar0
@omarsar0
πŸ“…
Aug 14, 2025
259d ago
πŸ†”49521780

GPT-5 (with high reasoning effort) achieves near-perfect accuracy on a high-quality ophthalmology question-answering dataset. Based on these other reports, GPT-5 seems to be a very strong model at medical reasoning. https://t.co/asKCWMBwVh

Media 1
πŸ–ΌοΈ Media
O
omarsar0
@omarsar0
πŸ“…
Aug 14, 2025
259d ago
πŸ†”51501991

Speed Always Wins Very nice and comprehensive new report on recent efficient architectures for LLMs. https://t.co/X1VRpLj2kN

Media 1
πŸ–ΌοΈ Media
O
omarsar0
@omarsar0
πŸ“…
Aug 14, 2025
259d ago
πŸ†”60707506

Paper: https://t.co/BTjJkW9zHA https://t.co/sv4wAULSwN

Media 1Media 2
πŸ–ΌοΈ Media
L
llama_index
@llama_index
πŸ“…
Aug 14, 2025
259d ago
πŸ†”33642418

The way we develop AI agent applications is changing, see an example of how we vibe-code a UI for extraction agents πŸ‘‡ Turn LlamaExtract agents into @streamlit web apps using AI-assisted "vibe coding" with Cursor. In this example, we transform an invoice extraction agent into an easy to use web-app. πŸ“ Use our sample cursor prompt that instructs the AI assistant on what to build, our requirements, and sample data. πŸš€ Completely transform from basic scripts to full Streamlit applications πŸ“Š Integrate with LlamaCloud to create and deploy extraction agents based on your pre-defined schemas πŸ’‘ Sample data and schemas included so you can immediately start building your own document processing applications This repository demonstrates our approach to rapid AI-assisted development - start with a working script, use natural language prompts with coding assistants like Cursor/Claude, and iteratively build production-ready applications. Check out the complete vibe coding example and get started: https://t.co/XaKuTUdbcN

Media 2
πŸ–ΌοΈ Media
S
s_batzoglou
@s_batzoglou
πŸ“…
Aug 14, 2025
260d ago
πŸ†”02792112

An exciting new paper proteomics paper describes novel insights on protein-disease mechanisms by using @seer-based MS combined with genomic data on the Genes & Health cohort. From the abstract by Maik Peitzner, Claudia Langenberg and colleagues: "... we idenIfy >1,200 significant variant-protein associations (n=895 cis-protein quantitative trait loci (pQTL)), half of which are novel. [...] consolidate a potential role of 21 proteins in the pathology of 44 diseases: e.g., a novel role of high IGLV3-21 in the development of Grave’s 36 disease elucidating B-cell mediated autoimmunity."

Media 1
πŸ–ΌοΈ Media
J
jxnlco
@jxnlco
πŸ“…
Aug 14, 2025
259d ago
πŸ†”52400569

Office hours from the latest session. https://t.co/91XrRjxR2w

Media 1
πŸ–ΌοΈ Media
πŸ”jxnlco retweeted
J
jason liu
@jxnlco
πŸ“…
Aug 14, 2025
259d ago
πŸ†”52400569

Office hours from the latest session. https://t.co/91XrRjxR2w

Media 1
❀️4
likes
πŸ”1
retweets
πŸ–ΌοΈ Media
E
emollick
@emollick
πŸ“…
Aug 14, 2025
259d ago
πŸ†”97119463

Some signs that catching up in the AI model space is rapidly becoming challenging for even the most highly capitalized companies. https://t.co/dj9ivbXVpF

@markgurman β€’ Wed Aug 13 18:18

BREAKING: Apple prepares ambitious AI devices comeback with multiple robots, smart speaker with a screen, lifelike version of Siri with conversational abilities, redesigned Siri, new Home OS, major home security push & more. Details on the plans here β€” https://t.co/KsQIrKl4w

Media 1
πŸ–ΌοΈ Media
I
iScienceLuvr
@iScienceLuvr
πŸ“…
Aug 14, 2025
259d ago
πŸ†”46749723

Meta releases DINOv3 Everyone talks about Llama but I think Meta's contributions to computer vision (SAM, DINOv2, etc.) are highly underappreciated. They're now releasing a newer iteration with large model (7B param), better data curation, and improved dense features. This is sure to be the foundation for many computer vision use-cases going forward.

@AIatMeta β€’ Thu Aug 14 16:19

Introducing DINOv3: a state-of-the-art computer vision model trained with self-supervised learning (SSL) that produces powerful, high-resolution image features. For the first time, a single frozen vision backbone outperforms specialized solutions on multiple long-standing dense p

Media 1
πŸ–ΌοΈ Media
_
_akhaliq
@_akhaliq
πŸ“…
Aug 14, 2025
259d ago
πŸ†”28125613

Decentralized Aerial Manipulation of a Cable-Suspended Load using Multi-Agent Reinforcement Learning https://t.co/izRATDkpnU

πŸ–ΌοΈ Media
_
_akhaliq
@_akhaliq
πŸ“…
Aug 14, 2025
259d ago
πŸ†”94583682

discuss with author: https://t.co/CnkN6kDzqL

Media 1
πŸ–ΌοΈ Media
_
_akhaliq
@_akhaliq
πŸ“…
Aug 14, 2025
259d ago
πŸ†”95899070

Story2Board A Training-Free Approach for Expressive Storyboard Generation https://t.co/odZyRPE4SO

Media 1
πŸ–ΌοΈ Media
_
_akhaliq
@_akhaliq
πŸ“…
Aug 14, 2025
259d ago
πŸ†”45538317

discuss with author: https://t.co/KQdtJUTJSc

Media 1
πŸ–ΌοΈ Media
_
_akhaliq
@_akhaliq
πŸ“…
Aug 14, 2025
259d ago
πŸ†”65554992

Mol-R1 Towards Explicit Long-CoT Reasoning in Molecule Discovery https://t.co/vVKhZHYPp1

Media 1
πŸ–ΌοΈ Media
_
_akhaliq
@_akhaliq
πŸ“…
Aug 14, 2025
259d ago
πŸ†”93292782

discuss with author: https://t.co/CYDzCuBA0z

Media 1
πŸ–ΌοΈ Media
_
_akhaliq
@_akhaliq
πŸ“…
Aug 14, 2025
259d ago
πŸ†”67605342

EngiBench A Framework for Data-Driven Engineering Design Research https://t.co/9lTi7UfALc

πŸ–ΌοΈ Media
_
_akhaliq
@_akhaliq
πŸ“…
Aug 14, 2025
259d ago
πŸ†”75436658

discuss with author: https://t.co/JbV3dVxVIR

Media 1
πŸ–ΌοΈ Media
_
_akhaliq
@_akhaliq
πŸ“…
Aug 14, 2025
259d ago
πŸ†”82945339

Seeing, Listening, Remembering, and Reasoning A Multimodal Agent with Long-Term Memory https://t.co/stSrINaYLN

πŸ–ΌοΈ Media
_
_akhaliq
@_akhaliq
πŸ“…
Aug 14, 2025
259d ago
πŸ†”50091419

discuss with author: https://t.co/1Kx5w2flUc

Media 1
πŸ–ΌοΈ Media
_
_akhaliq
@_akhaliq
πŸ“…
Aug 14, 2025
259d ago
πŸ†”98809957

D2F Diffusion LLMs Can Do Faster-Than-AR Inference via Discrete Diffusion Forcing https://t.co/p1luIEA7D4

πŸ–ΌοΈ Media
_
_akhaliq
@_akhaliq
πŸ“…
Aug 14, 2025
259d ago
πŸ†”38372089

discuss with author: https://t.co/0wlZRy6GDN

Media 1
πŸ–ΌοΈ Media
_
_akhaliq
@_akhaliq
πŸ“…
Aug 14, 2025
259d ago
πŸ†”13855533

Anycoder one shotted a working @Google Gemma 3 270M transformers.js chatbot app https://t.co/8OjWOoE35w

Media 1
πŸ–ΌοΈ Media
_
_akhaliq
@_akhaliq
πŸ“…
Aug 14, 2025
259d ago
πŸ†”59254721

app: https://t.co/esPDyHDu94

Media 1
πŸ–ΌοΈ Media
_
_akhaliq
@_akhaliq
πŸ“…
Aug 14, 2025
259d ago
πŸ†”83329969

gemma app: https://t.co/2KFWyaHY8N

Media 1
πŸ–ΌοΈ Media
πŸ”huggingface retweeted
M
MiroMindAI
@miromind_ai
πŸ“…
Aug 08, 2025
266d ago
πŸ†”94262041

Hello World! Meet Miromind Open Deep Research (Miro ODR): https://t.co/FbTzNotk6y Miro ODR is the most powerful open-source Deep Research model out there today β€” fully open, fully reproducible, and truly yours to explore. We’ve opened up everything: core models, datasets, training pipelines, AI infrastructure, and our DR Agent framework. No black boxes. No β€œsecret sauce.” Just code, ready for you to run and build on without the headaches. We’ll be pushing out monthly open-source updates, co-creating the strongest Deep Research models together with the community.

Media 1
❀️58
likes
πŸ”9
retweets
πŸ–ΌοΈ Media
πŸ”huggingface retweeted
X
Tiezhen WANG
@Xianbao_QIAN
πŸ“…
Aug 14, 2025
259d ago
πŸ†”39268852

A fully open-sourced, top tier Deep Research framework. Guess which one it is? https://t.co/3ilSHqFHPr

Media 1
❀️73
likes
πŸ”12
retweets
πŸ–ΌοΈ Media
H
huggingface
@huggingface
πŸ“…
Aug 14, 2025
259d ago
πŸ†”34095262

270 million (not billion) parameters! https://t.co/DQt3KDlFFL⚑️⚑️⚑️

@osanseviero β€’ Thu Aug 14 16:04

Introducing Gemma 3 270M πŸ”₯ 🀏A tiny model! Just 270 million parameters 🧠 Very strong instruction following πŸ€– Fine-tune in just a few minutes, with a large vocabulary to serve as a high-quality foundation https://t.co/E0BB5nlI1k https://t.co/XntprMBqSC

Media 1
πŸ–ΌοΈ Media
I
iScienceLuvr
@iScienceLuvr
πŸ“…
Aug 14, 2025
259d ago
πŸ†”90575212

Sample More to Think Less: Group Filtered Policy Optimization for Concise Reasoning "We introduce GFPO (Group Filtered Policy Optimization), which curbs this length explosion by sampling larger groups per problem during training and filtering responses to train on based on two key metrics: (1) response length and (2) token efficiency: reward per token ratio. By sampling more at training time, we teach models to think less at inference time. On the Phi-4-reasoning model, GFPO cuts GRPO's length inflation by 46-71% across challenging STEM and coding benchmarks (AIME 24/25, GPQA, Omni-MATH, LiveCodeBench) while maintaining accuracy. Optimizing for reward per token further increases reductions in length inflation to 71-85%. We also propose Adaptive Difficulty GFPO, which dynamically allocates more training resources to harder problems based on real-time difficulty estimates, improving the balance between computational efficiency and accuracy especially on difficult questions. GFPO demonstrates that increased training-time compute directly translates to reduced test-time compute--a simple yet effective trade-off for efficient reasoning."

Media 1
πŸ–ΌοΈ Media
I
iScienceLuvr
@iScienceLuvr
πŸ“…
Aug 14, 2025
259d ago
πŸ†”93828724

Noise Hypernetworks: Amortizing Test-Time Compute in Diffusion Models "we replace reward guided test-time noise optimization in diffusion models with a Noise Hypernetwork that modulates initial input noise." "We show that our approach recovers a substantial portion of the quality gains from explicit test-time optimization at a fraction of the computational cost."

Media 1
πŸ–ΌοΈ Media
I
iScienceLuvr
@iScienceLuvr
πŸ“…
Aug 14, 2025
259d ago
πŸ†”98565627

Towards Comprehensive Cellular Characterisation of H&E slides "we introduce HistoPLUS, a state-of-the-art model for cell analysis, trained on a novel curated pan-cancer dataset of 108,722 nuclei covering 13 cell types. In external validation across 4 independent cohorts, HistoPLUS outperforms current state-of-the-art models in detection quality by 5.2% and overall F1 classification score by 23.7%, while using 5x fewer parameters."

Media 1
πŸ–ΌοΈ Media
I
iScienceLuvr
@iScienceLuvr
πŸ“…
Aug 14, 2025
259d ago
πŸ†”41919862

Time Is a Feature: Exploiting Temporal Dynamics in Diffusion Language Models "Our work here reveals a critical phenomenon, temporal oscillation, where correct answers often emerge in the middle process, but are overwritten in later denoising steps. To address this issue, we introduce two complementary methods that exploit temporal consistency: 1) Temporal Self-Consistency Voting, a training-free, test-time decoding strategy that aggregates predictions across denoising steps to select the most consistent output; and 2) a post-training method termed Temporal Consistency Reinforcement, which uses Temporal Semantic Entropy (TSE), a measure of semantic stability across intermediate predictions, as a reward signal to encourage stable generations."

Media 1
πŸ–ΌοΈ Media
I
iScienceLuvr
@iScienceLuvr
πŸ“…
Aug 14, 2025
259d ago
πŸ†”09120476

OverFill: Two-Stage Models for Efficient Language Model Decoding "OverFill begins with a full model for prefill, processing system and user inputs in parallel. It then switches to a dense pruned model, while generating tokens sequentially. Leveraging more compute during prefill, OverFill improves generation quality with minimal latency overhead."

Media 1
πŸ–ΌοΈ Media