Your curated collection of saved posts and media

Showing 10 posts Β· last 14 days Β· by score
βž• Add New Post
A
arankomatsuzaki
@arankomatsuzaki
πŸ“…
Apr 10, 2026
13d ago
πŸ†”98191322

proj: https://t.co/fCl3QBpnoE abs: https://t.co/i50YCzUKax repo: https://t.co/ml7dOTqYy1

Media 1
πŸ–ΌοΈ Media
A
arankomatsuzaki
@arankomatsuzaki
πŸ“…
Apr 10, 2026
13d ago
πŸ†”86539809

Squeeze Evolve: A Unified Framework for Verifier-Free Evolution Across AIME 2025, GPQA-Diamond, ARC-AGI-V2, MMMU-Pro, etc: - Up to ~3x API cost reduction - Up to ~10x increase in fixed-budget serving throughput https://t.co/MoTgGNAYns

Media 1
πŸ–ΌοΈ Media
G
gerardsans
@gerardsans
πŸ“…
Apr 10, 2026
13d ago
πŸ†”82203756
⭐0.40

@plainionist The industry has misled the public with terms like β€œIn-Context Learning”, β€œSkills” or β€œReasoning”. It’s trivial to prove no real learning is happening: the model’s weights are never changed. Remove the context, and the alleged β€œlearning” disappears instantly. Puff.

P
PrismML
@PrismML
πŸ“…
Apr 16, 2026
7d ago
πŸ†”82896134

Today we’re announcing Ternary Bonsai: Top intelligence at 1.58 bits Using ternary weights {-1, 0, +1}, we built a family of models that are 9x smaller than their 16-bit counterparts while outperforming most models in their respective parameter classes on standard benchmarks. We’re open-sourcing the models under the Apache 2.0 license in three sizes: 8B (1.75 GB), 4B (0.86 GB), and 1.7B (0.37 GB).

Media 1
πŸ–ΌοΈ Media
R
RedHat_AI
@RedHat_AI
πŸ“…
Apr 17, 2026
6d ago
πŸ†”02520952

Qwen3.6-35B-A3B just dropped. Red Hat AI has an NVFP4 quantized checkpoint ready. 35B params, 3B active, quantized with LLM Compressor. Preliminary GSM8K Platinum: 100.69% recovery (slightly above baseline). Early release. Let us know what you think! https://t.co/i5Fc4P7NVN

Media 1
πŸ–ΌοΈ Media
V
victormustar
@victormustar
πŸ“…
Apr 17, 2026
6d ago
πŸ†”46958899

Sharing my current setup to run Qwen3.6 locally in a good agentic setup (Pi + llama.cpp). Should give you a good overview of how good local agents are today: # Start llama.cpp server: llama-server \ -hf unsloth/Qwen3.6-35B-A3B-GGUF:Q4_K_XL \ --jinja \ --chat-template-kwargs '{"preserve_thinking":true}' \ --temp 0.6 --top-p 0.95 --top-k 20 --min-p 0 # Configure Pi: { "providers": { "llama-cpp": { "baseUrl": "http://127.0.0.1:8080/v1", "api": "openai-completions", "apiKey": "none", "models": [ { "id": "unsloth/Qwen3.6-35B-A3B-GGUF:Q4_K_XL" } ] } } }

@Alibaba_Qwen β€’ Thu Apr 16 13:23

⚑ Meet Qwen3.6-35B-A3B:Now Open-SourceοΌπŸš€πŸš€ A sparse MoE model, 35B total params, 3B active. Apache 2.0 license. πŸ”₯ Agentic coding on par with models 10x its active size πŸ“· Strong multimodal perception and reasoning ability 🧠 Multimodal thinking + non-thinking modes Efficient. Pow

Media 1
πŸ–ΌοΈ Media
R
rasbt
@rasbt
πŸ“…
Apr 11, 2026
12d ago
πŸ†”58256302

@DanielWulikk Have to think a bit about how to best visualize it, but if you are interested, I have a working from-scratch code implementation of Gemma 4 E2B in the meantime to see how per-layer embeddings are implemented: https://t.co/jyiq1vyJnH https://t.co/fVrSBWHNHl

Media 1
πŸ–ΌοΈ Media
E
e_volkmann
@e_volkmann
πŸ“…
Apr 08, 2026
15d ago
πŸ†”30881771

Introducing gyaradax πŸ‰: A JAX solver for local flux-tube gyrokinetics with custom CUDA kernels for acceleration. This entire code was vibecoded by @ggalletti_ and me in a month. Validated against GKW (CPU-only Fortran code) with 10x speedups. Details and code in the replies. https://t.co/22PrHjItR5

πŸ–ΌοΈ Media
_
_akhaliq
@_akhaliq
πŸ“…
Apr 10, 2026
13d ago
πŸ†”70916384

DMax Aggressive Parallel Decoding for dLLMs paper: https://t.co/y421NkegRD https://t.co/Y7Ut9Gxly8

πŸ–ΌοΈ Media
I
iScienceLuvr
@iScienceLuvr
πŸ“…
Apr 10, 2026
13d ago
πŸ†”90441301

abs: https://t.co/tSLRNXAgY3 code: https://t.co/ldWVGmuC4O blog post: https://t.co/yI0NmdaKHm

Media 1Media 2
πŸ–ΌοΈ Media
Page 1 of 109Next β†’