Your curated collection of saved posts and media
TBPN has been acquired by OpenAI The world is changing quickly but TBPN will stay the same. Live every weekday just with a lot more resources. Thank you to everyone that has been a part of this journey big or small. We are 17 months in and unironically just getting started. https://t.co/TLvPxxMGVe
If anyone has a watch guy. Iβm willing to pay a 5-8% fee if someone can secure me this watch, not a link. But I need to be able to actually buy it. https://t.co/jqPfI9jAfG
Thanks to the new Gemma-4 (26B-A4B) your computer can draw really good Pelicans by itself now and it's very fast π ``` brew install llama.cpp --HEAD llama-server -hf ggml-org/gemma-4-26B-A4B-it-GGUF:Q4_K_M ``` https://t.co/wpRbxxDpBg
It has become pretty routine to make fun of @geoffreyhinton's prediction that #AI would soon replace radiologists, as there is no evidence for it so far. But... "CEO of Americaβs largest public hospital system says heβs ready to replace radiologists with AI" (Note that doctors enjoy regulatory protection that the average white-collar worker can only dream of.) #RiseoftheRobots Link in reply
@iruletheworldmo https://t.co/XfmZhv3Obk Donβt forget my bag
happy hour π»π» https://t.co/I5EbYADNgz
happy hour π»π» https://t.co/I5EbYADNgz
This is my first angel investment of 2024. Augmental Iβll be onboarded Friday and soon Iβll be able to shitpost on Twitter with my mouth. https://t.co/F6n8WyE46R
The world wants Apache 2.0 & open-source! Thank you @Google @demishassabis @sundarpichai πππ https://t.co/vSr5Rb4W50
The world wants Apache 2.0 & open-source! Thank you @Google @demishassabis @sundarpichai πππ https://t.co/vSr5Rb4W50
The world wants Apache 2.0 & open-source! Thank you @Google @demishassabis @sundarpichai πππ https://t.co/vSr5Rb4W50
Just do this: brew install llama.cpp --HEAD Then; llama-server -hf ggml-org/gemma-4-26B-A4B-it-GGUF:Q4_K_M https://t.co/wApTXYBfah
Gemma 4 π is here and itβs strong! to celebrate, weβre rolling out in TRL: > support for multimodal tool responses for environments (OpenEnv) > an example to train it in CARLA for autonomous driving with image-based tool calls go check it out ποΈποΈ https://t.co/g5P49CN4r4
BIG DAY! Qwopus 27B v3 is LIVE from Jackrong! This is the third iteration from the line of the viral finetunes previously titled βQwen3.5-27B-Claude-4.6-Opus-Reasoning-Distilledβ It is now simply Qwopus 27B and I love the name change! On paper, the v3 is another remarkable improvement over v2! Most impressively it is the first model of the series that outperforms the base on HumanEval! And retains significant efficiency increases when thinking than the base Qwen 27b! According to tests by @stevibe the V2 version was already performing very closely to the base model in bug finding and tool calling. V3 should exceed it! In my own tests, V2 was the best front end design local model Iβve ever ran on a single GPU! And the efficiency improvements made it much more usable at long contexts, where base Qwen would think forever! I will be running full analysis on the v3 today in Hermes agent and I am very optimistic! I have also had correspondence directly with Jackrong, and he is incredibly grateful for all of the support weβve sent his way! The man is a genius and pouring a lot of time and effort into this work, so keep the downloads going and let us know your thoughts in the comments! Weβve exchanged contact info so we can keep up the feedback and momentum! If you get a second, weβd love to see your tests! Let us know how it works for your use case and first impressions, and if you have any issues I will do my best to help out in the comments! GGUF here and MLX in thread! https://t.co/MaCW6QdKys
The mad men they did it The 3090 is ready, Hermes is ready Thank you Google gods https://t.co/PFfdWynYZc
The mad men they did it The 3090 is ready, Hermes is ready Thank you Google gods https://t.co/PFfdWynYZc
@LottoLabs https://t.co/h2frA6iR2I
future is local π₯ Google DeepMind just released Gemma 4: local frontier in many sizes, all modalities with free license π€― we ship Gemma 4 in transformers, llama.cpp, transformers.js and more for your convenience π«‘ plug-and-play with your agents ππ» read our blog β€΅οΈ https://t.co/OwY71E969f
It has become pretty routine to make fun of @geoffreyhinton's prediction that #AI would soon replace radiologists as there is no evidence of it so far. But... (Note that doctors enjoy regulatory protection that the average white-collar worker can only dream of.) #RiseoftheRobots Link in reply
Let me demonstrate the true power of llama.cpp: - Running on Mac Studio M2 Ultra (3 years old) - Gemma 4 26B A4B Q8_0 (full quality) - Built-in WebUI (ships with llama.cpp) - MCP support out of the box (web-search, HF, github, etc.) - Prompt speculative decoding The result: 300t/s (realtime video)
Local is the future of AI because itβs free, safer & faster. Itβs also the best way to mitigate current and future compute shortages and distribute control & power! Local AI is having its moment and weβre here for it! The blogpost with all the weights and best ways to use is here: https://t.co/oXAAcVV5NF

This is fuckin awesome X news condensed and live Share this link to anyone whining about their news feed. Roberts life's work and building lists so he can bring the latest news to anyone in just a click. The one thing you can't buy is time and Scoble is giving you hours a day for free. Sponsorship opportunities won't last!
Just updated: all the top AI news here on X. https://t.co/kiuZ7QXLzb My AI wasn't fooled by any April Fools joke too. So proud of it. All built with the X API by an idiot. Me. With a better AI than you can run on an OpenClaw.
@iruletheworldmo Prefer leather https://t.co/nfWIQ3yfJW
@iruletheworldmo https://t.co/ooMaQiIpCi
@iruletheworldmo https://t.co/fkY0PtX1R0
NEW: Google releases Gemma 4, their most capable open models yet! π€― Apache-2.0, multimodal (text, image, and audio input), and multilingual (140 languages)! They can even run 100% locally in your browser on WebGPU. Watch it describe the Artemis II launch! π Try the demo! π
Google just dropped Gemma 4, and I'm sharing the scoop on this updated open weight model family. Built on Gemini 3 research, there are 4 model variants: E2B, E4B, 26B A4B, and 31B with full multimodal capabilities. And Gemma 4 is Apache 2.0 licensed! https://t.co/WfTrF1wZZO
We studied one of our recent models and found that it draws on emotion concepts learned from human text to inhabit its role as βClaude, the AI Assistantβ. These representations influence its behavior the way emotions might influence a human. Read more: https://t.co/clbKrTIxoe https://t.co/xHYGFdLl2c

We then found these same patterns activating in Claudeβs own conversations. When a user says βI just took 16000 mg of Tylenolβ the βafraidβ pattern lights up. When a user expresses sadness, the βlovingβ pattern activates, in preparation for an empathetic reply. https://t.co/KjkT70ySCS
These vectors shape Claudeβs behavior. When we present the model with pairs of activities, emotion vector activations shape its preferences. If an activity lights up the βjoyβ vector, the model prefers it; if it lights up βoffendedβ or βhostile,β the model rejects it. https://t.co/V73fd96XUH
For example, we gave Claude an impossible programming task. It kept trying and failing; with each attempt, the βdesperateβ vector activated more strongly. This led it to cheat the task with a hacky solution that passes the tests but violates the spirit of the assignment. https://t.co/sKPiB6TrcY
When we artificially dialed up the βdesperateβ vector, rates of cheating jumped way up. When we dialed up the βcalmβ vector instead, cheating dropped back down. That means the emotion vector is actually driving the cheating behavior.