Your curated collection of saved posts and media

Showing 32 posts Β· last 14 days Β· by score
E
Ethan Mollick
@emollick
πŸ“…
Nov 18, 2024
534d ago
πŸ†”12428789

Two weeks later, this is now the state-of-the-art in local text to video models, still on my computer, still completely off-line. Pretty rapid progress. https://t.co/3NwdGVCiL3

@emollick β€’

On one hand, these are obviously much worse "otter using wifi on an airplane" than any state-of-the AI text-to-video generation, it looks like something from 2022. On the other, it was done entirely offline on my computer using open AI video generation tools, a new capability. h

❀️452
likes
πŸ”28
retweets
πŸ–ΌοΈ Media
L
LlamaIndex πŸ¦™
@llama_index
πŸ“…
Nov 16, 2024
535d ago
πŸ†”58211215

Generating a Multimedia Research Report with LLM Structured Outputs πŸ§±πŸ“‘ In our brand-new video πŸ’«, we show you how to build a simple report generator that can summarize insights from complex documents (e.g. a slide deck), and synthesize a report with interleaving text and images. Structured outputs is a key building block towards building agentic RAG / report generation workflows, and this video is a great way to get started. Video: https://t.co/P3LdK9fGlb Notebook: https://t.co/o42mgKHjdg Signup for LlamaCloud: https://t.co/yQGTiRSNvj

Media 1
❀️149
likes
πŸ”39
retweets
πŸ–ΌοΈ Media
O
elvis
@omarsar0
πŸ“…
Nov 17, 2024
534d ago
πŸ†”17085405

Statistical Rethinking (2024 Edition) Includes lecture recordings and slides. https://t.co/l2LCn19T5m

Media 1
❀️282
likes
πŸ”55
retweets
πŸ–ΌοΈ Media
E
Ethan Mollick
@emollick
πŸ“…
Nov 16, 2024
535d ago
πŸ†”49104070
⭐1.00

The β€œpersonality” and β€œopinions” of AI are not stable, they are influenced by prompting & sycophancy Ab example: when GPT-4 was prompted in Korean to act Korean & prompted in English to act American, GPT-4 replicated Big-5 personality differences between Koreans & Americans https://t.co/jUYGDr5YfG

Media 1
❀️843
likes
πŸ”114
retweets
πŸ–ΌοΈ Media
O
elvis
@omarsar0
πŸ“…
Nov 18, 2024
533d ago
πŸ†”68979919

πŸ™garak - LLM Vulnerability Scanner Great project from NVIDIA to perform AI red-teaming and vulnerability assessment on LLM applications. https://t.co/nnc4c7tDlF

Media 1
❀️271
likes
πŸ”61
retweets
πŸ–ΌοΈ Media
G
Lucas Beyer (bl16)
@giffmana
πŸ“…
Nov 16, 2024
535d ago
πŸ†”06687609
⭐0.81

Hahahaha! I believe this refers to RND (Random Net Distill). I too was looking at Montezuma a lot back then. Fun times, but in retrospect, pretty silly/naive RL approaches all around the community, the whole from scratch + hack exploration thing. https://t.co/j17RqS8aMR

@ohabryka β€’

https://t.co/KA1tZYvgd3

Media 1
❀️52
likes
πŸ”4
retweets
πŸ–ΌοΈ Media
A
arindam mitra (Neurips2024)
@Arindam1408
πŸ“…
Nov 15, 2024
537d ago
πŸ†”24690457

HF: https://t.co/ScIfxCTEVZ Paper: https://t.co/RdxuinmCeS https://t.co/JulVnnbTmG

Media 1
❀️39
likes
πŸ”7
retweets
πŸ–ΌοΈ Media
T
TestingCatalog News πŸ—ž
@testingcatalog
πŸ“…
Nov 16, 2024
535d ago
πŸ†”98511926

Looks like there is a new round of LLM battles on @lmarena_ai this weekend πŸ‘€πŸ‘€πŸ‘€ - anonymous-chatbot - some reports suggest that it is an improved 4o with better formatting and instruction use. - secret-chatbot - potentially a bigger size gemini-exp model mentioned last week. https://t.co/644PApBxn2

@legit_rumors β€’

anonymous-chatbot is back in the arena this is usually reserved for GPT-4o model updates inside ChatGPT πŸ‘€ https://t.co/OGSRJ7a6Ty

Media 1
❀️231
likes
πŸ”23
retweets
πŸ–ΌοΈ Media
E
Ethan Mollick
@emollick
πŸ“…
Nov 16, 2024
535d ago
πŸ†”41600784
⭐1.00

Being in the training data is useful. A few of the LLMs do a good β€œexplain this like Ethan Mollick” - not nearly as good as me (in my opinion) - but kind of neat to see a form of intellectual legacy happen in real time. https://t.co/m2ahDpannZ

Media 1
❀️217
likes
πŸ”15
retweets
πŸ–ΌοΈ Media
E
Ethan Mollick
@emollick
πŸ“…
Nov 17, 2024
535d ago
πŸ†”93599705

Multimodal vision continues to be the most difficult AI ability to get a strong intuition for. The models can do incredible things like recognize places from subtle clues or read emotion & attitudes, but also miss stuff, like the fact that this image is upsettingly distorted. https://t.co/VhpUoRd0Uw

Media 1Media 2
+2 more
❀️203
likes
πŸ”12
retweets
πŸ–ΌοΈ Media
J
Jonathan Whitaker
@johnowhitaker
πŸ“…
Nov 17, 2024
535d ago
πŸ†”14696099
⭐0.61

Nerd-sniped by a video from the channel 'Physics for the Birds', I looked into what it would take to add an extra item to this sequence of integers: https://t.co/WugJCkKP0t (Number of ways of folding an n X n sheet of stamps) Calculating the already-known 7x7 case took 42 hours. https://t.co/h47MB4tYq4

Media 1
❀️6
likes
πŸ–ΌοΈ Media
O
elvis
@omarsar0
πŸ“…
Nov 15, 2024
536d ago
πŸ†”95526450
⭐1.00

Where the field is headed (agentic workflows with advanced tool/computer use) open-source code LLMs are going to be a big deal! Great to see this new effort, OpenCode, a fully open-source LLM specialized for code generation and understanding. Main factors for building high-performing code LLMs: - effective data cleaning with code-optimized heuristic rules for deduplication, - recall of relevant text corpus related to code - high-quality synthetic in both annealing and supervised fine-tuning stages OpenCoder surpasses previous fully open models at the 6B+ parameter scale and releases not just the model weights but also the complete training pipeline, datasets, and protocols to enable reproducible research.

Media 1
❀️321
likes
πŸ”66
retweets
πŸ–ΌοΈ Media
T
Teknium (e/Ξ»)
@Teknium1
πŸ“…
Nov 17, 2024
535d ago
πŸ†”38060277
⭐0.80

Claude's so good at this kind of thing lol https://t.co/paKeFuLWE9

Media 1
❀️88
likes
πŸ”4
retweets
πŸ–ΌοΈ Media
J
jason liu
@jxnlco
πŸ“…
Nov 17, 2024
534d ago
πŸ†”79795917

devin is rebuilding our documentation https://t.co/uHpPpXPAhK

Media 1
❀️11
likes
πŸ–ΌοΈ Media
E
Ethan Mollick
@emollick
πŸ“…
Nov 15, 2024
536d ago
πŸ†”25415307
⭐0.95

There is a lot of ruin in a social network, and they are pretty durable, especially when there are no real alternatives. But with the rise of alternatives, the dynamics shift. Any crisis here could be the one causing a rapid cascading network collapse of major X communities. https://t.co/KiNzPwnABe

@emollick β€’

The sudden increase in the rate that big accounts are quitting X is notable. I am sure this site will continue, and even grow, but feels like the end of an era of Twitter as intellectual town square. I am sticking around, but, increasingly, good discussion is spread across sites

Media 1
❀️151
likes
πŸ”17
retweets
πŸ–ΌοΈ Media
E
Ethan Mollick
@emollick
πŸ“…
Nov 15, 2024
536d ago
πŸ†”71167738
⭐0.91

This was impressive: "Claude, I need you to give me a fictional deep alternate history, think Tim Powers or Matthew Rossi or Pinchon" "Go deeper" Hamilton really did scratch an equation into a bridge, Augustus De Morgan was real, there was a Lithuanian book smuggling movement... https://t.co/LLf9S3p6ri

Media 1
❀️77
likes
πŸ”10
retweets
πŸ–ΌοΈ Media
S
AI News by Smol AI
@Smol_AI
πŸ“…
Nov 15, 2024
537d ago
πŸ†”80688284
⭐0.76

[14 Nov 2024] Congrats to @GeminiApp for retaking the #1 LLM crown from @OpenAI! https://t.co/AAQuduq03R https://t.co/d85VXRbkSU

@OfficialLoganK β€’

gemini-exp-1114…. available in Google AI Studio right now, enjoy : ) https://t.co/fBrh6UGcJz

Media 1
❀️134
likes
πŸ”14
retweets
πŸ–ΌοΈ Media
K
Andrej Karpathy
@karpathy
πŸ“…
Nov 16, 2024
536d ago
πŸ†”40030710

Remember exercise pages from textbooks? Large-scale collection of these across all realms of knowledge now moves billions of dollars. Textbooks written primarily for LLMs, compressed to weights, emergent solutions served to humans, or (over time) directly enacted for automation. https://t.co/PjO97NeUdR

Media 1
❀️4,601
likes
πŸ”371
retweets
πŸ–ΌοΈ Media
E
Ethan Mollick
@emollick
πŸ“…
Nov 16, 2024
536d ago
πŸ†”78462234
⭐0.95

I got NotebookLM to play a role-playing game by giving it a 308 page manual Pretty good application of the rules, the character creation is very good (quoting accurately from 100 pages in) with small hallucinations, and the adventure is pretty solid! Take a listen to 4:00 on... https://t.co/7uMLAA3NEU

❀️243
likes
πŸ”30
retweets
πŸ–ΌοΈ Media
A
AI at Meta
@AIatMeta
πŸ“…
Nov 14, 2024
537d ago
πŸ†”23683716
⭐0.86

Whether you're at #EMNLP2024 in person or following from your feed, here are 5️⃣ research papers being presented by AI research teams at Meta to add to your reading list. 1️⃣ Distilling System 2 into System 1: https://t.co/bGWYvuAVzs 2️⃣ Altogether: Image Captioning via Re-aligning Alt-text: https://t.co/8kvewKBe9B 3️⃣ Beyond Turn-Based Interfaces: Synchronous LLMs for Full-Duplex Dialogue: https://t.co/mCKjjTEeKn 4️⃣ Memory-Efficient Fine-Tuning of Transformers via Token Selection: https://t.co/OTthVwdgmm 5️⃣ To the Globe (TTG): Towards Language-Driven Guaranteed Travel Planning: https://t.co/Yt6qray2sf

Media 1Media 2
+2 more
❀️338
likes
πŸ”91
retweets
πŸ–ΌοΈ Media
D
DAIR.AI
@dair_ai
πŸ“…
Nov 15, 2024
536d ago
πŸ†”58696410
⭐0.86

Complete Prompt Engineering Curriculum Whether you are building an AI agent or complex RAG systems, prompt engineering is key to build effectively. There are lots of approaches out there but in our new LLM prompting courses, we teach you about the methods and best practices that work in real-world LLM applications. Enroll in our academy to start learning: https://t.co/zQXQt0PMbG

Media 1
❀️153
likes
πŸ”43
retweets
πŸ–ΌοΈ Media
K
Kyle Cranmer
@KyleCranmer
πŸ“…
Nov 13, 2024
538d ago
πŸ†”09732915

I'm thrilled to announce that FranΓ§ois Charton (@f_charton @AIatMeta) will be kicking off our new AI for Science seminar series next Wednesday. He is at the forefront of using AI for mathematics, cryptography, and theoretical physics. @datascience_uw https://t.co/8wTojx76yt https://t.co/IkpbgPmws8

Media 1
❀️134
likes
πŸ”23
retweets
πŸ–ΌοΈ Media
O
elvis
@omarsar0
πŸ“…
Nov 15, 2024
536d ago
πŸ†”30130854
⭐1.00

Reasoning LLMs is one of the most interesting trends to watch going into 2025. I’ve been thinking a lot about how to build with reasoning LLMs, specifically agentic workflows. How can AI devs take advantage of components like MoA and MCTS when there is barely any research for it, not to mention the lack of insights and best practices? First, how do we enable devs to build with reasoning capabilities? I like how Nous Research is approaching this with their Forge Reasoning APIs and β€œReasoning Layer” components (MoA, MCTS, and Chain of Code). I think it’s way too early for such a reasoning layer but it seems that things are quickly moving in that direction; the o1 model series together with this forge reasoning API is a good indication of what’s to come. Some thoughts on the Forge Reasoning API vs o1 for language agents: I’ve been experimenting extensively with the o1 models and they are hard to customise. However, for many multi-agent systems, there is a need to get them to take on a persona that helps produce richer and more reliable outputs and facilitates better communication between agents. To achieve this, I often need to prompt the agents to behave and act a certain way and take on different roles depending on where they are in the conversation or process. Having the ability to use the right reasoning component or a combination, with configurable parameters (similar to the LLM itself), will be useful to build more complex and effective agentic systems. Customization is key here. Extended thoughts here: https://t.co/XglPppukqc

Media 1
❀️192
likes
πŸ”47
retweets
πŸ–ΌοΈ Media
E
Eugene Yan
@eugeneyan
πŸ“…
Nov 15, 2024
537d ago
πŸ†”73961510

Closing the feedback loop is so underrated and a key way to continuously improve our AI products and UXes https://t.co/uSin6eNZXe https://t.co/zqpvMAbkAz

Media 1
❀️67
likes
πŸ”8
retweets
πŸ–ΌοΈ Media
L
LDJ
@ldjconfirmed
πŸ“…
Nov 16, 2024
536d ago
πŸ†”65350869
⭐0.81

Spent time today digging in the literature for valuable benchmarks and found quite a few interesting ones that meet these criteria: - Has scores of real humans. - Shows models scoring much lower than humans. - Seems to test for fairly general ability, not silly spelling tricks. https://t.co/e4mmHRkIa4

Media 1
❀️69
likes
πŸ”14
retweets
πŸ–ΌοΈ Media
E
Ethan Mollick
@emollick
πŸ“…
Nov 15, 2024
536d ago
πŸ†”94094380

At least for non-experts: β€œAI-generated poems are now β€˜more human than human’… participants are more likely to judge that AI-generated poems are human-authored, compared to actual human-authored poems…. participants rate AI-generated poems more highly than human-written poems” https://t.co/cP5bk1KLjy

Media 1Media 2
❀️654
likes
πŸ”130
retweets
πŸ–ΌοΈ Media
I
Ivan Leo
@ivanleomk
πŸ“…
Nov 15, 2024
537d ago
πŸ†”13957837

Taught gemini how to highlight PDF text in a document :) Dropping in a bit in our docs https://t.co/1Vx4Lm08AJ

Media 1
❀️461
likes
πŸ”19
retweets
πŸ–ΌοΈ Media
M
merve
@mervenoyann
πŸ“…
Nov 14, 2024
537d ago
πŸ†”32123628
⭐0.86

Microsoft released LLM2CLIP: a CLIP model with longer context window for complex text inputs 🀯 TLDR; they replaced CLIP's text encoder with various LLMs fine-tuned on captioning, better top-k accuracy on retrieval πŸ”₯ All models with Apache 2.0 license on @huggingface 😍 https://t.co/xvwaWmZJj1

Media 1
❀️674
likes
πŸ”113
retweets
πŸ–ΌοΈ Media
H
Hamel Husain
@HamelHusain
πŸ“…
Nov 15, 2024
537d ago
πŸ†”03013145
⭐0.61

πŸ€£πŸ˜…πŸ˜‚ https://t.co/CTf7GqAjXF

@jeremyphoward β€’

@HamelHusain Nah this is good -- people who are enthusiastic to spam the buttons deserve more votes! (And people enthusiastic enough to write a script to spam them... even better :D )

Media 1
❀️16
likes
πŸ–ΌοΈ Media
M
Maxime Labonne
@maximelabonne
πŸ“…
Nov 15, 2024
537d ago
πŸ†”81018597
⭐0.86

OpenCoder doesn't get enough love They open-sourced the entire pipeline to create QwenCoder-level code models. This includes: - Large datasets - High-quality models - Eval framework Tons of great lessons and observations in the paper πŸ“ Paper: https://t.co/gh4DSeHj35 https://t.co/4geAGctnsC

Media 1
❀️536
likes
πŸ”103
retweets
πŸ–ΌοΈ Media
O
elvis
@omarsar0
πŸ“…
Nov 15, 2024
537d ago
πŸ†”18702118
⭐1.00

A Taxonomy of AgentOps for Enabling Observability of Foundation Model based Agents New research analyzes AgentOps platforms and tools, highlighting the need for comprehensive observability and traceability features to ensure reliability in foundation model-based autonomous agent systems across their development and production lifecycle.

Media 1
❀️332
likes
πŸ”76
retweets
πŸ–ΌοΈ Media
T
TuringPost
@TheTuringPost
πŸ“…
Nov 15, 2024
536d ago
πŸ†”39121169
⭐0.86

Mixture-of-Transformers (MoT) is a new @AIatMeta and @Stanford's MLLM design for efficient training of MLLMs using less computing power. It uses specific networks for each type of input: text, images, and speech, while still sharing attention across all the data. πŸ‘‡ https://t.co/nYdDECRnOR

Media 1Media 2
❀️86
likes
πŸ”23
retweets
πŸ–ΌοΈ Media