🐦 Twitter Post Details

Viewing enriched Twitter post

@DrJimFan

A viral paper "Language Model Represents Space and Time" recently claims that LLMs learn "world models". As much as I like @tegmark's works, I disagree with their definition of world model. World model is a core concept in AI agent and decision making. It is our mental simulation of how the world works given interventions (or lack thereof). A world model captures causality and intuitive physics, telling the agent what is likely and what is impossible. It can and should be used for counterfactual reasoning, i.e. "what ifs": what would happen if I knock over a cup of water? Where would I have been if I had not taken that bus? Yann LeCun @ylecun says it well in his position paper (https://t.co/MJxLffbK5Q). I quote: "Using such world models, animals can learn new skills with very few trials. They can predict the consequences of their actions, they can reason, plan, explore, and imagine new solutions to problems. Importantly, they can also avoid making dangerous mistakes when facing an unknown situation." The first use of the term World Model in deep policy learning is attributed to @hardmaru & @SchmidhuberAI: https://t.co/tWDuQRNTRh. In their seminal paper, an agent masters shooting skills in the popular game Doom (demo below) by learning in imagination, using an internal world model as a "physics simulator". To put in a simple Python math formula, world model learns a function F(s[0:t-1], a) -> s[t:], which takes as input the observed past and current action, and outputs plausible future states. Now the definition of World Model in Tegmark's paper seems to be about predicting GPS coordinates and time eras. I see this as just a classification task with no causal learning and simulation going on. You cannot make meaningful interventions against that model, nor can you optimize any decision making in a closed feedback loop. As for the "space & time neurons", I think they are most similar to the "sentiment neuron" that OpenAI published in 2017: https://t.co/QFnP2pjUSQ. Predicting GPS is conceptually no different from predicting sentiment in my opinion. I don't think their experimental results are wrong - just that their conclusion is on shaky grounds. I welcome any debate! Paper link: https://t.co/4ly12nPS1N

📊 Media Metadata

{
  "media": [
    {
      "type": "video",
      "url": "https://pbs.twimg.com/ext_tw_video_thumb/1709942174044065792/pu/img/SIYXjIFHPTooRg18.jpg",
      "media_url": "https://pbs.twimg.com/ext_tw_video_thumb/1709942174044065792/pu/img/SIYXjIFHPTooRg18.jpg",
      "filename": "media_0.mp4"
    }
  ],
  "nlp": {
    "processed_at": "2025-08-06T12:22:24.918506",
    "sentiment": "negative",
    "topics": [
      "LLMs",
      "Machine Learning",
      "Deep Learning",
      "Neural Networks",
      "AI Safety",
      "NLP"
    ],
    "ner": {
      "entities": [
        {
          "entity": "Language Model Represents Space and Time",
          "type": "paper"
        },
        {
          "entity": "LLMs",
          "type": "acronym"
        },
        {
          "entity": "world models",
          "type": "concept"
        },
        {
          "entity": "AI agent",
          "type": "concept"
        },
        {
          "entity": "decision making",
          "type": "concept"
        },
        {
          "entity": "causality",
          "type": "concept"
        },
        {
          "entity": "intuitive physics",
          "type": "concept"
        },
        {
          "entity": "counterfact",
          "type": "concept"
        },
        {
          "entity": "tegmark",
          "type": "person"
        }
      ]
    }
  },
  "original_structure": "had_media_only",
  "enhanced_from_raw_response": true,
  "enhanced_at": "2025-08-13T17:20:00Z"
}

🔧 Raw API Response

{
  "user": {
    "created_at": "2012-12-12T22:11:27.000Z",
    "default_profile_image": false,
    "description": "@NVIDIA Senior AI Scientist. @Stanford PhD. Join me on the frontier of AI Agents, LLM & Robotics. MineDojo (NeurIPS Best Paper), Voyager. Ex: @OpenAI, @GoogleAI",
    "fast_followers_count": 0,
    "favourites_count": 6203,
    "followers_count": 142744,
    "friends_count": 2804,
    "has_custom_timelines": true,
    "is_translator": false,
    "listed_count": 2880,
    "location": "Views my own. Get in touch →",
    "media_count": 638,
    "name": "Jim Fan",
    "normal_followers_count": 142744,
    "possibly_sensitive": false,
    "profile_banner_url": "https://pbs.twimg.com/profile_banners/1007413134/1672408318",
    "profile_image_url_https": "https://pbs.twimg.com/profile_images/1554922493101559808/SYSZhbcd_normal.jpg",
    "screen_name": "DrJimFan",
    "statuses_count": 2957,
    "translator_type": "none",
    "url": "https://t.co/H4rXo4Ei8X",
    "verified": false,
    "withheld_in_countries": [],
    "id_str": "1007413134"
  },
  "id": "1709947595525951787",
  "conversation_id": "1709947595525951787",
  "full_text": "A viral paper \"Language Model Represents Space and Time\" recently claims that LLMs learn \"world models\". As much as I like @tegmark's works, I disagree with their definition of world model.\n\nWorld model is a core concept in AI agent and decision making. It is our mental simulation of how the world works given interventions (or lack thereof). \n\nA world model captures causality and intuitive physics, telling the agent what is likely and what is impossible. It can and should be used for counterfactual reasoning, i.e. \"what ifs\": what would happen if I knock over a cup of water? Where would I have been if I had not taken that bus?\n\nYann LeCun @ylecun says it well in his position paper (https://t.co/MJxLffbK5Q). I quote:\n\n\"Using such world models, animals can learn new skills with very few trials. They can predict the consequences of their actions, they can reason, plan, explore, and imagine new solutions to problems. Importantly, they can also avoid making dangerous mistakes when facing an unknown situation.\"\n\nThe first use of the term World Model in deep policy learning is attributed to @hardmaru & @SchmidhuberAI:  https://t.co/tWDuQRNTRh. In their seminal paper, an agent masters shooting skills in the popular game Doom (demo below) by learning in imagination, using an internal world model as a \"physics simulator\".\n\nTo put in a simple Python math formula, world model learns a function F(s[0:t-1], a) -> s[t:], which takes as input the observed past and current action, and outputs plausible future states.\n\nNow the definition of World Model in Tegmark's paper seems to be about predicting GPS coordinates and time eras. I see this as just a classification task with no causal learning and simulation going on. You cannot make meaningful interventions against that model, nor can you optimize any decision making in a closed feedback loop.\n\nAs for the \"space & time neurons\", I think they are most similar to the \"sentiment neuron\" that OpenAI published in 2017: https://t.co/QFnP2pjUSQ. Predicting GPS is conceptually no different from predicting sentiment in my opinion. I don't think their experimental results are wrong - just that their conclusion is on shaky grounds.\n\nI welcome any debate! Paper link: https://t.co/4ly12nPS1N",
  "reply_count": 74,
  "retweet_count": 213,
  "favorite_count": 1219,
  "hashtags": [],
  "symbols": [],
  "user_mentions": [
    {
      "id_str": "2530947115",
      "name": "Max Tegmark",
      "screen_name": "tegmark",
      "profile": "https://twitter.com/tegmark"
    }
  ],
  "urls": [],
  "media": [
    {
      "media_url": "https://pbs.twimg.com/ext_tw_video_thumb/1709942174044065792/pu/img/SIYXjIFHPTooRg18.jpg",
      "type": "video",
      "video_url": "https://video.twimg.com/ext_tw_video/1709942174044065792/pu/vid/avc1/320x240/2dZo_BJVzHtMmXW_.mp4?tag=12"
    }
  ],
  "url": "https://twitter.com/DrJimFan/status/1709947595525951787",
  "created_at": "2023-10-05T15:04:10.000Z",
  "#sort_index": "1710325869213020400",
  "view_count": 523433,
  "quote_count": 37,
  "is_quote_tweet": false,
  "is_retweet": false,
  "is_pinned": false,
  "is_truncated": true,
  "startUrl": "https://twitter.com/drjimfan/status/1710325869213020336"
}