🐦 Twitter Post Details

Viewing enriched Twitter post

@yi_ding

A trend to watch: Retrieval in the Loop. Had a good discussion with @_superAGI @ishaanbhola @philipvollet @mlejva @silennai regarding this subject and finally got some time to get some thoughts down. We currently think of retrieval mostly as providing context in one shot, either at the beginning of a chat or the beginning of an agent process. However what we have found at @llama_index is that it’s often more useful to do retrieval at every step, either every chat in ContextChatEngine or every agent step in our Data Agents. Why? Because the LLM can almost inevitably benefit from new relevant context. Remember at this point ChatGPT’s knowledge cutoff is almost 2 years old. I saw @atroyn predict that in the future we will be doing retrieval on every token or even in the attention heads themselves. While I’m not sure they’ll go that far for latency reasons among others, I would very much welcome it if the LLM providers would formalize a “context message” just like the “system message” today. If you’re building a LLM application today the key thing to ask yourself is “would the LLM benefit from more context?” In most cases the answer is yes.

🔧 Raw API Response

{
  "user": {
    "created_at": "2008-05-09T19:07:43.000Z",
    "default_profile_image": false,
    "description": "Building LITS and DevRel @llama_index. Prev Messaging Apps @Apple, HFT @ GETCO, @Citadel",
    "fast_followers_count": 0,
    "favourites_count": 8951,
    "followers_count": 733,
    "friends_count": 1726,
    "has_custom_timelines": false,
    "is_translator": false,
    "listed_count": 15,
    "location": "San Jose, CA",
    "media_count": 59,
    "name": "Yi Ding -- prod/acc",
    "normal_followers_count": 733,
    "possibly_sensitive": false,
    "profile_image_url_https": "https://pbs.twimg.com/profile_images/1641119962977935360/pZLKP1HY_normal.jpg",
    "screen_name": "yi_ding",
    "statuses_count": 1337,
    "translator_type": "none",
    "verified": false,
    "withheld_in_countries": [],
    "id_str": "14716791"
  },
  "id": "1694475851239415869",
  "conversation_id": "1694475851239415869",
  "full_text": "A trend to watch: Retrieval in the Loop.\n\nHad a good discussion with @_superAGI @ishaanbhola @philipvollet @mlejva @silennai regarding this subject and finally got some time to get some thoughts down.\n\nWe currently think of retrieval mostly as providing context in one shot, either at the beginning of a chat or the beginning of an agent process.\n\nHowever what we have found at @llama_index is that it’s often more useful to do retrieval at every step, either every chat in ContextChatEngine or every agent step in our Data Agents. Why? Because the LLM can almost inevitably benefit from new relevant context. Remember at this point ChatGPT’s knowledge cutoff is almost 2 years old.\n\nI saw @atroyn predict that in the future we will be doing retrieval on every token or even in the attention heads themselves. While I’m not sure they’ll go that far for latency reasons among others, I would very much welcome it if the LLM providers would formalize a “context message” just like the “system message” today.\n\nIf you’re building a LLM application today the key thing to ask yourself is “would the LLM benefit from more context?” In most cases the answer is yes.",
  "reply_count": 2,
  "retweet_count": 7,
  "favorite_count": 32,
  "hashtags": [],
  "symbols": [],
  "user_mentions": [
    {
      "id_str": "1658049741111042048",
      "name": "SuperAGI",
      "screen_name": "_superAGI",
      "profile": "https://twitter.com/_superAGI"
    },
    {
      "id_str": "270505738",
      "name": "IshaanBhola",
      "screen_name": "ishaanbhola",
      "profile": "https://twitter.com/ishaanbhola"
    },
    {
      "id_str": "421795636",
      "name": "Philip Vollet",
      "screen_name": "philipvollet",
      "profile": "https://twitter.com/philipvollet"
    },
    {
      "id_str": "258300417",
      "name": "Vasek Mlejnsky",
      "screen_name": "mlejva",
      "profile": "https://twitter.com/mlejva"
    },
    {
      "id_str": "1090463126909304836",
      "name": "Silen Naihin",
      "screen_name": "silennai",
      "profile": "https://twitter.com/silennai"
    }
  ],
  "urls": [],
  "media": [
    {
      "media_url": "https://pbs.twimg.com/media/F4P8zg3aMAEdwNi.jpg",
      "type": "photo"
    }
  ],
  "url": "https://twitter.com/yi_ding/status/1694475851239415869",
  "created_at": "2023-08-23T22:24:59.000Z",
  "#sort_index": "1694475851239415869",
  "view_count": 5066,
  "quote_count": 1,
  "is_quote_tweet": false,
  "is_retweet": false,
  "is_pinned": false,
  "is_truncated": true,
  "startUrl": "https://twitter.com/yi_ding/status/1694475851239415869"
}