🐦 Twitter Post Details

Viewing enriched Twitter post

@arankomatsuzaki

Sequoia: Scalable, Robust, and Hardware-aware Speculative Decoding Improves the decoding speed of Vicuna-33B by up to 2.37x and Llama2-70B offloading speed by up to 10.33x https://t.co/4Tws6hP3E2 https://t.co/W2HwiIsX8e

Media 1

📊 Media Metadata

{
  "data": [
    {
      "id": "",
      "type": "photo",
      "url": null,
      "media_url": "https://pbs.twimg.com/media/GGv80D2XYAAaJvI.jpg",
      "media_url_https": null,
      "display_url": null,
      "expanded_url": null
    }
  ],
  "score": 0.934,
  "scored_at": "2025-08-09T13:46:07.552137",
  "import_source": "manual_curation_2024",
  "links_checked": true,
  "checked_at": "2025-08-10T10:31:58.771663",
  "media": [
    {
      "type": "photo",
      "url": "https://crmoxkoizveukayfjuyo.supabase.co/storage/v1/object/public/media/posts/1759778428801187871/media_0.jpg?",
      "filename": "media_0.jpg",
      "original_url": "https://pbs.twimg.com/media/GGv80D2XYAAaJvI.jpg"
    }
  ],
  "storage_migrated": true
}

🔧 Raw API Response

{
  "user": {
    "created_at": "2016-11-04T06:57:37.000Z",
    "default_profile_image": false,
    "description": "ML research & startup with @EnricoShippole",
    "fast_followers_count": 0,
    "favourites_count": 9848,
    "followers_count": 89404,
    "friends_count": 78,
    "has_custom_timelines": true,
    "is_translator": false,
    "listed_count": 1193,
    "location": "",
    "media_count": 1831,
    "name": "Aran Komatsuzaki",
    "normal_followers_count": 89404,
    "possibly_sensitive": false,
    "profile_image_url_https": "https://pbs.twimg.com/profile_images/1561220982328754176/JOYS5kab_normal.jpg",
    "screen_name": "arankomatsuzaki",
    "statuses_count": 4455,
    "translator_type": "none",
    "url": "https://t.co/aZGCShnLYq",
    "verified": true,
    "withheld_in_countries": [],
    "id_str": "794433401591693312"
  },
  "id": "1759778428801187871",
  "conversation_id": "1759778428801187871",
  "full_text": "Sequoia: Scalable, Robust, and Hardware-aware Speculative Decoding\n\nImproves the decoding speed of Vicuna-33B by up to 2.37x and Llama2-70B offloading speed by up to 10.33x\n\nhttps://t.co/4Tws6hP3E2 https://t.co/W2HwiIsX8e",
  "reply_count": 3,
  "retweet_count": 21,
  "favorite_count": 86,
  "hashtags": [],
  "symbols": [],
  "user_mentions": [],
  "urls": [
    {
      "url": "https://t.co/4Tws6hP3E2",
      "expanded_url": "https://arxiv.org/abs/2402.12374",
      "display_url": "arxiv.org/abs/2402.12374"
    }
  ],
  "media": [
    {
      "media_url": "https://pbs.twimg.com/media/GGv80D2XYAAaJvI.jpg",
      "type": "photo"
    }
  ],
  "url": "https://twitter.com/arankomatsuzaki/status/1759778428801187871",
  "created_at": "2024-02-20T03:14:07.000Z",
  "#sort_index": "1759778428801187871",
  "view_count": 10309,
  "quote_count": 3,
  "is_quote_tweet": false,
  "is_retweet": false,
  "is_pinned": false,
  "is_truncated": false,
  "startUrl": "https://twitter.com/arankomatsuzaki/status/1759778428801187871"
}