🐦 Twitter Post Details

Viewing enriched Twitter post

@AlphaSignalAI

Big news. Meta just announced SeamlessM4T, a competitor to Google translate. *turn sound on* SeamlessM4T is a multimodal foundational speech/text translation and transcription model capable of handling: 📥 101 languages for speech input ⌨️ 96 Languages for text input/output 🗣️ 35 languages for speech output It achieves SOTA by using Fairseq2, a new modeling toolkit, and the largest open dataset for multimodal translation, totaling 470k hours. This unified model enables multiple tasks without relying on multiple separate models: ▸ Speech-to-speech translation (S2ST) ▸ Speech-to-text translation (S2TT) ▸ Text-to-speech translation (T2ST) ▸ Text-to-text translation (T2TT) ▸ Automatic speech recognition (ASR)

🔧 Raw API Response

{
  "user": {
    "created_at": "2012-11-07T07:19:36.000Z",
    "default_profile_image": false,
    "description": "Covering the latest breakthroughs in AI. ML Engineer/Researcher now building AlphaSignal → A technical newsletter read by 120,000+ AI developers.",
    "fast_followers_count": 0,
    "favourites_count": 4230,
    "followers_count": 64949,
    "friends_count": 702,
    "has_custom_timelines": true,
    "is_translator": false,
    "listed_count": 1408,
    "location": "Join 120,000+ readers →",
    "media_count": 359,
    "name": "Lior⚡",
    "normal_followers_count": 64949,
    "possibly_sensitive": false,
    "profile_banner_url": "https://pbs.twimg.com/profile_banners/931470139/1681303371",
    "profile_image_url_https": "https://pbs.twimg.com/profile_images/1599792074336964608/CobSHV8l_normal.jpg",
    "screen_name": "AlphaSignalAI",
    "statuses_count": 2400,
    "translator_type": "none",
    "url": "https://t.co/AyubevadmD",
    "verified": false,
    "withheld_in_countries": [],
    "id_str": "931470139"
  },
  "id": "1694416112203534819",
  "conversation_id": "1694416112203534819",
  "full_text": "Big news. Meta just announced SeamlessM4T, a competitor to Google translate. *turn sound on*\n\nSeamlessM4T is a multimodal foundational speech/text translation and transcription model capable of handling:\n\n📥 101 languages for speech input\n⌨️ 96 Languages for text input/output\n🗣️ 35 languages for speech output\n\nIt achieves SOTA by using Fairseq2, a new modeling toolkit, and the largest open dataset for multimodal translation, totaling 470k hours.\n\nThis unified model enables multiple tasks without relying on multiple separate models:\n\n▸ Speech-to-speech translation (S2ST)\n▸ Speech-to-text translation (S2TT)\n▸ Text-to-speech translation (T2ST)\n▸ Text-to-text translation (T2TT)\n▸ Automatic speech recognition (ASR)",
  "reply_count": 6,
  "retweet_count": 23,
  "favorite_count": 87,
  "hashtags": [],
  "symbols": [],
  "user_mentions": [],
  "urls": [],
  "media": [
    {
      "media_url": "https://pbs.twimg.com/ext_tw_video_thumb/1694415245362491392/pu/img/YfcfsPQSKMO76vj5.jpg",
      "type": "video",
      "video_url": "https://video.twimg.com/ext_tw_video/1694415245362491392/pu/vid/1264x720/jndBrOvY6tZYb_AZ.mp4?tag=12"
    }
  ],
  "url": "https://twitter.com/AlphaSignalAI/status/1694416112203534819",
  "created_at": "2023-08-23T18:27:36.000Z",
  "#sort_index": "1694416112203534819",
  "view_count": 8620,
  "quote_count": 3,
  "is_quote_tweet": false,
  "is_retweet": false,
  "is_pinned": false,
  "is_truncated": true,
  "startUrl": "https://twitter.com/alphasignalai/status/1694416112203534819"
}