The TC-STAR project is envisaged as a long-term effort to advance research in all core technologies for Speech-to-Speech Translation (SST). SST technology is a combination of Automatic Speech Recognition (ASR), Spoken Language Translation (SLT) and Text to Speech (TTS) (speech synthesis). The objectives of the project are ambitious: making a breakthrough in SST that significantly reduces the gap between human and machine translation performance. The project targets a selection of unconstrained conversational speech domains—speeches and broadcast news—and three languages: European English, European Spanish, and Mandarin Chinese. Accurate translation of unrestricted speech is well beyond the capability of today’s state-of-the-art research systems. Therefore, advances are needed to improve the state-of the-art technologies for speech recognition and speech translation.
🙌🏼 Excited to share our work on Speech Foundation Model for data crowdsourcing at COLING 2025 🙌🏼
Our co-author Laurent Besacier (@laurent_besacie) at NAVER LABS Europe will be presenting -- don't miss it.
👉🏼 Details: https://mt.fbk.eu/1-paper-accepted-at-coling-2025
Exciting news: @iwslt is co-located with #ACL2025NLP again this year! 🎉
Interested in speech processing? Check out the new task on instruction following — any model can participate! 🚀
📅 Data release: April 1
⏳ Submission deadline: April 15
Don’t miss it! 💬 #NLP #SpeechTech
Weekly pick from the #MeetweenScientificWatch: “Video-SALMONN: Speech-enhanced audio-visual large language models” – Redefining video comprehension with speech-aware AV-LLMs and groundbreaking QA accuracy. 🎥🎤🤖
I’m glad to announce that our work “How "Real" is Your Real-Time Simultaneous Speech-to-Text Translation System?” has been accepted at the Transactions of @aclanthology (TACL)! 🎉
The preprint is available here: