The TC-STAR project is envisaged as a long-term effort to advance research in all core technologies for Speech-to-Speech Translation (SST). SST technology is a combination of Automatic Speech Recognition (ASR), Spoken Language Translation (SLT) and Text to Speech (TTS) (speech synthesis). The objectives of the project are ambitious: making a breakthrough in SST that significantly reduces the gap between human and machine translation performance. The project targets a selection of unconstrained conversational speech domains—speeches and broadcast news—and three languages: European English, European Spanish, and Mandarin Chinese. Accurate translation of unrestricted speech is well beyond the capability of today’s state-of-the-art research systems. Therefore, advances are needed to improve the state-of the-art technologies for speech recognition and speech translation.
Our pick of the week by @mgaido91: "AlignFormer: Modality Matching Can Achieve Better Zero-shot Instruction-Following Speech-LLM" by @RuchaoFan, Bo Ren, Yuxuan Hu, Rui Zhao, Shujie Liu, Jinyu Li (2024).
#NLProc #Speech #instructionfollowing #zeroshot #speechtech #speechllm
AI is transforming cultural heritage, but what have we learned?
Come and join the #AI4Culture movement at our Final Conference on March 10 in Hilversum to explore AI’s current & future impact on cultural heritage.
Details & Registration: https://pretix.eu/EFHA/AI4Culture/
@EU_HaDEA
BOUQuET💐: an OPEN INITIATIVE aimed at building an evaluation dataset for massively multilingual text-to-text MT.
Let’s make MT available for any written language!
We are inviting everyone to contribute: ➡️
More details at: https://arxiv.org/abs/2502.04314
I am happy to announce that I will speak about our recent work "How "Real" is Your Real-Time Simultaneous Speech-to-Text Translation System?" at the SlatorCon in March 🎊
📃 Preprint available here: