Today's task: model compression!!
🎯 Goal: Compress a large, general-purpose multimodal model, making speech translation more efficient ⚡️, deployable 📲, and sustainable ♻️, while preserving translation quality ⭐️
#AI #SpeechTech #ModelCompression #LLMcompression
First up, a new task for 2025:
*Instruction-following for speech processing!*
Explore instruction-following for speech ⇨
Integrate speech foundation models with LLMs across tasks such as speech translation, recognition, summarization, and QA.
🔗:
📢Workshop gratuito 05/02: “Lo stato dell'arte nelle tecnologie per il riconoscimento del parlato.”
Diretta YouTube: https://www.youtube.com/live/i4x7w8fIIXo?si=wYvvrO3-MSh7Yik4
Registrazione: https://www.eventbrite.com/e/biglietti-lo-stato-dellarte-nelle-tecnologie-per-il-riconoscimento-del-parlato-1109098797359?aff=oddtdtcreator
I'm happy to share that our paper "Prepending or Cross-Attention for Speech-to-Text? An Empirical Comparison" has been accepted at @naacl @naaclmeeting 2025! #NAACL2025
@Lam19Tk @mgaido91 👏
📃 Preprint:
⏰ Code will be released soon
#NLProc #Speech