Beomseok Lee

PhD Student

    Short bio

    Beomseok Lee is a PhD student at the University of Trento, conducting research at Fondazione Bruno Kessler and NAVER LABS Europe.

    Before starting his PhD, Beomseok worked as a full-time research engineer at SAMSUNG Research (Samsung Electronics R&D hub) Global AI Center, Seoul where he specialized in End-to-end (E2E) Speech-to-text translation. His current research focuses on E2E Spoken Language Understanding with an emphasis on multi-task, multi-lingual and multi-modal approaches. He holds a Computer Science (CS) Master's degree from Korea Advanced Institute of Science & Technology (KAIST, Korea) and a CS Bachelor's degree from Sungkyunkwan University (SKKU, Korea).

    Research topics

    Spoken Language Understanding, Multi-modality

    Publications

    1. Lee, Beomseok; Calapodescu, Ioan; Gaido, Marco; Negri, Matteo; Besacier, Laurent,
      in «»,
      Proceedings of Interspeech2024,
      ,
      vol. ,
      n. ,
      2024
      , pp. 817-
      821

    πŸ™ŒπŸΌ Excited to share our work on Speech Foundation Model for data crowdsourcing at COLING 2025 πŸ™ŒπŸΌ

    Our co-author Laurent Besacier (@laurent_besacie) at NAVER LABS Europe will be presenting -- don't miss it.

    πŸ‘‰πŸΌ Details: https://mt.fbk.eu/1-paper-accepted-at-coling-2025

    Exciting news: @iwslt is co-located with #ACL2025NLP again this year! πŸŽ‰
    Interested in speech processing? Check out the new task on instruction following β€” any model can participate! πŸš€
    πŸ“… Data release: April 1
    ⏳ Submission deadline: April 15
    Don’t miss it! πŸ’¬ #NLP #SpeechTech

    Weekly pick from the #MeetweenScientificWatch: β€œVideo-SALMONN: Speech-enhanced audio-visual large language models” – Redefining video comprehension with speech-aware AV-LLMs and groundbreaking QA accuracy. πŸŽ₯πŸŽ€πŸ€–

    I’m glad to announce that our work β€œHow "Real" is Your Real-Time Simultaneous Speech-to-Text Translation System?” has been accepted at the Transactions of @aclanthology (TACL)! πŸŽ‰

    The preprint is available here:

    Load More