Last week at the @fbk-mt.bsky.social seminars, we hosted Elizabeth Salesky from Google DeepMind, presenting her work on "Translation and Language Modeling with Pixels"
#NLProc #tokenization #MT
Last week at the @fbk-mt.bsky.social seminars, we hosted Elizabeth Salesky from Google DeepMind, presenting her work on "Translation and Language Modeling with Pixels"
#NLProc #tokenization #MT
🚀 JOB ALERT 3: The FBK's MT Unit is hiring!
Join us as a Researcher in Responsible & Trustworthy NLP and advance ethical, fair, and transparent language technologies. If you care about building safe and accountable AI systems, you can apply here:
👉 jobs.fbk.eu/Annunci/Offe...
🚀 We're hiring a Researcher in Responsible & Trustworthy NLP! Join our research group @fbk-mt.bsky.social at Fondazione Bruno Kessler to work on fairness and trustworthiness in multilingual technologies.
📅 Deadline: Dec 10, 2025
🔗 Apply: jobs.fbk.eu/Annunci/Offe...
🚀 New paper: Speech Discrete Tokens or Continuous Features?
📄 aclanthology.org/2025.emnlp-m...
🧩 A comprehensive benchmark of SpeechLLMs using HuBERT/WavLM with Qwen & LLaMA.
✨ Continuous features outperform overall, while discrete tokens excel at phoneme-level detail.
🚀 Exciting news from the @fbk-mt.bsky.social group!
@bsavoldi.bsky.social , @linaconti.bsky.social, @matteo-negri.bsky.social & @luisabentivogli.bsky.social are attending #EMNLP2025 in Suzhou 🇨🇳!
Come to our sessions & let's connect:
🔗 mt.fbk.eu/fbk-mt-at-em...
We’re also hiring postdocs!⚡
🚀 SimulMEGA: MoE Routers as advanced policy makers for Simultaneous Speech Translation 🎧🌍
Mixture-of-Experts routing → smarter decisions on when & how to translate, balancing latency vs quality in real-time speech. Paper link at arxiv.org/pdf/2509.012...
🚀 AdvST: Adversarial training aligns speech and text distributions without parallel data! Combines adversarial learning + hidden-state swapping to fix length mismatch & boost low-resource speech translation. ieeexplore.ieee.org/document/108...
🚀 Boost rare-phrase translation in speech! Uses **bilingual dictionaries** (e.g., "climate change"→"Klimawandel") to dynamically bias outputs.
✅ **+21%** recall in streaming ST
✅ **+85%** in multimodal LLMs
🔗: arxiv.org/abs/2506.09175
🔍 Stiamo studiando come l'AI viene usata in Italia e per farlo abbiamo costruito un sondaggio!
👉 bit.ly/sondaggio_ai...
(è anonimo, richiede ~10 minuti, e se partecipi o lo fai girare ci aiuti un sacco🙏)
Ci interessa anche raggiungere persone che non si occupano e non sono esperte di AI!
📢 Come and join our group!
We offer a fully funded 3-year PhD position:
📔 Automatic translation with large multimodal models: iecs.unitn.it/education/ad...
📍Full details for application: iecs.unitn.it/education/ad...
📅 Deadline May 12, 2025
#NLProc #FBK
ReShape Attention bridges speech & text models without extra parameters. Achieves +8.5% BLEU in translation by leveraging acoustic cues, outperforming cascade/E2E methods. Efficient & scalable. Check the paper by Kano et al. (2025) at: ieeexplore.ieee.org/stamp/stamp.....
New research fuels the debate between cascaded and E2E speech translation! The challenge of error propagation is addressed by incorporating multiple ASR candidates, along with HuBERT features to preserve acoustic information lost after ASR. Check the paper by Min et al. at: arxiv.org/pdf/2502.00377.