Professor of Signal Processing
Head of Department of Informatics
@kingscollegelondon.bsky.social
IEEE Workshop on Applications of Signal Processing to Audio and Acoustics (WASPAA)
WASPAA 2025 will be held Oct. 12-15, 2025 at Granlibakken Tahoe, Tahoe City, CA, USA.
Abstract deadline: April 23, 2025 (23:59 AOE)
Paper deadline: April 30, 2025 (23:59
San Diego Dec 2-7, 25 and Mexico City Nov 30-Dec 5, 25. Comments to this account are not monitored. Please send feedback to townhall@neurips.cc.
Automated posting of sound-related articles uploaded to arxiv.org (eess.AS + cs.SD)
Source: https://github.com/dsuedholt/bsky-paperbot-sound/
Inspired by @paperposterbot.bsky.social and https://twitter.com/ArxivSound
Multi-instrumentalist musician, sound engineer, audio software dev, ML engineer
Postdoc Researcher @Télécom Paris, Institut Polytechnique de Paris
Prev PhD @INRIA Prev intern @Meta @Baidu
I work on generative models and audio applications
https://xiaoyubie1994.github.io/
PhD student in multimodal learning for audio understanding at telecom-paris
Music machine learning, MIR, ML, DSP
PhD Student @ Telecom Paris, ADASP team. Previously research scientist intern @ Deezer, Sony CSL (Music Team).
AI/ML for audio and music signal processing and synthesis.
human computer musical instruments
https://hugofloresgarcia.art/
phd candidate @northwestern
research intern @adobe
prev @spotify, @descript
chicago // honduras
computers and music are (still) fun
Musician, Engineer, AI Researcher - @mitofficial.bsky.social @medialab.bsky.social
Efficient+Controllable Audio Generation @ UCSD | Interning Stability AI, Adobe | Teaching drums @ POW Percussion
Lecturer at the University of Edinburgh. Member of Centre of Speech Technology Research (CSTR).
Does research on machine learning at Sony AI, Barcelona. Works on audio analysis, synthesis, and retrieval. Likes tennis, music, and wine.
https://serrjoa.github.io/
Audio AI research engineer w/ Lemonaide. prev. Neutone, Okio. MSc in Audio Computation at UPF. I also fly planes and play the cello sometimes!
Assoc. Professor at UC Berkeley
Artificial and biological intelligence and language
Linguistics Lead at Project CETI 🐳
PI Berkeley Biological and Artificial Language Lab 🗣️
College Principal of Bowles Hall 🏰
https://www.gasperbegus.com
Associate professor @ Cornell Tech
AI researcher.
CTO at HANCE, Associate Professor at NTNU.
Compression, generative, audio, time series
Canadian in NYC (she/her) teaching music and data analysis at Brooklyn College and the Graduate Center, CUNY. Co-Editor-in-Chief of Journal of New Music Research.
Studying language in biological brains and artificial ones at the Kempner Institute at Harvard University.
www.tuckute.com
Principal Research Scientist at IBM Research AI in New York. Speech, Formal/Natural Language Processing. Currently LLM post-training, structured SDG and RL. Opinions my own and non stationary.
ramon.astudillo.com
We're the Music Understanding Synthesis and AI Creativity Group at UCSD! (PIs: Julian McAuley, Taylor Berg-Kirkpatrick, Shlomo Dubnov)
https://ucsd-musaic.github.io
Machine Learning for Music/Speech | Senior Research Engineer at Native Instruments/iZotope | Previously Intern @Microsoft, @Sony, @AudioshakeAI
Now: Audio & Multimodal ML PhD in the Music and Audio Research Lab @ NYU
Prev: Data Developer at Sonos and Northwestern, Research Intern at Adobe + Bosch Research
Research Scientist @SonyAI
PhD from Seoul National University
Previous intern @MERL, @Sony, and @Supertone
Postdoc researcher @telecomparis. Previously @CNRS/LS2N @c4dm. Machine learning for audio. https://changhongw.github.io/
Sr. Research Scientist at the Samsung AI Center in Cambridge. Affiliated lecturer at the University of Cambridge. SpeechBrain <3
I study machine listening methods for bioacoustics and automated sensing of natural environments. And I enjoy natural environments.
https://johnmartinsson.org/
Core member of Climate AI Nordics | ML researcher at RISE
unsound lab cat @ georgia tech
PhD Student @ltiatcmu.bsky.social
I work in speech processing.
wanchichen.github.io
Milanese-Californian Digital Speech and Audio Processing Technologist @ Apple
Official account for the SANE series of workshops. The one-day events annually gather researchers and students in speech and audio from the Northeast of the American continent, alternately in Boston and NYC.
🌐 saneworkshop.org
I’m a PhD student in University of Illinois Urbana-Champaign working on audio inverse problems.
My website: https://xzwy.github.io/alanweiyang.github.io/
Senior Manager, Foundational Research , @GoogleDeepMind
Googler, Ex @Dolby & @Broadcom
Talks and Investments 👉🏽 http://portfolio.v1vek.com
HCI Assistant Professor at UMich researching accessibility, audio AI, sound interaction, XR, and health. Director, Soundability Lab. Previously, Google, Apple, Microsoft, UW, and MIT Media Lab.
https://dhruv-jain.com
Assistant Professor at University of Michigan | PhD from UC San Diego | Human-Centered Generative AI for Content Creation
research on llm + music (https://seungheondoh.github.io/).
PhD Candidate @ Music and Audio Computing Lab, KAIST. Previously an intern @Adobe, @BytedanceTalk, @Naver, @Chartmetric.
Associate professor at Télécom Paris in machine listening and audio applied to extended reality
MS Music Tech @ Georgia Tech
Research Scientist @ Sound Scene Understanding Team, RIKEN-AIP, Japan
Audio and AI researcher. Faculty in Siebel School at UIUC and Visiting Academic at Amazon Lab126. A working dad. Some obsolete hobbies: music, photography, drawing, and writing. Still active interests: cooking.
🏠 https://minjekim.com
つくばあたりで音声・音響・聴覚の研究や教育をしています。
INGRESS ENL vengavenga
Speech • Language • Learning
https://grzegorz.chrupala.me
@ Tilburg University
Music and artificial intelligence.
Researcher at Stability AI.
Musician at BRNRT Collective.
Previously at Dolby and Universitat Pompeu Fabra.
artintech.substack.com
www.jordipons.me
AudioML research scientist at https://audioshake.ai, before: post-doc @inria@social.numerique.gouv.fr, Editor at https://bsky.app/profile/joss-openjournals.bsky.social
All in 17.68% of grey, located in Frankfurt (Germany)
🌮🤖 Speech and language modeling researcher.
Principal SWE @ Google Deepmind.
♊🌊 Gemini Audio and Astra core team.
http://rjryan.me/ https://google.github.io/tacotron
PhD student WAVLab@LTI, CMU
Multimodality and multilinguality
prev. predoc Google Deepmind
Physicist. Audio and deep learning research at Dolby Labs. Physics, audio, AI, science, technology and society.
Personal account @contraidees.bsky.social
Postdoctoral Researcher at KTH University, Sweden | 🤖🧠 Machine Learning, Time Series, Neuroscience & Dynamical Systems | Physics PhD from University of Buenos Aires
📚: https://scholar.google.com/citations?user=q5sweuIAAAAJ&hl=en
assistant professor at Warsaw University of Technology || music/audio AI researcher || experienced touring musician 🥁 || prev @apple
Machine Learning PhD Student
@ Blei Lab & Columbia University.
Working on probabilistic ML | uncertainty quantification | LLM interpretability.
Excited about everything ML, AI and engineering!
graduate student, BUT Speech, Brno, Czechia
Past: SPIRE Lab, IISc, India
AI Researcher @themetavoice
Interested in Audio & Music
Previously: Native Instruments, TikTok
ml, audio, cv, nlp, speech, bioacoustics // Assoc. Prof. at Université de Toulon, researcher at LIS CNRS UMR 7020, director of http://www.master-mir.eu in marine robotics and AI
interspeech2026.org
27 September – 1 October, ICC, Sydney, Australia
'Speaking Together'
Proudly hosted by the Australasian Speech Science and Technology Association (ASSTA) and the International Speech Communication Association (ISCA).
Here do discuss AI, Music, History, and much more!
Researcher in audio and speech generative models (SampleRNN, MelGAN, DAC, …)
Research Scientist @AdobeResearch. Ex @DescriptApp, @Mila_Quebec
https://ritheshkumar.com
Tenured Assistant Professor at CentraleSupélec.
Signal processing and machine learning for speech and audio.
sleglaive.github.io
Fixed-term researcher (RTDA) @polimi
working on audio signal processing, music informatics, spatial audio and generative models (https://lucacoma.github.io/)
Research Scientist at Google Deepmind working on audio/speech generation.
I'm working at CMU (2021-). I was working at NTT (2001-2011), MERL (2012-2017), and JHU (2017-2020). Speech and Audio Processing is my main research topic.
Challenge on Detection and Classification of Acoustic Scenes and Events.
https://dcase.community/
Head of Sound Design AI Research at Adobe. Machine learning and signal processing for audio & video. Musician. He/him.
www.justinsalamon.com
Postdoctoral researcher at Meta
Research scientist at #Inria. Audio signal processing, Acoustics, Machine Learning, Bicycle Riding, Lindy Hop Dancing.
Lecturer in speech and language technology, CSTR, University of Edinburgh.
https://homepages.inf.ed.ac.uk/clai/
machine learning researcher @Apple | PhD from @CoML_ENS | speech, ml and cognition.
Speech, language, and deep learning at the Technion. But also psychology, philosophy, and history. And Jazz improv.
I work on speech and language technologies at Google. I like languages, history, maps, traveling, cycling, and buying way too many books.
Guitarist, Researcher Google DeepMind. Opinions are my own.
Researcher at Adobe Research. Machine learning on audio. Screamer. Oaklander born in Barcelona. Titan. He/they 🌈
www.urinieto.com
Audio ML Research @ Auto-Tune 🎤🎵
Bay Area SSBM & RL gamer
Love to talk Cognitive Science, Linguistics, Bio-inspired Learning, Topological Signal Processing & TDA
Assistant professor at USherbrooke. Creator of the ODAS framework. Research in speech, multichannel audio processing, robot audition, embedded AI.
francoisgrondin.com
🎧 Machine Listening Researcher
Principal Scientist (Director) at Google DeepMind in Japan. 波瀬小⇒一志中⇒鈴鹿高専⇒名工大 (IBM T.J. Watson Research intern)⇒東芝欧州研究所⇒Google (Speech🇬🇧⇒Brain🇯🇵) ⇒Google DeepMind. 3rd generation Korean in Japan.
Professor/Admin @ Ohio State. All opinions expressed on this channel are my personal opinions and do not represent that of my employer.
Center for Language and Speech Processing at Johns Hopkins University
#NLProc #MachineLearning #AI http://tinyurl.com/clspy2ube
AI researcher in music, audio, LLMs.
Reader (Associate Professor), @qmuleecs.bsky.social Queen Mary University of London - research on AI for audio. Website: https://www.seresearch.qmul.ac.uk/cmai/people/ebenetos/
Research Scientist @ Meta GenAI in NYC.
Working on audio/speech for LLaMA.
Previously: PhD @ JHU CLSP
desh2608.github.io
Postdoctoral Researcher @ Inria Montpellier (IROKO, Pl@ntNet)
SSL for plant images
Interested in Computer Vision, Natural Language Processing, Machine Listening, and Biodiversity Monitoring
Website: ilyassmoummad.github.io
Auditory Signal Processing/Objective Metrics/Hearing Assistive Technologies.
Twitter: @kyama0321
WEB: https://sites.google.com/site/kyama0321/en
Scientist at CNRS.
https://audio.ls2n.fr
A science game to test your musical memory: https://tunetwins.app
AI for Music • Research Scientist @ Suno
Researcher in computer audition, machine learning, and HCI. Sr. Research Scientist, @AdobeResearch. Previously @DescriptApp, @Northwestern.
https://pseeth.github.io/
AI scientist & consultant :: prev Amazon Alexa, Toshiba, Cam Uni :: voice & language tech :: powered by coffee :: photographer :: Cambridge UK
https://www.catherinebreslin.co.uk
Full professor of inclusive speech communication at TU Delft, The Netherlands. Former president of the International Speech Communication Association (ISCA). Mother of 3🌈
Speech and audio research scientist @MERL. saneworkshop.org co-founder. IguanaTex developer.
🌐 jonathanleroux.org
🐙 github.com/Jonathan-LeRoux/
🎓 scholar.google.com/citations?user=aUpxty8AAAAJ&hl=en