autrainer
A Modular and Extensible Deep Learning Toolkit for Computer Audition Tasks. (GitHub) 
Published in IEEE Transactions on Pattern Analysis and Machine Intelligence, 2023
Our top-cited work on speech emotion recognition.
Recommended citation: Wagner, J., Triantafyllopoulos, A., et al. (2023). "Dawn of the Transformer Era in Speech Emotion Recognition: Closing the Valence Gap" IEEE Transactions on Pattern Analysis and Machine Intelligence. vol. 45, no. 9, pp. 10745-10759.
Download Paper
Published in Proceedings of the IEEE, 2023
A narrative review of affective speech synthesis.
Recommended citation: Triantafyllopoulos, A., et al. (2023). "An Overview of Affective Speech Synthesis and Conversion in the Deep Learning Era" Proceedings of the IEEE. vol. 111, no. 10, pp. 1355-1381.
Download Paper
Published in Proceedings of the IEEE (forthcoming), 2025
A narrative review of audio foundation models.
Recommended citation: Triantafyllopoulos, A. et al. (2025). "Computer audition: From task-specific machine learning to foundation models." Proceedings of the IEEE. (forthcoming).
Download Paper
Published:
Overview of digital health in noisy, industrial settings (https://www.tapas-etn-eu.org/events/te6)
Published:
Quick overview of the principles underlying adaptive denoising
Published:
Simple overview of affective computing and its connection to psychology research and praxis
Published:
Overview of our efforts on low-footprint speech denoising and detection of extracochlear electrode placement
Graduate course (10 ECTS), Technical University of Munich, Chair of Health Informatics, 2025
Our quarterly practical course on audio, speech, and language.
Learning to listen by reward.
Using audio to monitor the environment.
Using speech for the diagnosis and treatment of mental disorders.
Using speech for the diagnosis of respiratory diseases.
Using speech to recognize emotions.
Decomposing an audio stream to its individual salient components.