Anti-Stotter

Assistive Anti-Stutter AI-Technology
(BMFTR 2024-2026)

Around 80 million people worldwide (including around 800,000 in Germany alone) are affected by chronic stuttering. This neurological disorder is currently incurable. Despite speech therapy, which more than 70 percent of those affected undergo, speech flow usually remains impaired for life. Stuttering often leads to severe psychosocial stress and significantly reduces the quality of life of those affected.

The project team is developing an inconspicuous in-ear headphone system that can improve the speech flow of people who stutter immediately and without effort. The system uses artificial intelligence (AI) in speech synthesis to generate audio feedback while speaking. This feedback specifically activates a neurocognitive mechanism in the brain that bypasses the pathological component of stuttering. By aligning the system with the neural principles of human hearing, it is possible for the first time to improve speech fluency even with long-term use. In addition, laboratory and field studies are being conducted to comprehensively evaluate the effectiveness and suitability of the system for everyday use. The result promises to be a pioneering solution that could represent a breakthrough in the technology-assisted treatment of stuttering.

project website

offical project brief

Project Partners

References

2025

  1. StutterCut: Uncertainty-Guided Normalised Cut for Dysfluency Segmentation
    Suhita Ghosh, Melanie Jouaiti, Jan-Ole Perschewski, and Sebastian Stober
    In Interspeech 2025, Aug 2025
  2. Investigating Inclusivity of Whisper for Dysfluent Speech
    Evelyn Starzew, Suhita Ghosh, and Valerie Krug
    In 12th edition of the Disfluency in Spontaneous Speech Workshop (DiSS 2025), Sep 2025

2024

  1. Anonymising Elderly and Pathological Speech: Voice Conversion Using DDSP and Query-by-Example
    Suhita Ghosh, Melanie Jouaiti, Arnab Das, Yamini Sinha, Tim Polzehl, Ingo Siegert, and Sebastian Stober
    In Interspeech 2024, Sep 2024
  2. Improving Voice Quality in Speech Anonymization With Just Perception-Informed Losses
    Suhita Ghosh, Tim Thiele, Frederic Lorbeer, and Sebastian Stober
    In Audio Imagination: NeurIPS 2024 Workshop AI-Driven Speech, Music, and Sound Generation, 2024
  3. T-DVAE: A Transformer-Based Dynamical Variational Autoencoder for Speech
    Jan-Ole Perschewski and Sebastian Stober
    In Artificial Neural Networks and Machine Learning – ICANN 2024, 2024

2023

  1. Improving voice conversion for dissimilar speakers using perceptual losses
    Suhita Ghosh, Yamini Sinha, Ingo Siegert, and Sebastian Stober
    In 49. Jahrestagung für Akustik DAGA 2023, Hamburg, Mar 2023
  2. Anonymization of Stuttered Speech – Removing Speaker Information while Preserving the Utterance
    Jan Hintz, Sebastian Bayerl, Yamini Sinha, Suhita Ghosh, Martha Schubert, Sebastian Stober, Korbinian Riedhammer, and Ingo Siegert
    In 3rd Symposium on Security and Privacy in Speech Communication, Aug 2023
  3. StarGAN-VC++: Towards Emotion Preserving Voice Conversion Using Deep Embeddings
    Arnab Das, Suhita Ghosh, Tim Polzehl, Ingo Siegert, and Sebastian Stober
    In 12th ISCA Speech Synthesis Workshop (SSW2023), Aug 2023
  4. Emo-StarGAN: A Semi-Supervised Any-to-Many Non-Parallel Emotion-Preserving Voice Conversion
    Suhita Ghosh, Arnab Das, Yamini Sinha, Ingo Siegert, Tim Polzehl, and Sebastian Stober
    In INTERSPEECH 2023, Aug 2023

2022

  1. Voice Privacy - leveraging multi-scale blocks with ECAPA-TDNN SE-Res2NeXt extension for speaker anonymization
    Razieh Khamsehashari, Yamini Sinha, Jan Hintz, Suhita Ghosh, Tim Polzehl, Clarlos Franzreb, Sebastian Stober, and Ingo Siegert
    In 2nd Symposium on Security and Privacy in Speech Communication, Sep 2022