• English
  • Deutsch
  • Log In
    Password Login
    Research Outputs
    Fundings & Projects
    Researchers
    Institutes
    Statistics
Repository logo
Fraunhofer-Gesellschaft
  1. Home
  2. Fraunhofer-Gesellschaft
  3. Konferenzschrift
  4. Controlling the Perceived Sound Quality for Dialogue Enhancement with Deep Learning
 
  • Details
  • Full
Options
2020
Conference Paper
Title

Controlling the Perceived Sound Quality for Dialogue Enhancement with Deep Learning

Abstract
Speech enhancement attenuates interfering sounds in speech signals but may introduce artifacts that perceivably deteriorate the output signal. We propose a method for controlling the trade-off between the attenuation of the interfering background signal and the loss of sound quality. A deep neural network estimates the attenuation of the separated background signal such that the sound quality, quantified using the Artifact-related Perceptual Score, meets an adjustable target. Subjective evaluations indicate that consistent sound quality is obtained across various input signals. Our experiments show that the proposed method is able to control the tradeoff with an accuracy that is adequate for real-world dialogue enhancement applications.
Author(s)
Uhle, C.
Torcoli, M.
Paulus, J.
Mainwork
IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP 2020. Proceedings  
Conference
International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2020  
Open Access
DOI
10.1109/ICASSP40776.2020.9053789
Additional link
Full text
Language
English
Fraunhofer-Institut für Integrierte Schaltungen IIS  
  • Cookie settings
  • Imprint
  • Privacy policy
  • Api
  • Contact
© 2024