• English
  • Deutsch
  • Log In
    Password Login
    or
  • Research Outputs
  • Projects
  • Researchers
  • Institutes
  • Statistics
Repository logo
Fraunhofer-Gesellschaft
  1. Home
  2. Fraunhofer-Gesellschaft
  3. Konferenzschrift
  4. Training phoneme models for singing with "songified" speech data
 
  • Details
  • Full
Options
2015
Conference Paper
Titel

Training phoneme models for singing with "songified" speech data

Abstract
Speech recognition in singing is a task that has not been widely researched so far. Singing possesses several characteristics that differentiate it from speech. Therefore, algorithms and models that were developed for speech usually perform worse on singing. One of the bottlenecks in many algorithms is the recognition of phonemes in singing. We noticed that this recognition step can be improved when using singing data in model training, but to our knowledge, there are no large datasets of singing data annotated with phonemes. However, such data does exist for speech. We therefore propose to make phoneme recognition models more robust for singing by training them on speech data that has artificially been made more ""song-like"". We test two main modifications on speech data: Time stretching and pitch shifting. Artificial vibrato is also tested. We then evaluate models trained on different combinations of these modified speech recordings. The utilized modeling algorithms are Neural Networks and Deep Belief Networks.
Author(s)
Kruspe, Anna M.
Hauptwerk
16th International Society for Music Information Retrieval Conference, ISMIR 2015. Proceedings
Konferenz
International Society for Music Information Retrieval (ISMIR Conference) 2015
Thumbnail Image
Externer Link
Externer Link
Language
English
google-scholar
Fraunhofer-Institut für Digitale Medientechnologie IDMT
Tags
  • vocal analysis

  • phoneme recognition

  • Cookie settings
  • Imprint
  • Privacy policy
  • Api
  • Send Feedback
© 2022