Quality Enhancement of Highly Degraded Music Using Deep Learning-Based Prediction Models for Lost Frequencies

  • Arthur C. Serra PUC-Rio
  • Antonio José G. Busson PUC-Rio
  • Álan L. V. Guedes PUC-Rio
  • Sérgio Colcher PUC-Rio

Resumo


Audio quality degradation can have many causes. For musical applications, this fragmentation may lead to highly unpleasant experiences. Restoration algorithms may be employed to reconstruct missing parts of the audio in a similar way as for image reconstruction — in an approach called audio inpainting. Current state-of-the art methods for audio inpainting cover limited scenarios, with well-defined gap windows and little variety of musical genres. In this work, we propose a Deep-Learning-based (DL-based) method for audio inpainting accompanied by a dataset with random fragmentation conditions that approximate real impairment situations. The dataset was collected using tracks from different music genres to provide a good signal variability. Our best model improved the quality of all musical genres, obtaining an average of 12.9 dB of PSNR, although it worked better for musical genres in which acoustic instruments are predominant.
Palavras-chave: Audio quality enhancement, Audio reconstruction, Neural Networks, Autoencoder
Publicado
05/11/2021
Como Citar

Selecione um Formato
SERRA, Arthur C.; BUSSON, Antonio José G.; GUEDES, Álan L. V.; COLCHER, Sérgio. Quality Enhancement of Highly Degraded Music Using Deep Learning-Based Prediction Models for Lost Frequencies. In: SIMPÓSIO BRASILEIRO DE SISTEMAS MULTIMÍDIA E WEB (WEBMEDIA), 1. , 2021, Minas Gerais. Anais [...]. Porto Alegre: Sociedade Brasileira de Computação, 2021 . p. 205-211.