Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Active Restoration of Lost Audio Signals Using Machine Learning and Latent Information
Université Frères Mentouri I, Algeria.
Blekinge Institute of Technology, Faculty of Computing, Department of Computer Science.ORCID iD: 0000-0002-4390-411X
2024 (English)In: Intelligent Systems and Applications: Proceedings of the 2023 Intelligent Systems Conference (IntelliSys) Volume 1 / [ed] Kohei Arai, Springer, 2024, 822, Vol. 822, p. 1-16Conference paper, Published paper (Refereed)
Abstract [en]

Digital audio signal reconstruction of a lost or corrupt segment using deep learning algorithms has been explored intensively in recent years. Nevertheless, prior traditional methods with linear interpolation, phase coding and tone insertion techniques are still in vogue. However, we found no research work on reconstructing audio signals with the fusion of dithering, steganography, and machine learning regressors. Therefore, this paper proposes the combination of steganography, halftoning (dithering), and state-of-the-art shallow and deep learning methods. The results (including comparing the SPAIN, Autoregressive, deep learning-based, graph-based, and other methods) are evaluated with three different metrics. The observations from the results show that the proposed solution is effective and can enhance the reconstruction of audio signals performed by the side information (e.g., Latent representation) steganography provides. Moreover, this paper proposes a novel framework for reconstruction from heavily compressed embedded audio data using halftoning (i.e., dithering) and machine learning, which we termed the HCR (halftone-based compression and reconstruction). This work may trigger interest in optimising this approach and/or transferring it to different domains (i.e., image reconstruction). Compared to existing methods, we show improvement in the inpainting performance in terms of signal-to-noise ratio (SNR), the objective difference grade (ODG) and Hansen’s audio quality metric. In particular, our proposed framework outperformed the learning-based methods (D2WGAN and SG) and the traditional statistical algorithms (e.g., SPAIN, TDC, WCP).

Place, publisher, year, edition, pages
Springer, 2024, 822. Vol. 822, p. 1-16
Series
Lecture Notes in Networks and Systems (LNNS), ISSN 23673370, E-ISSN 23673389
Keywords [en]
Audio reconstruction; Halftoning; Steganography; Machine learning
National Category
Signal Processing Computer Sciences
Research subject
Applied Signal Processing; Telecommunication Systems
Identifiers
URN: urn:nbn:se:bth-25930DOI: 10.1007/978-3-031-47721-8_1ISI: 001261691200001Scopus ID: 2-s2.0-85182506380ISBN: 9783031477201 (print)OAI: oai:DiVA.org:bth-25930DiVA, id: diva2:1832673
Conference
Intelligent Systems Conference, IntelliSys 2023, Amsterdam, Sept 7-8 2023
Available from: 2024-01-30 Created: 2024-01-30 Last updated: 2024-08-30Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopusarxiv.org

Authority records

Cheddad, Abbas

Search in DiVA

By author/editor
Cheddad, Abbas
By organisation
Department of Computer Science
Signal ProcessingComputer Sciences

Search outside of DiVA

GoogleGoogle Scholar

doi
isbn
urn-nbn

Altmetric score

doi
isbn
urn-nbn
Total: 170 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf