IDEAS home Printed from https://ideas.repec.org/a/gam/jsusta/v15y2023i14p10762-d1189852.html
   My bibliography  Save this article

ResNet Based on Multi-Feature Attention Mechanism for Sound Classification in Noisy Environments

Author

Listed:
  • Chao Yang

    (The 10th Research Institute of China Electronics Technology Group Corporation, Chengdu 610036, China
    School of Information and Electronic Engineering, Zhejiang University of Science and Technology, Hangzhou 310023, China
    These authors contributed equally to this work.)

  • Xingli Gan

    (School of Information and Electronic Engineering, Zhejiang University of Science and Technology, Hangzhou 310023, China
    These authors contributed equally to this work.)

  • Antao Peng

    (School of Information and Electronic Engineering, Zhejiang University of Science and Technology, Hangzhou 310023, China)

  • Xiaoyu Yuan

    (School of Information and Electronic Engineering, Zhejiang University of Science and Technology, Hangzhou 310023, China)

Abstract

Environmental noise affects people’s lives and poses challenges for urban sound classification. Traditional algorithms such as Mel frequency cepstral coefficients (MFCCs) struggle due to audio signal complexity. This study applied an attention mechanism to a deep residual network (ResNet) deep learning network to overcome the structural impact of urban noise on audio signals and improve classification accuracy. We propose a three-feature fusion ResNet + attention method (Net50_SE) to maximize information representation in environmental sound signals. This method uses residual structured convolutional neural networks (CNNs) for feature extraction in sound classification tasks. Additionally, an attention module is added to suppress environmental noise impact and focus on different feature map channels. The experimental results demonstrate the effectiveness of our method, achieving 93.2% accuracy compared with 82.87% with CNN and 84.77% with long short-term memory (LSTM). Our model provides higher accuracy and confidence in urban sound classification.

Suggested Citation

  • Chao Yang & Xingli Gan & Antao Peng & Xiaoyu Yuan, 2023. "ResNet Based on Multi-Feature Attention Mechanism for Sound Classification in Noisy Environments," Sustainability, MDPI, vol. 15(14), pages 1-15, July.
  • Handle: RePEc:gam:jsusta:v:15:y:2023:i:14:p:10762-:d:1189852
    as

    Download full text from publisher

    File URL: https://www.mdpi.com/2071-1050/15/14/10762/pdf
    Download Restriction: no

    File URL: https://www.mdpi.com/2071-1050/15/14/10762/
    Download Restriction: no
    ---><---

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:gam:jsusta:v:15:y:2023:i:14:p:10762-:d:1189852. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    We have no bibliographic references for this item. You can help adding them by using this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: MDPI Indexing Manager (email available below). General contact details of provider: https://www.mdpi.com .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.