skip to main content

Convolutional Neural Network based Audio Event Classification

Lim, Minkyu ; Lee, Donghyun ; Park, Hosung ; Kang, Yoseb ; Oh, Junseok ; Park, Jeong-Sik ; Jang, Gil-Jin ; Kim, Ji-Hwan

KSII Transactions on Internet and Information Systems, 2018, 12(6), , pp.2748-2760 [Periódico revisado por pares]

한국인터넷정보학회

Texto completo disponível

Citações Citado por
  • Título:
    Convolutional Neural Network based Audio Event Classification
  • Autor: Lim, Minkyu ; Lee, Donghyun ; Park, Hosung ; Kang, Yoseb ; Oh, Junseok ; Park, Jeong-Sik ; Jang, Gil-Jin ; Kim, Ji-Hwan
  • Assuntos: Analysis ; Audio event classification ; Convolutional neural networks ; Deep learning ; Machine learning ; Methods ; Neural networks ; Sound recordings ; 컴퓨터학
  • É parte de: KSII Transactions on Internet and Information Systems, 2018, 12(6), , pp.2748-2760
  • Notas: Korean Society for Internet Information
    KISTI1.1003/JNL.JAKO201821464986105
    http://doi.org/10.3837/tiis.2018.06.017
  • Descrição: This paper proposes an audio event classification method based on convolutional neural networks (CNNs). CNN has great advantages of distinguishing complex shapes of image. Proposed system uses the features of audio sound as an input image of CNN. Mel scale filter bank features are extracted from each frame, then the features are concatenated over 40 consecutive frames and as a result, the concatenated frames are regarded as an input image. The output layer of CNN generates probabilities of audio event (e.g. dogs bark, siren, forest). The event probabilities for all images in an audio segment are accumulated, then the audio event having the highest accumulated probability is determined to be the classification result. This proposed method classified thirty audio events with the accuracy of 81.5% for the UrbanSound8K, BBC Sound FX, DCASE2016, and FREESOUND dataset.
  • Editor: 한국인터넷정보학회
  • Idioma: Coreano;Inglês

Buscando em bases de dados remotas. Favor aguardar.