Loading...
Please wait, while we are loading the content...
Similar Documents
Sound Event Detection Using Derivative Features in Deep Neural Networks
| Content Provider | MDPI |
|---|---|
| Author | Kwak, Jin-Yeol Chung, Yong-Joo |
| Copyright Year | 2020 |
| Description | We propose using derivative features for sound event detection based on deep neural networks. As input to the networks, we used log-mel-filterbank and its first and second derivative features for each frame of the audio signal. Two deep neural networks were used to evaluate the effectiveness of these derivative features. Specifically, a convolutional recurrent neural network (CRNN) was constructed by combining a convolutional neural network and a recurrent neural networks (RNN) followed by a feed-forward neural network (FNN) acting as a classification layer. In addition, a mean-teacher model based on an attention CRNN was used. Both models had an average pooling layer at the output so that weakly labeled and unlabeled audio data may be used during model training. Under the various training conditions, depending on the neural network architecture and training set, the use of derivative features resulted in a consistent performance improvement by using the derivative features. Experiments on audio data from the Detection and Classification of Acoustic Scenes and Events 2018 and 2019 challenges indicated that a maximum relative improvement of 16.9% was obtained in terms of the F-score. |
| Starting Page | 4911 |
| e-ISSN | 20763417 |
| DOI | 10.3390/app10144911 |
| Journal | Applied Sciences |
| Issue Number | 14 |
| Volume Number | 10 |
| Language | English |
| Publisher | MDPI |
| Publisher Date | 2020-07-17 |
| Access Restriction | Open |
| Subject Keyword | Applied Sciences Artificial Intelligence Sound Event Detection Convolutional Recurrent Neural Network Derivative Features Mean-teacher Model Attention Model |
| Content Type | Text |
| Resource Type | Article |