Please use this identifier to cite or link to this item:
https://idr.l2.nitk.ac.in/jspui/handle/123456789/15004
Full metadata record
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Chatterjee C.C. | |
dc.contributor.author | Mulimani M. | |
dc.contributor.author | Koolagudi S.G. | |
dc.date.accessioned | 2021-05-05T10:16:11Z | - |
dc.date.available | 2021-05-05T10:16:11Z | - |
dc.date.issued | 2020 | |
dc.identifier.citation | ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings , Vol. 2020-May , , p. 661 - 665 | en_US |
dc.identifier.uri | https://doi.org/10.1109/ICASSP40776.2020.9054628 | |
dc.identifier.uri | http://idr.nitk.ac.in/jspui/handle/123456789/15004 | - |
dc.description.abstract | In this paper we propose a Transposed Convolutional Recurrent Neural Network (TCRNN) architecture for polyphonic sound event recognition. Transposed convolution layer, which caries out a regular convolution operation but reverts the spatial transformation and it is combined with a bidirectional Recurrent Neural Network (RNN) to get TCRNN. Instead of the traditional mel spectrogram features, the proposed methodology incorporates mel-IFgram (Instantaneous Frequency spectrogram) features. The performance of the proposed approach is evaluated on sound events of publicly available TUT-SED 2016 and Joint sound scene and polyphonic sound event recognition datasets. Results show that the proposed approach outperforms state-of-the-art methods. © 2020 Institute of Electrical and Electronics Engineers Inc.. All rights reserved. | en_US |
dc.title | Polyphonic sound event detection using transposed convolutional recurrent neural network | en_US |
dc.type | Conference Paper | en_US |
Appears in Collections: | 2. Conference Papers |
Files in This Item:
There are no files associated with this item.
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.