TY - GEN
T1 - A CNN-GRU approach to capture time-frequency pattern interdependence for snore sound classification
AU - Wang, Jianhong
AU - Strömfelt, Harald
AU - Schuller, Björn W.
N1 - Publisher Copyright:
© EURASIP 2018.
PY - 2018/11/29
Y1 - 2018/11/29
N2 - In this work, we propose an architecture named DualConvGRU Network to overcome the INTERPEECH 2017 ComParE Snoring sub-challenge. In this network, we devise two new models: the Dual Convolutional Layer, which is applied to a spectrogram to extract features; and the Channel Slice Model, which reprocess the extracted features. The first amalgamates an ensemble of information collected from two types of convolutional operations, with differing kernel dimension on the frequency axis and equal dimension on the time axis. Secondly, the dependencies within the convolutional layer channel axes are learnt, by feeding channel slices into a Gated Recurrent Unit (GRU) layer. By taking this approach, convolutional layers can be connected to sequential models without the use of fully connected layers. Compared with other state-of-the-art methods delivered to INTERPEECH 2017 ComParE Snoring sub-challenge, our method ranks 5th on performance of test data. Moreover, we are the only competitor to train a deep learning model solely on the provided training data, except for Baseline. The performance of our model exceeds the baseline too much.
AB - In this work, we propose an architecture named DualConvGRU Network to overcome the INTERPEECH 2017 ComParE Snoring sub-challenge. In this network, we devise two new models: the Dual Convolutional Layer, which is applied to a spectrogram to extract features; and the Channel Slice Model, which reprocess the extracted features. The first amalgamates an ensemble of information collected from two types of convolutional operations, with differing kernel dimension on the frequency axis and equal dimension on the time axis. Secondly, the dependencies within the convolutional layer channel axes are learnt, by feeding channel slices into a Gated Recurrent Unit (GRU) layer. By taking this approach, convolutional layers can be connected to sequential models without the use of fully connected layers. Compared with other state-of-the-art methods delivered to INTERPEECH 2017 ComParE Snoring sub-challenge, our method ranks 5th on performance of test data. Moreover, we are the only competitor to train a deep learning model solely on the provided training data, except for Baseline. The performance of our model exceeds the baseline too much.
KW - Channel Slice Model
KW - Dual Convolutional Layers
KW - DualConvGRU Network
UR - http://www.scopus.com/inward/record.url?scp=85059818866&partnerID=8YFLogxK
U2 - 10.23919/EUSIPCO.2018.8553521
DO - 10.23919/EUSIPCO.2018.8553521
M3 - Conference contribution
AN - SCOPUS:85059818866
T3 - European Signal Processing Conference
SP - 997
EP - 1001
BT - 2018 26th European Signal Processing Conference, EUSIPCO 2018
PB - European Signal Processing Conference, EUSIPCO
T2 - 26th European Signal Processing Conference, EUSIPCO 2018
Y2 - 3 September 2018 through 7 September 2018
ER -