• No results found

Speech Emotion Recognition Using Convolutional Recurrent Neural Networks with Attention Model

N/A
N/A
Protected

Academic year: 2020

Share "Speech Emotion Recognition Using Convolutional Recurrent Neural Networks with Attention Model"

Copied!
10
0
0

Loading.... (view fulltext now)

Full text

(1)

2017 2nd International Conference on Computer Engineering, Information Science and Internet Technology (CII 2017) ISBN: 978-1-60595-504-9

Speech Emotion Recognition Using

Convolutional-Recurrent Neural Networks with Attention Model

YAWEI MU, LUIS A. HERNÁNDEZ GÓMEZ, ANTONIO CANO MONTES, CARLOS ALCARAZ MARTÍNEZ, XUETIAN WANG and HONGMIN GAO

ABSTRACT

Speech Emotion Recognition (SER) plays an important role in human-computer interface and assistant technologies. In this paper, a new method is proposed using distributed Convolution Neural Networks (CNN) to automatically learn affect-salient features from raw spectral information, and then applying Bidirectional Recurrent Neural Network (BRNN) to obtain the temporal information from the output of CNN. In the end, an Attention Mechanism is implemented on the output sequence of the BRNN to focus on target emotion-pertinent parts of an utterance. This attention mechanism not only improves the classification accuracy, but also provides model’s interpretability. Experimental results show that this approach can gain 64.08% weighted accuracy and 56.41% unweighted accuracy for four-emotion classification in IEMOCAP dataset, which outperform previous results reported for this dataset.

KEYWORDS

Speech emotion recognition, Distributed convolution neural networks, Bidirectional recurrent neural networks, Attention mechanism.

INTRODUCTION

With the development of Artificial Intelligence, more and more studies and research attempt to investigate Speech Emotion Recognition (SER), which not only makes the communication between machine and human more natural and real, but also has great potential in the development of auxiliary technologies, such as clinical assistant to help monitor the patient’s emotional change, especially these who have difficulty to express themselves very well. Speech emotion recognition is a sequence classification problem, where the input is the length-variant utterance and the output is the actual emotion. Considering the characteristics of speech and various ways people express emotions, there are many challenges to overcome. Besides, there is still no consensus about the best features to recognize emotion from speech among academic community or industrial circles. In recent years, there have been many researches involving in this area.

_________________________________________

Corresponding author: Yaw ei Mu, muyaw ei822@gmail.com School of Information and Electronics, Beijing Institute of Technology, Beijing 100081, P. R. China.

Luis A. Hernández Gómez, Antonio Cano Montes, Carlos Alcaraz Martínez, Grupo de Aplicaciones del Procesado de Señal, Universidad Politécnica de Madrid, Avda. Complutense 30, Madrid, Spain

(2)

In a general way, it can be stated that there are two main strategies to address the development of SER. The first one is to manually extract as many as relevant short-term features from audio signal, such as MFCC, pitch and energy etc. [1,2]. The second one is to let a neural network extract features automatically, for example, using autoencoders [3,4], or CNN models [5]. W.Q. Zheng et al. [6] used deep convolutional neural networks (DCNN) on five emotions from IEMOCAP [7], achieving 40.02% accuracy in speaker-independent manner.

After affect-salient features are available, several classification algorithms can be applied, such as SVM [8], Hidden Markov Method [9], Random Forest [10], DNN [11], or RNN [10]. Convolutional Recurrent Neural Network (CRNN), a combination of CNN followed by Long Short-Term Memory (LSTM) networks, has been investigated in [12], showing very good performance on Berlin [13] dataset, which is a much smaller database than IEMOCAP.

Based on that previous research, in this paper we propose a similar architecture (CNN + LSTM) extended with an attention mechanism. The proposed model is tested on IEMOCAP.

More specifically, in the proposed architecture a Convolutional Neural Network is firstly applied to the two-dimensional spectral representation of the audio signal to extract emotion-relevant features from short term segments. Then Bidirectional Recurrent Neural Networks (BRNN) are further implemented to model the temporal information closely associated to target emotion. Inspired by [14], an attention model is used to weight the output sequence of BRNN and get the final emotion classification. The experimental results show that the combination of CNN and BRNN (CRNN) as well as the use of an attention mechanism can improve results previously reported on IEMOCAP dataset.

The rest of paper is organized as follows. Details on the use of IEMOCAP are firstly described in section 2. The CRNN model proposed for Speech Emotion Recognition is introduced in section 3. Afterwards, experimental setup and results are presented in section 4. Conclusions and future research are given at the end of the paper.

EXPERIMENTAL DATASET

(3)

classifier [16], weighting the cost function to balance the classes [17] and so forth. In this paper, we chose cost-sensitive function for its simplicity.

PROPOSED SER MODEL

As Fig. 2 illustrates, to feed our SER model, audio signals are first represented by time-frequency maps, which are obtained by Fast Fourier transform of the speech wave form. We used 256 FFT bins for each short-term frame, which has the length of 20ms with 10ms overlapping. Subsequently, we stack a context window of 32 adjacent FFT frames together as one mid-term segment, which results in 32*128 two-dimensional time-frequency array. Thus, the total length of one segment is 10ms*32+10ms = 330ms. 50% overlapping between segments was used to make the analysis of neural networks smoother. In general, emotion is carried by continuous segments, the length of which is highly dependent on the speakers and certain emotion. However, a segment longer than 250ms has been verified to contain efficient emotion information [18, 19].

[image:3.612.151.443.412.632.2]

Once input audio signal is represented as spectrogram segments, they share the same property as a sequence of images, from which CNN can be applied to extract hierarchical affect-salient features (see Fig. 2 and Fig. 3). Afterwards, as shown in Fig. 2, bidirectional LSTM can be further used to explore the dynamic temporal information related to the final emotion representation hidden in the sequence of mid-term segments. Before the final classification, attention interface is implemented to focus on those more discriminative part of the RNN output sequence. Fig. 2 shows the detailed schema of our proposed approach.

(4)
[image:4.612.164.429.59.339.2]

Figure 2. The proposed CRNN with attention model.

[image:4.612.158.435.260.469.2]

Figure 3. The convolutional architecture for one mid-term segment.

[image:4.612.155.435.495.693.2]
(5)

Figure 5. The architecture of bidirectional LSTM with 78 memory cells.

CONVOLUTIONAL NEURAL NETWORKS

In CNN input is usually considered as an image, so some certain properties can be encoded, such as local connectivity and weight sharing, which substantially reduces the number of parameters to be trained in the network. In our case, CNN plays a very important role in extracting discriminative features from the time-frequency representation of audio signal, which can be viewed as a sequence of images. As it can be seen in Fig. 2, the feature representation provided by the CNN layer is used as input to the Bidirectional LSTM layer. LSTM is applied in a parallel and time-distributed manner.

Also, as shown in Fig. 3, for the analysis of each mid-term spectrogram segment, CNN has the basic architecture of [INPUT-CONV-RELU-POOL-CONV-RELU-POOL-FLAT].

The CNN input’s size is 32*128, which is one mid-term segment from the time-frequency map. With 50% overlapping between segments, the distribution of the number of segments in the whole dataset is shown in Fig. 4.

To preserve as much as possible the information provided by the IEMOCAP, we choose the padding length with 78 segments for every utterance sample. Therefore, the complete convolutional net will contain 78 parallel time-distributed architectures as showed in Fig. 2.

BIDIRECTIONAL RECURRENT NEURAL NETWORKS

We assume that in modeling emotions both past and future audio information can be relevant. Consequently, a bidirectional LSTM was implemented on the features learnt by the CNN layer. Following the same time structure of the convolutional net, the bidirectional LSTM was designed to have 78 memory cells. The detailed LSTM architecture is shown in Fig. 5.

(6)

ATTENTION MODEL

Attention models have represented significant progress in many applications, such as machine translation [20], image captioning [21] and description [22]. Attention mechanisms allow neural networks to focus on the most pertinent piece of information, and predict accordingly attributes to this success. To be trained with the neural network, attention algorithms should be differentiable. A direct approach when classifying sequences is to give different weights to every sequence output based on its expected relevance. Fig. 6 illustrates the process of the attention interface we used for emotion recognition.

Following the same approach presented in 13, our attention mechanism gives different weights to each element in the sequence of outputs from the bidirectional LSTM. The attention algorithm is based on the scores obtained through the dot-product of the LSTM outputs and an attention vector (𝑤𝑖). The attention vector is

trained together with the neural network through the backpropagation algorithm. The score is subsequently fed into a SoftMax to produce the attention distribution (i.e., weights) (𝑎𝑖) along all the output sequence. The distribution of these attention weights

suggests to what extent different segments in the input utterance contributes to the target emotion recognition.

PERFORMANCE EVALUATION

Data Preprocessing

Attention model should be able to take care of the silence part of utterances, but considering the complexity of IEMOCAP, removing silence part [23] was still considered necessary so that attention model can pay more attention to other more specific parts of the utterance. In this paper, we use WebRTC Voice Activity Detector (VAD) provided by Google, for which a very easy-implemented interface has already been well-developed in Python [24]. An angry utterance sample after silence removal is shown in Fig. 7. We can observe that all silence part can be almost perfectly removed without destroying any original information structure.

(7)
[image:7.612.177.412.56.202.2]

Figure 7. An angry utterance with VAD.

Table 1. Experimental Setup.

Network Parameters Value

Convolutional layer

Number of CONV layers 2

Convolutional filter size 3*3

Number of filters 16

Max Pooling filter size 2*2

Activation function Leaky RELU

Bidirectional LSTM layer Number of cells 128

Activation function tanh

Attention Layer Dimension of attention distribution 256

Optimization (SGD)

Learning rate 0.01

Decay 1e-6

momentum 0.8

Model training Number of epoch 50

Batch size 64

Table 2. Weights of cost function for four emotions.

Emotion angry happy neutral sad

Weight 0.95 1.92 0.68 1.03

Table 3. Accuracy comparison among different architectures.

WA UA

RNN without Attention model 60.27% 49.96%

RNN with Attention model 61.13% 49.66%

CRNN without Attention model 62.36% 50.66%

(8)

In addition, we applied z-normalization to all utterance’s spectrograms with respect to the mean and standard deviation of spectrograms of neutral emotion in the training dataset. It can eliminate the variation of the characteristics of speakers to some extent and improve the final accuracy.

NETWORK SETTINGS

The parameter set-up for the whole experiments are shown in the following table. To balance the number of classes, we weighted the cost function based on the number of utterance for each emotion in the training dataset. Table 2 shows the weights when Session 2 is used as the validation dataset. We can observe that happy, the emotion with the larger number of training examples, has the highest weight.

EXPERIMENTAL RESULTS

We first tried to evaluate the efficiency of using CNNs for extracting discriminative and hierarchical features. Therefore, we compared our CNN + RNN model with a simple one HSF + RNN in which the first CNN layer was replaced by simple high-level statistical functions (HSF). The mean and standard deviation of spectral features within 32 adjacent short-term frames were used as HSF.

Therefore, in the HSF + RNN model the input dimension to RNN was 78*256. In both cases the configuration for the bidirectional LSTM was the same.

To assess the efficiency of the attention mechanism, we also compared the model using the attention weights with another one using simple temporal mean pooling; that is using the average of the RNN output sequence.

The results in terms of WA (weighted accuracy), which means the overall accuracy in validation dataset, and UA (unweighted accuracy), which is the average accuracy among four emotions, are shown in Table 3. In all the cases IEMOCAP session 2 was the validation dataset, so they correspond to speaker independent tests.

SUMMARY

In this paper, a novel approach based on the combination of distributed Convolutional and Recurrent Neural Networks (CRNN) together with an attention mechanism has been proposed. Experimental results give 64.08% weighted accuracy and 56.41% unweighted accuracy for IEMOCAP dataset, which, to the best of our knowledge, outperforms any other previous result on this dataset. The experimental results also indicate that CNN can extract more discriminative features from the raw spectrogram than the general statistical properties. We have also verified that attention model can improve the performance of SER systems.

(9)

ACKNOWLEDGEMENTS

This research was financially supported by Estudios y Soluciones Avanzadas de Ingeniería (ESAI).

Corresponding author: Yawei Mu, School of Information and Electronics, Beijing Institute of Technology, Beijing 100081, P. R. China.

REFERENCES

1. Eyben, F., Wöllmer, M., & Schuller, B. (2009, September). OpenEAR—introducing the Munich open-source emotion and affect recognition toolkit. In Affective Computing and Intelligent Interaction and Workshops, 2009. ACII 2009. 3rd International Conference on (pp. 1-6). IEEE. 2. Luengo, Iker, Eva Navas, and Inmaculada Hernáez. "Feature analysis and evaluation for automatic

emotion identification in speech." IEEE Transactions on Multimedia 12.6 (2010): 490-501.

3. Cibau, Neri E., Enrique M. Albornoz, and Hugo L. Rufiner. "Speech emotion recognition using a deep autoencoder." Anales de la XV Reunion de Procesamiento de la Informacion y Control 16 (2013): 934-939.

4. Zhou, Xi, Junqi Guo, and Rongfang Bie. "Deep Learning Based Affective Model for Speech Emotion Recognition." Ubiquitous Intelligence & Computing, Advanced and Trusted Computing, Scalable Computing and Communications, Cloud and Big Data Computing, Internet of People, and Smart World Congress (UIC/ATC/ScalCom/CBDCom/IoP/SmartWorld), 2016 Intl IEEE Conferences. IEEE, 2016.

5. Huang, Zhengwei, et al. "Speech emotion recognition using CNN." Proceedings of the 22nd ACM international conference on Multimedia. ACM, 2014.

6. Zheng, W. Q., J. S. Yu, and Y. X. Zou. "An experimental study of speech emotion recognition based on deep convolutional neural networks." Affective Computing and Intelligent Interaction (ACII), 2015 International Conference on. IEEE, 2015.

7. Information on http://sail.usc.edu/iemocap/.

8. Dahake, Prajakta P., Kailash Shaw, and P. Malathi. "Speaker dependent speech emotion recognition using MFCC and Support Vector Machine." Automatic Control and Dynamic Optimization Techniques (ICACDOT), International Conference on. IEEE, 2016.

9. Kim, Yelin, and Emily Mower Provost. "Emotion classification via utterance-level dynamics: A pattern-based approach to characterizing affective expressions." Acoustics, Speech and Signal Processing (ICASSP), 2013 IEEE International Conference on. IEEE, 2013.

10. Chernykh, Vladimir, Grigoriy Sterling, and Pavel Prihodko. "Emotion Recognition from Speech with Recurrent Neural Networks." arXiv preprint arXiv: 1701.08071 (2017).

11. Han, Kun, Dong Yu, and Ivan Tashev. "Speech emotion recognition using deep neural network and extreme learning machine." Interspeech. 2014.

12. Lim, Wootaek, Daeyoung Jang, and Taejin Lee. "Speech emotion recognition using convolutional and Recurrent Neural Networks." Signal and Information Processing Association Annual Summit and Conference (APSIPA), 2016 Asia-Pacific. IEEE, 2016.

13. Burkhardt, Felix, et al. "A database of german emotional speech." Interspeech. Vol. 5. 2005. 14. Seyedmahdad, Emad, Cha. "Automatic speech recognition using recurrent neural network with

local attention." Acoustics, Speech and Signal processing (ICASSP), 2017 Intl IEEE Conferences. IEEE, 2017.

15. Jia, Cangzhi, and Yun Zuo. "S-SulfPred: A sensitive predictor to capture S-sulfenylation sites based on a resampling one-sided selection undersampling-synthetic minority oversampling technique." Journal of Theoretical Biology 422 (2017): 84-89.

16. Sun, Yanmin, et al. "Cost-sensitive boosting for classification of imbalanced data." Pattern Recognition 40.12 (2007): 3358-3378.

(10)

18. Kim, Yelin, and Emily Mower Provost. "Emotion classification via utterance-level dynamics: A pattern-based approach to characterizing affective expressions." Acoustics, Speech and Signal Processing (ICASSP), 2013 IEEE International Conference on. IEEE, 2013.

19. Provost, Emily Mower. "Identifying salient sub-utterance emotion dynamics using flexible units and estimates of affective flow." Acoustics, Speech and Signal Processing (ICASSP), 2013 IEEE International Conference on. IEEE, 2013.

20. Bahdanau, Dzmitry, Kyunghyun Cho, and Yoshua Bengio. "Neural machine translation by jointly learning to align and translate." arXiv preprint arXiv: 1409.0473 (2014).

21. Xu, Kelvin, et al. "Show, attend and tell: Neural image caption generation with visual attention." International Conference on Machine Learning. 2015.

22. Karpathy, Andrej, and Li Fei-Fei. "Deep visual-semantic alignments for generating image descriptions." Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2015.

23. Zhong-Qiu Wang and Ivan Tashev. "Learning utterance-level representations for speech emotion and age/gender recognition using deep neural networks." IEEE International Conference on Acoustics, Speech and Signal Processing

Figure

Figure 1. Utterance distribution about four emotions.
Figure 3. The convolutional architecture for one mid-term segment.
Figure 7. An angry utterance with VAD.

References

Related documents

Of the 291 unique comment letters pertaining to Medicaid Section 1115 waiver applications in five states, 186 (64 percent) were submitted by citizens, of which 55 identified

Sessions including audio (VoIP), multipoint video, and all web conferencing features – PowerPoint, document, browser, application, region and complete desktop sharing – can be

This study reveals that, given a half-hour increase in the minimum nursing hours per resident day for licensed nurses, quality of patient care increases by 25 percent.. This

All food contact surfaces, equipment, and utensils used for the preparation, packaging, or handling of any cottage food products shall be washed, rinsed, and

Figure 5 considers the case of a hedge fund with short iShare positions of up to 100%. The average optimal asset allocation changes dramatically once again. As

In May 2013, the Standing Council on Tertiary Education, Skills and Employment (SCOTESE) and the Standing Council on Schools Education and Early Childhood (SCSEEC) established a

The ODR process could then follow in accordance with the parties’ previous agreement (per- haps mediation, arbitration, or crowdsourced resolutions). Depending on the out-

The concepts of 'nation,' 'nationalism,' 'nationalization,' 'millet,' 'ulus,' 'national identity' and alike will be taken under theoretical investigation and their