Sentiment Analysis Using Image-based Deep Spectrum Features

We test the suitability of our novel deep spectrum feature representation for performing speech-based sentiment analysis. Deep spectrum features are formed by passing spectrograms through a pre-trained image convolutional neural network (CNN) and have been shown to capture useful emotion information in speech; however, their usefulness for sentiment analysis is yet to be investigated. Using a data set of movie reviews collected from YouTube, we compare deep spectrum features combined with the bag-of-audio-words (BoAW) paradigm with a state-of-the-art Mel Frequency Cepstral Coefficients (MFCC) based BoAW system when performing a binary sentiment classification task. Key results presented indicate the suitability of both features for the proposed task. The deep spectrum features achieve an unweighted average recall of 74.5 %. The results provide further evidence for theeffectiveness of deep spectrum features as a robust feature representation for speech analysis.
Title: Sentiment Analysis Using Image-based Deep Spectrum Features
Lecturer: Shahin Amiriparian
Date: 17-10-2017
Building/Room: Eichleitnerstraße 30 / 207
Contact: U Augsburg/TUM