Suche

Publikationen


Book Chapters

  • A. Batliner and B. Möbius, "Prosody in Automatic Speech Processing," in The Oxford Handbook of Language Prosody (C. Gussenhoven and Aoju Chen, eds.), Oxford University Press. 20 pages, to appear 2020. [pdf]

Journal Papers

2019

  • S. Amiriparian, N. Cummins, M. Gerczuk, S. Pugachevskiy, S. Ottl, and B. Schuller, “'Are You Playing a Shooter Again?!' Deep Representation Learning for Audio-based Video Game Genre Recognition,” IEEE Transactions on Games, vol. 11, January 2019. 11 pages, to appear
  • J. Han, Z. Zhang, Z. Ren, and B. Schuller, “EmoBed: Strengthening Monomodal Emotion Recognition via Training with Crossmodal Emotion Embeddings,” IEEE Transactions on Affective Computing, vol. 10, 2019. 12 pages, to appear
  • C. Janott, M. Schmitt, C. Heiser, W. Hohenhorst, M. Herzog, M. C. Llatas, W. Hemmert, and B. Schuller, “VOTE versus ACLTE: Vergleich zweier Schnarchger¨auschklassifikationen mit Methoden des maschinellen Lernens, HNO, vol. 24, 2019. 9 pages, to appear
  • G. Keren, S. Sabato, and B. Schuller, “Analysis of Loss Functions for Fast Single-Class Classification,” Knowledge and Information Systems, vol. 59, 2019. 12 pages, invited as one of best papers from ICDM 2018, to appear
  • Z. Zhang, J. Han, K. Qian, C. Janott, Y. Guo, and B. Schuller, “Snore-GANs: Improving Automatic Snore Sound Classification with Synthesized Data,” IEEE Journal of Biomedical and Health Informatics, vol. 23, 2019. 11 pages, to appear
  • B. Schuller, “Responding to Uncertainty in Emotion Recognition,” Journal of Information, Communication & Ethics in Society, vol. 17, no. 2, 2019. 4 pages, invited contribution, to appear
  • S. Hantke, T. Olenyi, C. Hausner, and B. Schuller, “Large-scale Data Collection and Analysis via a Gamified Intelligent Crowdsourcing Platform,” International Journal of Automation and Computing, vol. 16, pp. 427-436, August 2019.
  • Z. Zhao, Z. Bao, Y. Zhao, Z. Zhang, N. Cummins, Z. Ren, and B. Schuller, “Exploring Deep Spectrum Representations via Attention-based Recurrent and Convolutional Neural Networks for Speech Emotion Recognition,” IEEE Access, pp. 9715-97525, July 2019. [open access]
  • Z. Zhang, J. Han, E. Coutinho, and B. Schuller, “Dynamic Difficulty Awareness Training for Continuous Emotion Prediction,” IEEE Transactions on Multimedia, vol. 20, pp. 1289-1301, May 2019.
  • J. Han, Z. Zhang, N. Cummins, and B. Schuller, “Adversarial Training in Affective Computing and Sentiment Analysis: Recent Advances and Perspectives,” IEEE Computational Intelligence Magazine, special issue on Computational Intelligence for Affective Computing and Sentiment Analysis, pp. 68-81, April 2019. [preprint]
  • K. Qian, M. Schmitt, C. Janott, Z. Zhang, C. Heiser, W. Hohenhorst, M. Herzog, W. Hemmert, and B. Schuller, “A Bag of Wavelet Features for Snore Sound Classification,” Annals of Biomedical Engineering, vol. 47, pp. 1000-1011, April 2019.
  • E. Parada-Cabaleiro, G. Costantini, A. Batliner, M. Schmitt, and B. W. Schuller, “DEMoS – An Italian Emotional Speech Corpus – Elicitation methods, machine learning, and perception,” Language Resources and Evaluation, vol. 53, pp. 1-43, February 2019.
  • B. Schuller, “Micro-Expressions – A Chance for Computers to Beat Humans at Revealing Hidden Emotions?,” IEEE Computer Magazine, vol. 52, pp. 4-5, February 2019.
  • B. Schuller, F. Weninger, Y. Zhang, F. Ringeval, A. Batliner, S. Steidl, F. Eyben, E. Marchi, A. Vinciarelli, K. Scherer, M. Chetouani, and M. Mortillaro, “Affective and Behavioural Computing: Lessons Learnt from the First Computational Paralinguistics Challenge,” Computer Speech and Language, vol. 53, pp. 156-180, January 2019.

2018

  • K. Veselkov and B. Schuller, “The age of data analytics: converting biomedical data into actionable insights,” Methods, Special Issue on Health Informatics and Translational Data Analytics, vol. 151, pp. 1-2, December 2018.
  • N. Cummins, B. W. Schuller, and A. Baird, “Speech analysis for health: Current state-of-the-art and the increasing impact of deep learning,” Methods, Special Issue on Health Informatics and Translational Data Analytics, vol. 151, pp. 41-54, December 2018.
  • K. Grabowski, A. Rynkiewicz, A. Lassalle, S. Baron-Cohen, B. Schuller, N. Cummins, A. E. Baird, J. Podgórska-Bednarz, A. Pieniazek, and I. Lucka, “Emotional expression in psychiatric conditions – new technology for clinicians,” Psychiatry and Clinical Neurosciences, vol. 73, pp. 50-62, November 2018.
  • S. Jing, X. Mao, L. Chen, M. C. Comes, A. Mencattini, G. Raguso, F. Ringeval, B. Schuller, C. D. Natale, and E. Martinelli, “A closed-form solution to the graph total variation problem for continuous emotion profiling in noisy environment,” Speech Communication, vol. 104, pp. 66–72, November 2018.
  • K. Qian, C. Janott, Z. Zhang, J. Deng, A. Baird, C. Heiser, W. Hohenhorst, M. Herzog, W. Hemmer, and B. Schuller, “Teaching Machines on Snoring: A Benchmark on Computer Audition for Snore Sound Excitation Localisation,” Archives of Acoustics, vol. 43, pp. 465-475, November 2018. [open access]
  • F. B. Pokorny, K. D. Bartl-Pokornya, C. Einspieler, D. Zhang, R. Vollmann, S. Bölte, H. Tager-Flusberg, M. Gugatschka, B.W. Schuller, and P. B. Marschik, “Typical vs. atypical: Combining auditory Gestalt perception and acoustic analysis of early vocalisations in Rett syndrome,” Research in Developmental Disabilities, vol. 82, pp. 109-119, November 2018.
  • J. Han, Z. Zhang, G. Keren, and B. Schuller, “Emotion Recognition in Speech with Latent Discriminative Representations Learning,” Acta Acustica united with Acustica, vol. 104, pp. 737-740, September/October 2018. [open access]
  • D. Schuller and B. Schuller, “The Age of Artificial Emotional Intelligence,” IEEE Computer Magazine, Special Issue on The Future of Artificial Intelligence, vol. 51, September 2018, pp. 38-46.
  • A. Mencattini, F. Mosciano, M. Colomba Comes, T. De Gregorio, G. Raguso, E. Daprati, F. Ringeval, B. Schuller, and E. Martinelli, “An emotional modulation model as signature for the identification of children developmental disorders,” Scientific Reports, 14487, 12 pages, September 2018. [open access]
  • G. Keren, N. Cummins, and B. Schuller, “Calibrated Prediction Intervals for Neural Network Regressors,” IEEE Access, vol. 6, pp. 54033-54041, September 2018. [open access]
  • E. Marchi, B. Schuller, A. Baird, S. Baron-Cohen, A. Lassalle, H. O’Reilly, D. Pigat, P. Robinson, I. Davies, T. Baltrusaitis, O. Golan, S. Fridenson-Hayo, S. Tal, S. Newman, N. MeirGoren, A. Camurri, S. Piana, S. Bölte, M. Sezgin, N. Alyuz, A. Rynkiewicz, and A. Baranger, “The ASC-Inclusion Perceptual Serious Gaming Platform for Autistic Children,” IEEE Transactions on Computational Intelligence and AI in Games, pp. 1-12, August 2018.
  • S. Hantke, A. Abstreiter, N. Cummins, and B. Schuller, “Trustability-based Dynamic Active Learning for Crowdsourced Labelling of Emotional Audio Data,” IEEE Access, vol. 6, 2018. pp. 42142-42155, July 2018. [open access]
  • O. Rudovic, J. Lee, M. Dai, B. Schuller, and R. W. Picard, “Personalized machine learning for robot perception of affect and engagement in autism therapy,” Science Robotics, vol. 3, doi: 10.1126/scirobotics.aao6760, 11 pages, June 2018.  [open access]
  • B. Schuller, “What Affective Computing Reveals on Autistic Children’s Face of Joy or Fear,” IEEE Computer Magazine, vol. 51,  pp. 7-8, June 2018. [open access]
  • B. Schuller, “Speech Emotion Recognition: Two Decades in a Nutshell, Benchmarks, and Ongoing Trends,” Communications of the ACM, vol. 61, pp. 90-99, May 2018. 
  • A. Baird, S. H. Jorgensen, E. Parada-Cabaleiro, S. Hantke, N. Cummins, and B. Schuller, “The Perception of Vocal Traits in Synthesized Voices: Age, Gender, and Human Likeness,” Journal of the Audio Engineering Society, Special Issue on Augmented and Participatory Sound and Music Interaction using Semantic Audio, vol. 66, pp. 277-285, April 2018.
  • Z. Zhang, J. Han, J. Deng, X. Xu, F. Ringeval, and B. Schuller, “Leveraging Unlabelled Data for Emotion Recognition with Enhanced Collaborative Semi-Supervised Learning,” IEEE Access, vol. 6, pp. 22196 - 22209, April 2018. [open access]
  • Z. Ren, K. Qian, Z. Zhang, V. Pandit, A. Baird, and B. Schuller, “Deep Scalogram Representations for Acoustic Scene Classification,” IEEE/CAA Journal of Automatica Sinica, 2018. vol. 5, pp. 662-669, April 2018.  [open access]
  • C. Janott, M. Schmitt, Y. Zhang, K. Qian, V. Pandit, Z. Zhang, C. Heiser, W. Hohenhorst, M. Herzog, W. Hemmert, and B. Schuller, “Snoring Classified: The Munich Passau Snore Sound Corpus,” Computers in Biology and Medicine, vol. 94, pp. 106-118, March 2018. 
  • M. Freitag, S. Amiriparian, S. Pugachevskiy, N. Cummins, and B. Schuller, “auDeep: Unsupervised Learning of Representations from Audio with Deep Recurrent Neural Networks,” Journal of Machine Learning Research, vol. 18, pp. 1-5, February 2017 - December 2018. [open access]
  • J. Deng, X. Xu, Z. Zhang, S. Fruhholz, and B. Schuller, “Semi-Supervised Autoencoders for Speech Emotion Recognition,” IEEE/ACM Transactions on Audio, Speech and Language Processing, vol. 26, pp. 31-43, January 2018.

2017

  • X. Xu, J. Deng, E. Coutinho, C. Wu, L. Zhao, and B. Schuller, “Connecting Subspace Learning and Extreme Learning Machine in Speech Emotion Recognition,” IEEE Transactions on Multimedia, vol. 20, 2017, 13 pages.
  • P. Tzirakis, G. Trigeorgis, M. A. Nicolaou, B. Schuller, and S. Zafeiriou, “End-to-End Multimodal Emotion Recognition using Deep Neural Networks,” IEEE Journal of Selected Topics in Signal Processing, Special Issue on End-to-End Speech and Language Processing, vol. 11, pp. 1301-1309, December 2017.
  • V. Pandit and B. Schuller, “A Novel Graphical Technique for Combinational Logic Representation and Optimization,” Complexity, vol. 2017, Article ID 9696342, 12 pages, 2017. [open access]
  • D. Schuller and B. Schuller, “Automatic Speech Emotion Recognition in Dialogue and Human-Computer Interaction: A Survey,” International Journal of Automation and Computing, vol. 15, 2018. 10 pages, invited contribution, to appear
  • B. Schuller and D. Schuller, “Maschinelle Profilierung – Die KI kennt Sie,” digma – Zeitschrift für Datenrecht und Informationssicherheit, vol. 1, no. 4, 2017. 6 pages, to appear

Conference Proceedings

2019

  • E. Parada-Cabaleiro, A. Batliner, and B. Schuller, “A Diplomatic Edition of Il Lauro Secco: Ground Truth for OMR of White Mensural Notation,” in Proceedings of the 20th International Society for Music Information Retrieval Conference (ISMIR), Delft, The Netherlands, November 2019. 7 pages, to appear
  • F. Ringeval, B. Schuller, M. Valstar, N. Cummins, R. Cowie, M. Soleymani, M. Schmitt, S. Amiriparian, E.-M. Messner, L. Tavabi, S. Song, S. Alisamir, S. Lui, Z. Zhao, and M. Pantic, “AVEC 2019 Workshop and Challenge: State-of-Mind, Depression with AI, and Cross-Cultural Affect Recognition,” in Proceedings of the 9th International Workshop on Audio/Visual Emotion Challenge (AVEC), Nice, France, October 2019. 8 pages, to appear
  • B. W. Schuller, A. Batliner, C. Bergler, F. Pokorny, J. Krajewski, M. Cychosz, R. Vollmann, S.-D. Roelen, S. Schnieder, E. Bergelson, A. Cristià, A. Seidl, L. Yankowitz, E. Nöth, S. Amiriparian, S. Hantke, and M. Schmitt, “The INTERSPEECH 2019 Computational Paralinguistics Challenge: Styrian Dialects, Continuous Sleepiness, Baby Sounds & Orca Activity,” in Proceedings of INTERSPEECH, Graz, Austria, September 2019. 5 pages, to appear
  • A. Baird, E. Coutinho, J. Hirschberg, and B. W. Schuller, “Sincerity in Acted Speech: Presenting the Sincere Apology Corpus and Results,” in Proceedings of INTERSPEECH, Graz, Austria, September 2019. 5 pages, to appear
  • A. Baird, S. Amiriparian, N. Cummins, S. Strumbauer, J. Janson, E.-M. Messner, H. Baumeister, N. Rohleder, and B. W. Schuller, “Using Speech to Predict Sequentially Measured Cortisol Levels During a Trier Social Stress Test,” in Proceedings of INTERSPEECH, Graz, Austria, September 2019. 5 pages, to appear 
  • Y. Guo, Z. Zhao, Y. Ma, and B. W. Schuller, “Speech Augmentation via Speaker-Specific Noise in Unseen Environment,” in Proceedings of INTERSPEECH, Graz, Austria, September 2019. 5 pages, to appear
  • A. Mallol-Ragolta, Z. Zhao, L. Stappen, N. Cummins, and B. W. Schuller, “A Hierarchical Attention Network-Based Approach for Depression Detection from Transcribed Clinical Interviews,” in Proceedings of INTERSPEECH, Graz, Austria, September 2019. 5 pages, to appear
  • M. Schmitt, N. Cummins, and B. W. Schuller, “Continuous Emotion Recognition in Speech – Do We Need Recurrence?,” in Proceedings of INTERSPEECH, Graz, Austria, September 2019. 5 pages, to appear 
  • X. Xu, J. Deng, N. Cummins, Z. Zhang, L. Zhao, and B. W. Schuller, “Autonomous emotion learning in speech: A view of zero-shot speech emotion recognition,” in Proceedings of INTERSPEECH 2019, Graz, Austria, September 2019. 5 pages, to appear
  • Z. Zhao, Z. Bao, Z. Zhang, N. Cummins, H. Wang, and B. W. Schuller, “Attention-enhanced Connectionist Temporal Classification for Discrete Speech Emotion Recognition,” in Proceedings of INTERSPEECH, Graz, Austria, September 2019. 5 pages, to appear
  • A. Baird, S. Amiriparian, and B. Schuller, “Can Deep Generative Audio be Emotional? Towards an Approach for Personalised Emotional Audio Generation,” in Proceedings of the the IEEE 21st International Workshop on Multimedia Signal Processing (MMSP 2019), Kuala Lumpur, Malaysia, September 2019. 5 pages, to appear
  • A. Baird, S. Amiriparian, M. Berschneider, M. Schmitt, and B. Schuller, “Predicting Blood Volume Pulse and Skin Conductance from Speech: Introducing a Novel Database and Results,” in Proceedings of the IEEE 21st International Workshop on Multimedia Signal Processing (MMSP 2019), Kuala Lumpur, Malaysia, September 2019. 5 pages, to appear
  • L. Stappen, V. Karas, N. Cummins, F. Ringeval, K. Scherer, and B. Schuller, “From Speech to Facial Activity: Towards Cross-modal Sequence-to-Sequence Attention Networks,” in Proceedings of the IEEE 21st International Workshop on Multimedia Signal Processing (MMSP 2019), Kuala Lumpur, Malaysia, September 2019. 6 pages, to appear
  • M. Schmitt and B. W. Schuller, “End-to-end Audio Classification with Small Datasets – Making It Work,” in Proceedings of the 27th European Signal Processing Conference (EUSIPCO), A Coruña, Spain, September 2019. 5 pages, to appear
  • M. Song, Z. Yang, A. Baird, E. Parada-Cabaleiro, Z. Zhang, Z. Zhao, and B. Schuller, “Audiovisual Analysis for Recognising Frustration during Game-Play: Introducing the Multimodal Game Frustration Database,” in Proceedings of the 8th Biannual Conference on Affective Computing and Intelligent Interaction (ACII 2019), Cambridge, UK, September 2019. to appear
  • S. Amiriparian, A. Awad, M. Gerczuk, L. Stappen, A. Baird, S. Ottl, and B. Schuller, “Audio-based Recognition of Bipolar Disorder Utilising Capsule Networks,” in Proceedings of the 32nd International Joint Conference on Neural Networks (IJCNN), Budapest, Hungary, July 2019. 7 pages, to appear
  • C. Li, Q. Zhang, Z. Zhao, L. Gu, N. Cummins, and B. Schuller, “Analysing and Inferring of Intimacy Based on fNIRS Signals and Peripheral Physiological Signals,” in Proceedings of the 32nd International Joint Conference on Neural Networks (IJCNN), Budapest, Hungary, July 2019. 8 pages, to appear
  • C. Janott, C. Rohrmeier, M. Schmitt, W. Hemmert, and B. Schuller, “Snoring – An Acoustic Definition,” in Proceedings of the 41st Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC 2019), Berlin, Germany, July 2019. 5 pages, to appear
  • J. Schiele, F. Rabe, M. Schmitt, M. Glaser, F. H¨aring, J. O. Brunner, B. Bauer, B. Schuller, C. Traidl-Hoffmann, and A. Damialis, “Automated Classification of Airborne Pollen using Neural Networks,” in Proceedings of the 41st Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Berlin, Germany, July 2019. 5 pages, to appear
  • V. Pandit, M. Schmitt, N. Cummins, and B. Schuller, “I know how you feel now, and here’s why!: Demystifying Timecontinuous High Resolution Text-based Affect Predictions In the Wild,” in Proceedings of the 32nd IEEE International Symposium on Computer-Based Medical Systems (CBMS), Cordoba, Spain, June 2019, pp. 465-470. [link]
  • O. Rudovic, B. Schuller, C. Breazeal, and R. Picard, “Personalized Estimation of Engagement from Videos Using Active Learning with Deep Reinforcement Learning,” in Proceedings of the 9th IEEE International Workshop on Analysis and Modeling of Faces and Gestures (AMFG), Long Beach, CA, June 2019. 10 pages, to appear
  • J. Schmid, M. Schneider, A. Höß, and B. Schuller, “A Comparison of AI-Based Throughput Prediction for Cellular Vehicle-To-Server Communication,” in Proceedings of the 15th International Wireless Communications and Mobile Computing Conference (IWCMC), Tangier, Morocco, June 2019, pp. 471-476. [link]
  • A. Mallol-Ragolta, M. Schmitt, A. Baird, N. Cummins, and B. Schuller, “Performance Analysis of Unimodal and Multimodal Models in Valence-Based Empathy Recognition,” in Workshop Proceedings  of the 14th IEEE International Conference on Automatic Face & Gesture Recognition (FG), Lille, France, May 2019, 5 pages. [link]
  • J. Han, Z. Zhang, Z. Ren, and B. Schuller, “Implicit Fusion by Joint Audiovisual Training for Emotion Recognition in Mono Modality,” in Proceedings of the 44th IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), Brighton, UK, IEEE, IEEE, May 2019, pp. 5861 - 5865. [link] [slides]
  • Z. Ren, Q. Kong, J. Han, M. D. Plumbley, and B. W. Schuller, “Attention-based Atrous Convolutional Neural Networks: Visualisation and Understanding Perspectives of Acoustic Scenes,” in Proceedings of the 44th IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), Brighton, UK, IEEE, IEEE, May 2019, pp. 56 - 60. [link] [slides]
  • L. Stappen, N. Cummins, E.-M. Rathner, H. Baumeister, J. Dineley, and B. Schuller, “Context Modelling Using Hierarchical Attention Networks for Sentiment and Self-Assessed Emotion Detection in Spoken Narratives,” in Proceedings of the 44th IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), Brighton, UK, IEEE, IEEE, May 2019, pp. 6680 - 6684. [link

2018

  • G. Keren, S. Sabato, and B. Schuller, “Fast Single-Class Classification and the Principle of Logit Separation,” in Proceedings of the International Conference on Data Mining (ICDM), Singapore, Singapore, November 2018, pp. 227-236. [link]
  • Z. Ren, Q. Kong, K. Qian, and B. Schuller, “Attention-based Convolutional Neural Networks for Acoustic Scene Classification,” in Proceedings of the 3rd Detection and Classification of Acoustic Scenes and Events 2018 Workshop (DCASE), Surrey, UK, November 2018, pp. 39-43. [link] [slides]
  • F. Ringeval, B. Schuller, M. Valstar, R. Cowie, H. Kaya, M. Schmitt, S. Amiriparian, N. Cummins, D. Lalanne, A. Michaud, E. Ciftci, H. Gülec, A. A. Salah, and M. Pantic, “AVEC 2018 Workshop and Challenge: Bipolar Disorder and Cross-Cultural Affect Recognition,” in Proceedings of the 8th International Workshop on Audio/Visual Emotion Challenge (AVEC), co-located with the 26th ACM International Conference on Multimedia (MM), Seoul, South Korea, October 2018, pp. 3-13.
  • S. Hantke, M. Schmitt, P. Tzirakis, and B. Schuller, “EAT - The ICMI 2018 Eating Analysis and Tracking Challenge,” in Proceedings of the 20th ACM International Conference on Multimodal Interaction (ICMI), Boulder, CO, October 2018, pp. 559-563.
  • B. Sertolli, N. Cummins, A. Sengur, and B. Schuller, “Deep End-to-End Representation Learning for Food Type Recognition from Speech,” in Proceedings of the 20th ACM International Conference on Multimodal Interaction (ICMI), Boulder, CO, October 2018, pp. 574-578.
  • Y. Guo, J. Han, Z. Zhang, B. Schuller, and Y. Ma, “Exploring a New Method for Food Likability Rating Based on DT-CWT Theory,” in Proceedings of the 20th ACM International Conference on Multimodal Interaction (ICMI), Boulder, CO, October 2018, pp. 569-573.
  • Z. Ren, N. Cummins, J. Han, S. Schnieder, J. Krajewski, and B. Schuller, “Evaluation of the Pain Level from Speech: Introducing a Novel Pain Database and Benchmarks,” in Proceedings of the 13th ITG Conference on Speech Communication, Oldenburg, Germany, October 2018, pp. 56-60.
  • A. Sengur, F. Demir, H. Lu, S. Amiriparian, N. Cummins, and B. Schuller, “Compact Bilinear Deep Features for Environmental Sound Recognition,” in Proceedings of the International Conference on Artificial Intelligence and Data Mining (IDAP), Malatya, Turkey, September 2018. 5 pages. [pdf]
  • A. Baird, E. Parada-Cabaleiro, C. Fraser, S. Hantke, and B. Schuller, “The Emotion of Synthetic Audio – A Dataset and Perception Results,” in Proceedings of the 12th Audio Mostly Conference on Interaction with Sound (Audio Mostly), Wrexham, UK, September 2018. 8 pages, to appear
  • V. Pandit, M. Schmitt, N. Cummins, F. Graf, L. Paletta, and B. Schuller, “How Good Is Your Model ‘Really’? On ‘Wildness’ of the In-the-wild Speech-based Affect Recognisers,” in Proceedings 20th International Conference on Speech and Computer (SPECOM), Leipzig, Germany, September 2018, pp. 490-500. 
  • J. Han, M. Schmitt, and B. Schuller, “You Sound Like Your Counterpart: Interpersonal Speech Analysis,” in Proceedings 20th International Conference on Speech and Computer (SPECOM), Leipzig, Germany, September 2018, pp. 188-197.
  • G. Keren, J. Han, and B. Schuller, “Scaling Speech Enhancement in Unseen Environments with Noise Embeddings,” in Proceedings The 5th International Workshop on Speech Processing in Everyday Environments held in conjunction with Interspeech 2018, Hyderabad, India, September 2018, pp. 25-29. [link]
  • B. Schuller, S. Steidl, A. Batliner, P. B. Marschik, H. Baumeister, F. Dong, S. Hantke, F. Pokorny, E.-M. Rathner, K. D. Bartl-Pokorny, C. Einspieler, D. Zhang, A. Baird, S. Amiriparian, K. Qian, Z. Ren, M. Schmitt, P. Tzirakis, and S. Zafeiriou, “The INTERSPEECH 2018 Computational Paralinguistics Challenge: Atypical & Self-Assessed Affect, Crying & Heart Beats,” in Proceedings of INTERSPEECH, Hyderabad, India, September 2018, pp. 122-126. [link]
  • S. Hantke, C. Stemp, and B. Schuller, “Annotator Trustability-based Cooperative Learning Solutions for Intelligent Audio Analysis,” in Proceedings of INTERSPEECH, Hyderabad, India, September 2018, pp. 3504-3508. [link]
  • A. Baird, E. Parada-Cabaleiro, S. Hantke, F. Burkhardt, N. Cummins, and B. Schuller, “The Perception and Analysis of the Likeability and Human Likeness of Synthesized Speech,” in Proceedings INTERSPEECH, Hyderabad, India, September 2018, pp. 2863-2867. [link]
  • E. Parada-Cabaleiro, G. Costantini, A. Batliner, A. Baird, and B. Schuller, “Categorical vs Dimensional Perception of Italian Emotional Speech,” in Proceedings of INTERSPEECH, Hyderabad, India, September 2018, pp. 3638-3642. [link]
  • E.-M. Rathner, J. Djamali, Y. Terhorst, B. Schuller, N. Cummins, G. Salamon, C. Hunger-Schoppe, and H. Baumeister, “How did you like 2017? Detection of language markers of depression and narcissism in personal narratives,” in Proceedings of INTERSPEECH, Hyderabad, India, September 2018, pp. 3388-3392. [link]
  • E.-M. Rathner, Y. Terhorst, N. Cummins, B. Schuller, and H. Baumeister, “State of mind: Classification through self-reported affect and word use in speech,” in Proceedings INTERSPEECH, Hyderabad, India, September 2018, pp. 267-271. [link]
  • S. Amiriparian, A. Baird, S. Julka, A. Alcorn, S. Ottl, S. Petrović, E. Ainger, N. Cummins, and B. Schuller, “Recognition of Echolalic Autistic Child Vocalisations Utilising Convolutional Recurrent Neural Networks,” in Proceedings of INTERSPEECH, Hyderabad, India, September 2018, pp. 2334-2338. [link]
  • Z. Zhang, J. Han, K. Qian, and B. Schuller, “Evolving Learning for Analysing Mood-Related Infant Vocalisation,” in Proceedings INTERSPEECH, Hyderabad, India, September 2018, pp. 142-146. [link]
  • Z. Zhang, A. Cristia, A. Warlaumont, and B. Schuller, "Automated Classification of Children's Linguistic versus Non-Linguistic Vocalisations," in Proceedings of INTERSPEECH, Hyderabad, India, September 2018, pp. 2588-2592. [link]
  • J. Han, Z. Zhang, M. Schmitt, Z. Ren, F. Ringeval, and B. Schuller, “Bags in Bag: Generating Context-Aware Bags for Tracking Emotions from Speech,” in Proceedings of INTERSPEECH, Hyderabad, India, September 2018, pp. 3082-3086. [link]
  • W. Han, H. Ruan, X. Chen, Z. Wang, H. Li, and B. Schuller, “Towards Temporal Modelling of Categorical Speech Emotion Recognition,” in Proceedings of INTERSPEECH, Hyderabad, India, September 2018, pp. 932–936. [link]
  • E. Parada-Cabaleiro, A. Batliner, M. Schmitt, and B. Schuller, “Musical-Linguistic Annotations of Il Lauro Secco,” in Proceedings of the 19th International Society for Music Information Retrieval Conference (ISMIR), Paris, France, September 2018. pp. 461-467. [pdf]
  • E. Parada-Cabaleiro, A. Batliner, M. Schmitt, S. Hantke, K. Scherer, and B. Schuller, “Identifying Emotions in Opera Singing: Implications of Adverse Acoustic Conditions,” in Proceedings of the 19th International Society for Music Information Retrieval Conference (ISMIR), Paris, France, September 2018. pp. 376-382. [pdf]
  • S. Amiriparian, M. Freitag, N. Cummins, M. Gerzcuk, S. Pugachevskiy, and B. W. Schuller, “A Fusion of Deep Convolutional Generative Adversarial Networks and Sequence to Sequence Autoencoders for Acoustic Scene Classification,” in Proceedings of the 26th European Signal Processing Conference (EUSIPCO), Rome, Italy, September 2018. 5 pages, pp.977-981. [link]
  • S. Amiriparian, M. Gerczuk, S. Ottl, N. Cummins, S. Pugachevskiy, and B. Schuller, “Bag-of-Deep-Features: Noise-Robust Deep Feature Representations for Audio Analysis,” in Proceedings of the 31st International Joint Conference on Neural Networks (IJCNN), Rio de Janeiro, Brazi, July 2018, pp. 2419-2425. [link] [ppt]
  • S. Amiriparian, M. Schmitt, N. Cummins, K. Qian, F. Dong, and B. Schuller, “Deep Unsupervised Representation Learning for Abnormal Heart Sound Classification,” in Proceedings of the 40th Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Honolulu, HI, July 2018, pp. 4776-4779. [link] [ppt]
  • F. Demir, A. Sengur, N. Cummins, S. Amiriparian, and B. Schuller, “Low-Level Texture Features for Snore Sound Discrimination,” in Proceedings of the 40th Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Honolulu, HI, July 2018, pp. 413-416. [link] [poster]
  • S. Hantke, C. Cohrs, M. Schmitt, B. Tannert, F. Lütkebohmert, M. Detmers, H. Schelhowe, and B. Schuller, “EmotAsS: An Emotion-driven Assistance System for Cognitively Impaired Individuals,” in Proceedings of the 16th International Conference on Computers Helping People with Special Needs (ICCHP), Linz, Austria, July 2018. pp. 486-494. [pdf]
  • S. Song, S. Zhang, B. Schuller, L. Shen, and M. Valstar, “Noise Invariant Frame Selection: A Simple Method to Address the Background Noise Problem for Text-Independent Speaker Verification,” in Proceedings of the 31st International Joint Conference on Neural Networks (IJCNN), Rio de Janeiro, Brazil, July 2018, pp. 1-8.
  • J. Schmid, P. Heß, A. Höß, and B. Schuller, “Passive monitoring and geo-based prediction of mobile network vehicle-to server communication,” in Proceedings of the 14th International Wireless Communications and Mobile Computing Conference (IWCMC), Limassol, Cyprus, June 2018, pp. 483-1488.
  • A. Baird, S. Hantke, and B. Schuller, “Responsible Speech and Language Analysis: On Auditability, Benchmarking, Confidence, Data-Reliance & Explainability,” in Proceedings of the Legal and Ethical Issues Workshop, satellite of the 11th Language Resources and Evaluation Conference (LREC), Miyazaki, Japan, May 2018, 4 pages. [link]
  • S. Hantke, T. Appel, and B. Schuller, “The Inclusion of Gamification Solutions to Enhance User Enjoyment on Crowdsourcing Platforms,” in Proceedings of the first Asian Conference on Affective Computing and Intelligent Interaction (ACII Asia), Beijing, China, May 2018, 6 pages. [link]
  • S. Hantke, T. Olenyi, C. Hausner, and B. Schuller, “VoiLA: An Online Intelligent Speech Analysis and Collection Platform,” in Proceedings of the first Asian Conference on Affective Computing and Intelligent Interaction (ACII Asia), Beijing, China, May 2018, 5 pages. [link]
  • Y. Li, J. Tao, B. Schuller, S. Shan, D. Jiang, and J. Jia, “MEC 2017: Multimodal Emotion Recognition Challenge 2017,” in Proceedings of the first Asian Conference on Affective Computing and Intelligent Interaction (ACII Asia), Beijing, China, May 2018. 5 pages. [link]
  • V. Pandit, N. Cummins, M. Schmitt, S. Hantke, F. Graf, L. Paletta, and B. Schuller, “Tracking Authentic and In-the-wild Emotions using Speech,” in Proceedings of the first Asian Conference on Affective Computing and Intelligent Interaction (ACII Asia), Beijing, China, May 2018. 6 pages. [link]
  • H.-J. Vögel, C. Süß, V. Ghaderi, R. Chadowitz, E. André, N. Cummins, B. Schuller, J. Härri, R. Troncy, B. Huet, M. Önen, A. Ksentini, J. Conradt, A. Adi, A. Zadorojniy, J. Terken, J. Beskow, A. Morrison, K. Eng, F. Eyben, S. A. Moubayed, and S. Müller, “Emotion-awareness for intelligent Vehicle Assistants: a research agenda,” in Proceedings of the First Workshop on Software Engineering for AI in Autonomous Systems, SEFAIAS, co-located with the 40th International Conference on Software Engineering, ICSE, Gothenburg, Sweden, May 2018, pp. 11-15.
  • G. Hagerer, N. Cummins, F. Eyben, and B. Schuller, “Robust Laughter Detection for Mobile Wellbeing Sensing on Wearable Devices,” in Proceedings of the 8th International Conference on Digital Health (DH), Lyon, France, April 2018, pp. 156-157. [link]
  • Z. Ren, N. Cummins, V. Pandit, J. Han, K. Qian, and B. Schuller, “Learning Image-based Representations for Heart Sound Classification,” in Proceedings of the 8th International Conference on Digital Health (DH), Lyon, France, April 2018, pp. 143-147. [link] [ppt]
  • N. Cummins, S. Amiriparian, S. Ottl, M. Gerczuk, M. Schmitt, and B. Schuller, “Multimodal Bag-of-Words for Cross Domains Sentiment Analysis,” in Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), Calgary, Canada, April 2018, pp. 4954-4958. [link]
  • P. Tzirakis, J. Zhang, and B. Schuller, “End-to-End Speech Emotion Recognition using Deep Neural Networks,” in Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), Calgary, Canada, April 2018, pp. 5089-5093. [link]
  • J. Han, Z. Zhang, Z. Ren, F. Ringeval, and B. Schuller, “Towards Conditional Adversarial Training for Predicting Emotions from Speech,” in Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), Calgary, Canada, April 2018, pp.6822-6826 [link] [ppt]
  • S. Hantke, N. Cummins, and B. Schuller, “What is my dog trying to tell me? The automatic recognition of the context and perceived emotion of dog barks,” in Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), Calgary, Canada, April 2018, pp. 5134-5138. [link] [poster]
  • S. Amiriparian, S. Julka, N. Cummins, and B. Schuller, “Deep Convolutional Recurrent Neural Networks for Rare Sound Event Detection,” in Proceedings of the 44. Jahrestagung für Akustik (DAGA), Munich, Germany, March 2018, pp. 1522-1525.
  • M. Schmitt and B. Schuller, “Deep Recurrent Neural Networks for Emotion Recognition in Speech,” in Proceedings of the 44. Jahrestagung für Akustik (DAGA), Munich, Germany, March 2018, pp. 1537-1540. [ppt]

2017

  • K. Qian, Z. Ren, V. Pandit, Z. Yang, Z. Zhang, and B. Schuller, “Wavelets Revisited for the Classification of Acoustic Scenes,” in Proceedings of the Detection and Classification of Acoustic Scenes and Events 2017 IEEE AASP Challenge Workshop (DCASE), Munich, Germany, November 2017. pp. 108-112. [pdf] [poster]
  • Z. Ren, V. Pandit, K. Qian, Z. Zhang, Z. Yang, and B. Schuller, “Deep Sequential Image Features on Acoustic Scene Classification,” in Proceedings of the Detection and Classification of Acoustic Scenes and Events 2017 IEEE AASP Challenge Workshop (DCASE), Munich, Germany, November 2017. pp. 113-117. [pdf] [poster]
  • S. Amiriparian, M. Freitag, N. Cummins, and B. Schuller, “Sequence to Sequence Autoencoders for Unsupervised Representation Learning from Audio,” in Proceedings of the Detection and Classification of Acoustic Scenes and Events 2017 IEEE AASP Challenge Workshop (DCASE), Munich, Germany, November 2017. pp. 17-21. [pdf] [ppt]
  • S. Amiriparian, N. Cummins, M. Freitag, K. Qian, Z. Ren, V. Pandit, and B. Schuller "The Combined Augsburg / Passau / TUM / ICL System for DCASE 2017", Technical Paper for the Challenge entry at the Detection and Classification of Acoustic Scenes and Events 2017 IEEE AASP Challenge Workshop (DCASE), Munich, Germany, November 2017. 1 page, to appear. [pdf]
  • S. Amiriparian, M. Freitag, N. Cummins, and B. Schuller, “Feature Selection in Multimodal Continuous Emotion Prediction,” in Proceedings of the 2nd International Workshop on Automatic Sentiment Analysis in the Wild (WASA 2017) held in conjunction with the 7th biannual Conference on Affective Computing and Intelligent Interaction (ACII), San Antonio, TX, October 2017. pp. 30–37. [link] [ppt]
  • S. Amiriparian, N. Cummins, S. Ottl, M. Gerczuk, and B. Schuller, “Sentiment Analysis Using Image-based Deep Spectrum Features,” in Proceedings of the 2nd International Workshop on Automatic Sentiment Analysis in the Wild (WASA 2017) held in conjunction with the 7th biannual Conference on Affective Computing and Intelligent Interaction (ACII), San Antonio, TX, October 2017. pp. 26–29. [link [ppt]
  • J. Han, Z. Zhang, M. Schmitt, M. Pantic, and B. Schuller, “From Hard to Soft: Towards more Human-like Emotion Recognition by Modelling the Perception Uncertainty,” in Proceedings of the 25th ACM International Conference on Multimedia (MM), Mountain View, CA, October 2017. pp. 890-897. [link[poster]
  • E. Parada-Cabaleiro, A. Batliner, A. E. Baird, and B. Schuller, “The Perception of Emotion in the Singing Voice,” in Proceedings of the 4th International Digital Libraries for Musicology workshop (DLfM) at the 18th International Society for Music Information Retrieval Conference (ISMIR), Suzhou, China, October 2017, pp. 29-36. [pdf] [ppt]
  • E. Parada-Cabaleiro, A. Batliner, A. E. Baird, and B. Schuller, “The SEILS dataset: Symbolically Encoded Scores in ModernAncient Notation for Computational Musicology,” in Proceedings of the 18th International Society for Music Information Retrieval Conference (ISMIR), Suzhou, China, October 2017. pp. 575-581. [pdf] [poster]

Technical Reports

  • G. Keren, M. Schmitt, T. Kehrenberg, and B. Schuller, “Weakly Supervised One-Shot Detection with Attention Siamese Networks,” arxiv.org, January 2018. 11 pages [pdf]