{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,9,30]],"date-time":"2025-09-30T04:40:10Z","timestamp":1759207210649,"version":"3.37.3"},"reference-count":84,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2019,1,1]],"date-time":"2019-01-01T00:00:00Z","timestamp":1546300800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"}],"funder":[{"DOI":"10.13039\/100011102","name":"Seventh Framework Programme","doi-asserted-by":"publisher","award":["289021"],"award-info":[{"award-number":["289021"]}],"id":[{"id":"10.13039\/100011102","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100007601","name":"HORIZON 2020","doi-asserted-by":"publisher","award":["645378"],"award-info":[{"award-number":["645378"]}],"id":[{"id":"10.13039\/501100007601","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100011102","name":"European Union Seventh Framework Programme","doi-asserted-by":"publisher","award":["338164"],"award-info":[{"award-number":["338164"]}],"id":[{"id":"10.13039\/100011102","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Computer Speech &amp; Language"],"published-print":{"date-parts":[[2019,1]]},"DOI":"10.1016\/j.csl.2018.02.004","type":"journal-article","created":{"date-parts":[[2018,3,15]],"date-time":"2018-03-15T19:01:09Z","timestamp":1521140469000},"page":"156-180","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":31,"special_numbering":"C","title":["Affective and behavioural computing: Lessons learnt from the First Computational Paralinguistics Challenge"],"prefix":"10.1016","volume":"53","author":[{"given":"Bj\u00f6rn","family":"Schuller","sequence":"first","affiliation":[]},{"given":"Felix","family":"Weninger","sequence":"additional","affiliation":[]},{"given":"Yue","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Fabien","family":"Ringeval","sequence":"additional","affiliation":[]},{"given":"Anton","family":"Batliner","sequence":"additional","affiliation":[]},{"given":"Stefan","family":"Steidl","sequence":"additional","affiliation":[]},{"given":"Florian","family":"Eyben","sequence":"additional","affiliation":[]},{"given":"Erik","family":"Marchi","sequence":"additional","affiliation":[]},{"given":"Alessandro","family":"Vinciarelli","sequence":"additional","affiliation":[]},{"given":"Klaus","family":"Scherer","sequence":"additional","affiliation":[]},{"given":"Mohamed","family":"Chetouani","sequence":"additional","affiliation":[]},{"given":"Marcello","family":"Mortillaro","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"year":"2006","author":"Albrecht","series-title":"Social Intelligence: The New Science of Success 2005","key":"10.1016\/j.csl.2018.02.004_bib0001"},{"key":"10.1016\/j.csl.2018.02.004_bib0002","series-title":"Proceedings of Interspeech","first-page":"178","article-title":"Detecting laughter and filled pauses using syllable-based features","author":"An","year":"2013"},{"key":"10.1016\/j.csl.2018.02.004_bib0003","series-title":"Proceedings of Interspeech","first-page":"191","article-title":"Robust and accurate features for detecting and diagnosing autism spectrum disorders","author":"Asgari","year":"2013"},{"key":"10.1016\/j.csl.2018.02.004_bib0004","doi-asserted-by":"crossref","first-page":"1581","DOI":"10.1121\/1.1391244","article-title":"The acoustic features of human laughter","volume":"110","author":"Bachorowski","year":"2001","journal-title":"J. Acoust. Soc. Am."},{"key":"10.1016\/j.csl.2018.02.004_bib0005","doi-asserted-by":"crossref","first-page":"1161","DOI":"10.1037\/a0025827","article-title":"Introducing the Geneva multimodal expression corpus for experimental research on emotion perception","volume":"12","author":"B\u00e4nziger","year":"2012","journal-title":"Emotion"},{"key":"10.1016\/j.csl.2018.02.004_bib0006","series-title":"Proceedings of ISCA Workshop on Speech and Emotion","first-page":"195","article-title":"Desperately seeking emotions: Actors, wizards, and human beings","author":"Batliner","year":"2000"},{"issue":"6","key":"10.1016\/j.csl.2018.02.004_bib0007","doi-asserted-by":"crossref","first-page":"847","DOI":"10.1080\/02699930541000057","article-title":"Distinctions between emotion and mood","volume":"19","author":"Beedie","year":"2005","journal-title":"Cognit. Emot."},{"key":"10.1016\/j.csl.2018.02.004_bib0008","series-title":"Proceedings of Interspeech","first-page":"1043","article-title":"Spontaneous-speech acoustic-prosodic features of children with autism and the interacting psychologist","author":"Bone","year":"2012"},{"doi-asserted-by":"crossref","unstructured":"Bone, D., Black, M. P., Li, M., Metallinou, A., Lee, S., Narayanan, S., 2011. Intoxicated speech detection by fusion of speaker normalized hierarchical features and GMM supervectors, 3217\u20133220.","key":"10.1016\/j.csl.2018.02.004_bib0009","DOI":"10.21437\/Interspeech.2011-805"},{"key":"10.1016\/j.csl.2018.02.004_bib0010","series-title":"Proceedings of Interspeech","first-page":"182","article-title":"Classifying language-related developmental disorders from speech cues: the promise and the potential confounds","author":"Bone","year":"2013"},{"key":"10.1016\/j.csl.2018.02.004_bib0011","doi-asserted-by":"crossref","first-page":"321","DOI":"10.1613\/jair.953","article-title":"SMOTE: Synthetic Minority Over-sampling Technique","volume":"16","author":"Chawla","year":"2002","journal-title":"J. Artif. Intel. Res."},{"issue":"1","key":"10.1016\/j.csl.2018.02.004_bib0012","doi-asserted-by":"crossref","first-page":"73","DOI":"10.1016\/S0010-0277(02)00017-3","article-title":"Using \u201cuh\u201d and \u201cum\u201d in spontaneous speaking","volume":"84","author":"Clark","year":"2002","journal-title":"Cognition"},{"issue":"4","key":"10.1016\/j.csl.2018.02.004_bib0013","doi-asserted-by":"crossref","first-page":"1402","DOI":"10.1016\/j.rasd.2011.01.026","article-title":"Differential language markers of pathology in autism, pervasive developmental disorders not otherwise specified and specific language impairment","volume":"5","author":"Demouy","year":"2011","journal-title":"Res. Autism Spectrum Disorders"},{"key":"10.1016\/j.csl.2018.02.004_bib0014","series-title":"Proceedings of ICMI","first-page":"371","article-title":"Emotion recognition in the wild challenge (EmotiW) challenge and workshop summary","author":"Dhall","year":"2013"},{"key":"10.1016\/j.csl.2018.02.004_bib0015","doi-asserted-by":"crossref","first-page":"1895","DOI":"10.1162\/089976698300017197","article-title":"Approximate statistical tests for comparing supervised classification learning algorithms","volume":"10","author":"Dietterich","year":"1998","journal-title":"Neural Comput."},{"unstructured":"Dumouchel, P., Dehak, N., Attabi, Y., Dehak, R., Boufaden, N.","key":"10.1016\/j.csl.2018.02.004_bib0016"},{"year":"2013","author":"Ekkekakis","series-title":"The Measurement of Affect, Mood, and Emotion: A Guide for Health-Behavioral Research","key":"10.1016\/j.csl.2018.02.004_bib0017"},{"doi-asserted-by":"crossref","unstructured":"Eyben, F., 2015. Real-time Speech and Music Classification by Large Audio Feature Space Extraction. Springer Theses, Springer International Publishing, Switzerland.","key":"10.1016\/j.csl.2018.02.004_bib0018","DOI":"10.1007\/978-3-319-27299-3"},{"key":"10.1016\/j.csl.2018.02.004_bib0019","series-title":"Proceedings of ACM Multimedia","first-page":"835","article-title":"Recent developments in openSMILE, the munich open-source multimedia feature extractor","author":"Eyben","year":"2013"},{"key":"10.1016\/j.csl.2018.02.004_bib0020","series-title":"Proceedings of ACII","first-page":"576","article-title":"openEAR \u2013 introducing the munich open-source emotion and affect recognition toolkit","author":"Eyben","year":"2009"},{"key":"10.1016\/j.csl.2018.02.004_bib0021","series-title":"Proceedings of ACM Multimedia","first-page":"1459","article-title":"openSMILE \u2013 the munich versatile and fast open-source audio feature extractor","author":"Eyben","year":"2010"},{"year":"1994","author":"First","series-title":"Diagnostic and Statistical Manual of Mental Disorders","key":"10.1016\/j.csl.2018.02.004_bib0022"},{"key":"10.1016\/j.csl.2018.02.004_bib0023","series-title":"Proceedings of Interspeech","first-page":"220","article-title":"Detecting autism, emotions and social signals using adaboost","author":"Gosztolya","year":"2013"},{"key":"10.1016\/j.csl.2018.02.004_bib0024","series-title":"Proceedings of Interspeech","first-page":"200","article-title":"Let me finish: automatic conflict detection using speaker overlap","author":"Gr\u00e8zes","year":"2013"},{"key":"10.1016\/j.csl.2018.02.004_bib0025","series-title":"Proceedings of Interspeech","first-page":"173","article-title":"Paralinguistic event detection from speech using probabilistic time-series smoothing and masking","author":"Gupta","year":"2013"},{"issue":"2","key":"10.1016\/j.csl.2018.02.004_bib0026","doi-asserted-by":"crossref","first-page":"451","DOI":"10.1214\/aos\/1028144844","article-title":"Classification by pairwise coupling","volume":"26","author":"Hastie","year":"1998","journal-title":"Annals Stat."},{"issue":"2","key":"10.1016\/j.csl.2018.02.004_bib0027","doi-asserted-by":"crossref","first-page":"392","DOI":"10.1016\/j.csl.2013.06.002","article-title":"Speaker state classification based on fusion of asymmetric simple partial least squares (simpls) and support vector machines","volume":"28","author":"Huang","year":"2014","journal-title":"Comput. Speech Lang."},{"key":"10.1016\/j.csl.2018.02.004_bib0028","series-title":"Proceedings of Interspeech","first-page":"278","article-title":"Modulation spectrum analysis for speaker personality trait recognition","author":"Ivanov","year":"2012"},{"key":"10.1016\/j.csl.2018.02.004_bib0029","series-title":"Proceedings of Interspeech","first-page":"153","article-title":"Non-linguistic vocalisation recognition based on hybrid GMM-SVM approach","author":"Janicki","year":"2013"},{"issue":"2","key":"10.1016\/j.csl.2018.02.004_bib0030","doi-asserted-by":"crossref","first-page":"420","DOI":"10.1016\/j.csl.2013.09.005","article-title":"Level of interest sensing in spoken dialog using decision-level fusion of acoustic and lexical evidence","volume":"28","author":"Jeon","year":"2014","journal-title":"Comput. Speech Lang."},{"key":"10.1016\/j.csl.2018.02.004_bib0031","series-title":"Proceedings of Interspeech","first-page":"534","article-title":"Intelligibility classification of pathological speech using fusion of multiple subsystems","author":"Kim","year":"2012"},{"key":"10.1016\/j.csl.2018.02.004_bib0032","series-title":"Proceedings of ACM Multimedia","first-page":"793","article-title":"Predicting the conflict level in television political debates: an approach based on crowdsourcing, nonverbal communication and gaussian processes","author":"Kim","year":"2012"},{"issue":"2","key":"10.1016\/j.csl.2018.02.004_bib0033","doi-asserted-by":"crossref","first-page":"187","DOI":"10.1109\/TAFFC.2014.2324564","article-title":"Predicting continuous conflict perceptionwith bayesian gaussian processes","volume":"5","author":"Kim","year":"2014","journal-title":"IEEE Trans. Affect. Comput."},{"key":"10.1016\/j.csl.2018.02.004_bib0034","series-title":"Proceedings of Interspeech","first-page":"187","article-title":"Classification of developmental disorders from speech signals using submodular feature selection.","author":"Kirchhoff","year":"2013"},{"key":"10.1016\/j.csl.2018.02.004_bib0035","series-title":"Proceedings of Interspeech","first-page":"2822","article-title":"Brno university of technology system for interspeech 2010 paralinguistic challenge","author":"Kockmann","year":"2010"},{"issue":"6","key":"10.1016\/j.csl.2018.02.004_bib0036","doi-asserted-by":"crossref","first-page":"618","DOI":"10.1016\/S0022-1031(02)00510-3","article-title":"Inferring speakers\u2019 physical attributes from their voices","volume":"38","author":"Krauss","year":"2002","journal-title":"J. Exper. Social Psychol."},{"issue":"9","key":"10.1016\/j.csl.2018.02.004_bib0037","doi-asserted-by":"crossref","first-page":"1162","DOI":"10.1016\/j.specom.2011.06.004","article-title":"Emotion recognition using a hierarchical binary decision tree approach","volume":"53","author":"Lee","year":"2011","journal-title":"Speech Commun."},{"key":"10.1016\/j.csl.2018.02.004_bib0038","series-title":"Proceedings of Interspeech","first-page":"215","article-title":"Ensemble of machine learning and acoustic segment model techniques for speech emotion and autism spectrum disorders recognition","author":"Lee","year":"2013"},{"issue":"5","key":"10.1016\/j.csl.2018.02.004_bib0039","doi-asserted-by":"crossref","first-page":"425","DOI":"10.1016\/j.specom.2009.01.004","article-title":"Peaks\u2013a system for the automatic evaluation of voice and speech disorders","volume":"51","author":"Maier","year":"2009","journal-title":"Speech Commun."},{"key":"10.1016\/j.csl.2018.02.004_bib0040","series-title":"Speech and Automata in Health Care (Speech Technology and Text Mining in Medicine and Healthcare)","first-page":"207","article-title":"Voice-enabled Assistive Robots for Handling Autism Spectrum Conditions: An Examination of theRole of Prosody","author":"Marchi","year":"2014"},{"key":"10.1016\/j.csl.2018.02.004_bib0041","series-title":"Proceedings of Interspeech","first-page":"115","article-title":"Typicality and emotion in the voice of children with autism spectrum condition: Evidence across three languages","author":"Marchi","year":"2015"},{"key":"10.1016\/j.csl.2018.02.004_bib0042","series-title":"Proceedings of Interspeech","first-page":"195","article-title":"Suprasegmental information modelling for autism disorder spectrum and specific language impairment classification","author":"Mart\u0131nez","year":"2013"},{"key":"10.1016\/j.csl.2018.02.004_bib0043","doi-asserted-by":"crossref","first-page":"325","DOI":"10.1080\/1368282031000154204","article-title":"Prosody in autism spectrum disorders: a critical review","volume":"38","author":"McCann","year":"2003","journal-title":"Int. J. Lang. Commun. Disorder"},{"key":"10.1016\/j.csl.2018.02.004_bib0044","series-title":"Proceedings of Interspeech","first-page":"2818","article-title":"Age and gender classification using fusion of acoustic and prosodic features","author":"Meinedo","year":"2010"},{"key":"10.1016\/j.csl.2018.02.004_bib0045","series-title":"Proceedings of ACM Multimedia Workshop on Social Signal Processing","first-page":"17","article-title":"The voice of personality: mapping nonverbal vocal behavior into trait attributions","author":"Mohammadi","year":"2010"},{"key":"10.1016\/j.csl.2018.02.004_bib0046","series-title":"Proceedings of Interspeech","first-page":"526","article-title":"Pitch and intonation contribution to speakers\u2019 traits classification","author":"Montaci\u00e9","year":"2012"},{"key":"10.1016\/j.csl.2018.02.004_bib0047","series-title":"Proceedings of Interspeech","first-page":"158","article-title":"Characteristic contours of syllabic-level units in laughter","author":"Oh","year":"2013"},{"issue":"4","key":"10.1016\/j.csl.2018.02.004_bib0048","doi-asserted-by":"crossref","first-page":"108","DOI":"10.1109\/MSP.2007.4286569","article-title":"Social signal processing [exploratory dsp]","volume":"24","author":"Pentland","year":"2007","journal-title":"IEEE Signal Process. Mag."},{"year":"1997","author":"Picard","series-title":"Affective Computing","key":"10.1016\/j.csl.2018.02.004_bib0049"},{"key":"10.1016\/j.csl.2018.02.004_bib0050","series-title":"Advances in Large Margin Classifiers","first-page":"61","article-title":"Probabilistic outputs for support vector machines and comparisons to regularized likelihood methods","author":"Platt","year":"1999"},{"key":"10.1016\/j.csl.2018.02.004_bib0051","series-title":"Proceedings of Interspeech","first-page":"210","article-title":"Random subset feature selection in automatic recognition of developmental disorders, affective states, and level of conflict from speech","author":"R\u00e4s\u00e4nen","year":"2013"},{"key":"10.1016\/j.csl.2018.02.004_bib0052","doi-asserted-by":"crossref","first-page":"1328","DOI":"10.1109\/TASL.2010.2090147","article-title":"Automatic intonation recognition for the prosodic assessment of language impaired children","volume":"19","author":"Ringeval","year":"2011","journal-title":"IEEE Trans. Audio Speech Lang. Process."},{"key":"10.1016\/j.csl.2018.02.004_bib0053","series-title":"Proceedings of Interspeech","doi-asserted-by":"crossref","first-page":"1210","DOI":"10.21437\/Interspeech.2016-766","article-title":"Automatic analysis of typical and atypical encoding of spontaneous emotion in the voice of children","author":"Ringeval","year":"2016"},{"key":"10.1016\/j.csl.2018.02.004_bib0054","series-title":"Proceedings of International Workshop on Audio\/Visual Emotion Challenge (AVEC), co-Located with ACM Multimedia","first-page":"3","article-title":"AV+EC 2015 \u2013 the first affect recognition challenge bridging across audio, video, and physiological data","author":"Ringeval","year":"2015"},{"key":"10.1016\/j.csl.2018.02.004_bib0055","series-title":"The New Handbook of Methods in Nonverbal Behavior Research","first-page":"199","article-title":"Conducting judgment studies: some methodological issues","author":"Rosenthal","year":"2005"},{"key":"10.1016\/j.csl.2018.02.004_bib0056","doi-asserted-by":"crossref","first-page":"807","DOI":"10.1007\/s11422-010-9272-8","article-title":"Solidarity and conflict: aligned and misaligned prosody as a transactional resource in intra- and intercultural communication involving power differences","volume":"5","author":"Roth","year":"2010","journal-title":"Cult. Stud. Sci. Educ."},{"issue":"1","key":"10.1016\/j.csl.2018.02.004_bib0057","doi-asserted-by":"crossref","first-page":"145","DOI":"10.1037\/0033-295X.110.1.145","article-title":"Core affect and the psychological construction of emotion","volume":"110","author":"Russell","year":"2003","journal-title":"Psychol. Rev."},{"issue":"7","key":"10.1016\/j.csl.2018.02.004_bib0058","doi-asserted-by":"crossref","first-page":"1259","DOI":"10.1080\/02699930902809375","article-title":"Emotion, core affect, and psychological construction","volume":"23","author":"Russell","year":"2009","journal-title":"Cognit. Emotion"},{"key":"10.1016\/j.csl.2018.02.004_bib0059","series-title":"Proceedings of Interspeech","first-page":"983","article-title":"Laying the foundation for in-car alcohol detection by speech","author":"Schiel","year":"2009"},{"key":"10.1016\/j.csl.2018.02.004_bib0060","series-title":"Proceedings of ITG Conference on Speech Communication","first-page":"264\u2014268","article-title":"Towards cross-lingual automatic diagnosis of autism spectrum condition in children\u2019s voices","author":"Schmitt","year":"2016"},{"issue":"2","key":"10.1016\/j.csl.2018.02.004_bib0061","doi-asserted-by":"crossref","first-page":"77","DOI":"10.1007\/s10772-011-9090-8","article-title":"Affective speaker state analysis in the presence of reverberation","volume":"14","author":"Schuller","year":"2011","journal-title":"Int. J. Speech Technol."},{"issue":"4","key":"10.1016\/j.csl.2018.02.004_bib0062","doi-asserted-by":"crossref","first-page":"97","DOI":"10.1109\/MSP.2012.2192211","article-title":"The computational paralinguistics challenge","volume":"29","author":"Schuller","year":"2012","journal-title":"IEEE Signal Process. Mag."},{"year":"2014","author":"Schuller","series-title":"Computational Paralinguistics: Emotion, Affect and Personality in Speech and Language Processing","key":"10.1016\/j.csl.2018.02.004_bib0063"},{"issue":"9\/10","key":"10.1016\/j.csl.2018.02.004_bib0064","doi-asserted-by":"crossref","first-page":"1062","DOI":"10.1016\/j.specom.2011.01.011","article-title":"Recognising realistic emotions and affect in speech: State of the art and lessons learnt from the first challenge","volume":"53","author":"Schuller","year":"2011","journal-title":"Speech Commun."},{"key":"10.1016\/j.csl.2018.02.004_bib0065","series-title":"Proceedings of Interspeech","first-page":"312","article-title":"The INTERSPEECH 2009 Emotion Challenge","author":"Schuller","year":"2009"},{"issue":"1","key":"10.1016\/j.csl.2018.02.004_bib0066","first-page":"4","article-title":"Paralinguistics in speech and language \u2013 state-of-the-art and the challenge","volume":"27","author":"Schuller","year":"2013","journal-title":"Special Issue on Paralinguistics in Naturalistic Speech and Language, Comput. Speech Lang."},{"key":"10.1016\/j.csl.2018.02.004_bib0067","series-title":"Proceedings of Interspeech","first-page":"2794","article-title":"The INTERSPEECH 2010 Paralinguistic Challenge \u2013 Age, Gender, and Affect","author":"Schuller","year":"2010"},{"issue":"1","key":"10.1016\/j.csl.2018.02.004_bib0068","doi-asserted-by":"crossref","first-page":"100","DOI":"10.1016\/j.csl.2014.08.003","article-title":"A survey on perceived speaker traits: Personality, likability, pathology, and the first challenge","volume":"29","author":"Schuller","year":"2015","journal-title":"Comput. Speech Lang."},{"key":"10.1016\/j.csl.2018.02.004_bib0069","series-title":"Proceedings of Interspeech","first-page":"254","article-title":"The INTERSPEECH 2012 speaker trait challenge","author":"Schuller","year":"2012"},{"key":"10.1016\/j.csl.2018.02.004_bib0070","series-title":"Proceedings of Interspeech","first-page":"3201","article-title":"The INTERSPEECH 2011 speaker state challenge","author":"Schuller","year":"2011"},{"issue":"2","key":"10.1016\/j.csl.2018.02.004_bib0071","first-page":"346","article-title":"Medium-term speaker states \u2013 a review on intoxication, sleepiness and the first challenge","volume":"28","author":"Schuller","year":"2014","journal-title":"Special Issue on Broadening the View on Speaker Analysis, Comput. Speech Lang."},{"key":"10.1016\/j.csl.2018.02.004_bib0072","series-title":"Proceedings of Interspeech","first-page":"205","article-title":"Gmm based speaker variability compensated system for interspeech 2013 compare emotion challenge","author":"Sethu","year":"2013"},{"year":"2009","author":"Steidl","series-title":"Automatic Classification of Emotion-Related User States in Spontaneous Children\u2019s Speech","key":"10.1016\/j.csl.2018.02.004_bib0073"},{"key":"10.1016\/j.csl.2018.02.004_bib0074","series-title":"Proceedings of International Congress of Phonetic Sciences (ICPhS)","first-page":"1958","article-title":"Acoustic features of four types of laughter in natural conversational speech","author":"Tanaka","year":"2011"},{"key":"10.1016\/j.csl.2018.02.004_bib0075","series-title":"Proceedings of the International Conference on Affective Computing and Intelligent Interaction","first-page":"981","article-title":"Affective computing: A review","author":"Tao","year":"2005"},{"key":"10.1016\/j.csl.2018.02.004_bib0076","doi-asserted-by":"crossref","first-page":"215","DOI":"10.1177\/1362361310363281","article-title":"Computational prosodic markers for autism","volume":"14","author":"Van Santen","year":"2010","journal-title":"Autism"},{"issue":"2","key":"10.1016\/j.csl.2018.02.004_bib0077","doi-asserted-by":"crossref","first-page":"93","DOI":"10.1023\/B:JONB.0000023654.73558.72","article-title":"Laughter in conversation: Features of occurrence and acoustic structure","volume":"28","author":"Vettin","year":"2004","journal-title":"J. Nonverbal Behav."},{"key":"10.1016\/j.csl.2018.02.004_bib0078","doi-asserted-by":"crossref","first-page":"1743","DOI":"10.1016\/j.imavis.2008.11.007","article-title":"Social signal processing: survey of an emerging domain","volume":"27","author":"Vinciarelli","year":"2009","journal-title":"Image Vis. Comput."},{"issue":"1","key":"10.1016\/j.csl.2018.02.004_bib0079","doi-asserted-by":"crossref","first-page":"69","DOI":"10.1109\/T-AFFC.2011.27","article-title":"Bridging the gap between social animal and unsocial machine: a survey of social signal processing","volume":"3","author":"Vinciarelli","year":"2012","journal-title":"IEEE Trans. Affect. Comput."},{"key":"10.1016\/j.csl.2018.02.004_bib0080","doi-asserted-by":"crossref","first-page":"60","DOI":"10.1007\/978-3-642-34584-5_5","article-title":"From nonverbal cues to perception: personality and social attractiveness","author":"Vinciarelli","year":"2012","journal-title":"Cognitive Behav. Syst."},{"key":"10.1016\/j.csl.2018.02.004_bib0081","series-title":"Proceedings of the International Conference on Multimedia and Expo","first-page":"474","article-title":"Comparing feature sets for acted and spontaneous speech in view of automatic emotion recognition","author":"Vogt","year":"2005"},{"key":"10.1016\/j.csl.2018.02.004_bib0082","series-title":"Proceedings of the Interspeech","first-page":"168","article-title":"Using phonetic patterns for detecting social cues in natural conversations","author":"Wagner","year":"2013"},{"year":"2005","author":"Witten","series-title":"Data Mining: Practical Machine Learning Tools and Techniques","key":"10.1016\/j.csl.2018.02.004_bib0083"},{"year":"2006","author":"Young","series-title":"The HTK book (v3.4)","key":"10.1016\/j.csl.2018.02.004_bib0084"}],"container-title":["Computer Speech &amp; Language"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0885230816303928?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0885230816303928?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2022,8,16]],"date-time":"2022-08-16T04:00:47Z","timestamp":1660622447000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S0885230816303928"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019,1]]},"references-count":84,"alternative-id":["S0885230816303928"],"URL":"https:\/\/doi.org\/10.1016\/j.csl.2018.02.004","relation":{},"ISSN":["0885-2308"],"issn-type":[{"type":"print","value":"0885-2308"}],"subject":[],"published":{"date-parts":[[2019,1]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"Affective and behavioural computing: Lessons learnt from the First Computational Paralinguistics Challenge","name":"articletitle","label":"Article Title"},{"value":"Computer Speech & Language","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.csl.2018.02.004","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"Crown Copyright \u00a9 2018 Published by Elsevier Ltd. All rights reserved.","name":"copyright","label":"Copyright"}]}}