Mousavi M., Negro R. and Riccardi G. An Unsupervised Approach to Extract Life-Events from Personal Narratives in the Mental Health Domain (Conference) Eighth Italian Conference on Computational Linguistics, 2022. (BibTeX | Tags: Machine Learning, Signal Annotation and Interpretation) Bayerl P. S., Tammewar A., Riedhammer K. and Riccardi G. Detecting Emotion Carriers By Combining Acoustic and Lexical Representations (Conference) IEEE Automatic Speech Recognition and Understanding Conference, 2021. (BibTeX | Tags: Affective Computing, Machine Learning) Torres M. J., Ravanelli M., Medina-Devilliers S., Lerner D. M. and Riccardi G. Interpretable SincNet-based Deep Learning for Emotion Recognition in Individuals with Autism (Conference) IEEE Conf. Engineering in Medicine and Biology, Conference, 2021. (Links | BibTeX | Tags: Affective Computing, Autism, Machine Learning, Signal Annotation and Interpretation) Torres M. J., Clarkson T., Hauschild K., Luhmann C. C., Lerner D. M. and Riccardi G. Facial emotions are accurately encoded in the brains of those with autism: A deep learning approach (Article) Biological Psychiatry: Cognitive Neuroscience and Neuroimaging, 2021. (Links | BibTeX | Tags: Affective Computing, Autism, Machine Learning, Signal Annotation and Interpretation) Marinelli F., Cervone A., Tortoreto G., Stepanov E. A., Di Fabbrizio G., Riccardi G. Active Annotation: bootstrapping annotation lexicon and guidelines for supervised NLU learning (Conference) 2019. (Links | BibTeX | Tags: Machine Learning, Signal Annotation and Interpretation) Tortoreto G., Stepanov E. A., Cervone A., Dubiel M., Riccardi G. Affective Behaviour Analysis of On-line User Interactions: Are On-line Support Groups more Therapeutic than Twitter? (Conference) 2019. (Links | BibTeX | Tags: Discourse, Health, Language Analytics, Machine Learning) Mayor Torres, J.M., Clarkson, T., Luhmann, C. C., Riccardi, G., Lerner, M.D. 2019. (Links | BibTeX | Tags: Affective Computing, Autism, Machine Learning) Stepanov E. A., Lathuilierev S., Chowdhury S. A., Ghosh A., Vieriu R.D., Sebe N. and Riccardi G. Depression Severity Estimation from Multiple Modalities (Article) 2018, (EXCELLENT Paper AWARD). (Links | BibTeX | Tags: Health Analytics, Machine Learning) Dias R., Conboy M. H., Gabany M. J., Clarke A. L. , Osterweil J. L., Avrunin S. G., Arney D., Goldman M. J., Riccardi G., Yule J. S., Zenati A. M. 2018. (Links | BibTeX | Tags: Health Analytics, Interactive Systems, Machine Learning, Signal Annotation and Interpretation) Mayor Torres, J.M., Clarkson, T., Stepanov E. A. , Luhmann C. C., Lerner, M.D., Riccardi, G. Enhanced Error Decoding from Error-Related Potentials using Convolutional Neural Networks (Conference) 2018. (Links | BibTeX | Tags: Affective Computing, Autism, Machine Learning) Dias RD, Conboy HM, Gabany JM, Clarke LA, Osterweil LJ, Arney D, Goldman JM, Riccardi G, Avrunin GS, Yule SJ, Zenati MA. Intelligent Interruption Management System to Enhance Safety and Performance in Complex Surgical and Robotic Procedures (Proceeding) 2018. (BibTeX | Tags: Interactive Systems, Machine Learning) Cervone A., Tortoreto G., Mezza S., Gambi E. and Riccardi G Roving Mind: a balancing act between open–domain and engaging dialogue systems (Conference) 2017. (Links | BibTeX | Tags: Conversational and Interactive Systems , Interactive Systems, Machine Learning, Natural Language Processing, Speech Processing) Celli F., Stepanov A. E., Poesio M. and Riccardi G. Predicting Brexit: Classifying Agreement is Better than Sentiment and Pollsters (Proceeding) Proc. PEOPLES Workshop at COLING, Osaka 2016., 2016. (Abstract | Links | BibTeX | Tags: Affective Computing, Machine Learning, Signal Annotation and Interpretation) Celli F., Ghosh A., Alam F. and Riccardi G. Information Processing and Management, Nov 2015, 2015. (Abstract | Links | BibTeX | Tags: Machine Learning, Natural Language Processing, Signal Annotation and Interpretation) Mogessie M, Riccardi G. and Ronchetti M. Predicting Students’ Final Exam Scores from their Course Activities (Article) Proc. IEEE Frontiers in Education, El Paso ( USA), 2015., 2015. (Abstract | Links | BibTeX | Tags: Education Analytics, Machine Learning) Chowdhury A,, Danieli M. and Riccardi G. The Role of Speakers and Context in Classifying Competition in Overlapping Speech (Conference) 2015. (Abstract | Links | BibTeX | Tags: Conversational and Interactive Systems , Machine Learning, Signal Annotation and Interpretation) Chowdhury A, Calvo M., Ghosh A., Stepanov A. E., Bayer A. O., Riccardi G., Garcia F. and Sanchis E. Selection and Aggregation Techniques for Crowdsourced Semantic Annotation Task (Conference) 2015. (Abstract | Links | BibTeX | Tags: Machine Learning, Signal Annotation and Interpretation, Statistical Machine Translation) Ghosh A, Danieli M. and Riccardi G. Annotation and Prediction of Stress and Workload from Physiological and Inertial Signals (Conference) 2015. (Abstract | Links | BibTeX | Tags: Health Analytics, Machine Learning, Signal Annotation and Interpretation) Ghosh A, Mayor Torres J.M., Danieli M. and Riccardi G. Detection of Essential Hypertension with Physiological Signals from Wearable Devices (Conference) 2015. (Abstract | Links | BibTeX | Tags: Health Analytics, Machine Learning, Signal Annotation and Interpretation) Stepanov E., Bayer A. O., Riccardi G., The UniTN Discourse Parser in CoNLL 2015 Shared Task (Conference) 2015. (Abstract | Links | BibTeX | Tags: Discourse, Machine Learning, Natural Language Processing) Vinciarelli A., Esposito A., Andre’ E., Bonin F., Chetouani M., Cohn F. J., Cristani M., Fuhrmann F., Gilmartin E., Hammal Z., Heylen D., Kaiser R., Koutsombogera M., Potamianos A., Renals S., Riccardi G., Salah A. G. Cognitive Computation, pp. 1-17, April 2015, 2015. (Abstract | Links | BibTeX | Tags: Conversational and Interactive Systems , Machine Learning, Speech Processing) Parodi S, Riccardi G, Castagnino N, Tortolina L, Maffei M, Zoppoli G, Nencioni A, Ballestrero A, Patrone F. Systems Medicine in Oncology: Signaling-networks modeling and new generation decision-support systems (Book) Methods Molecular Biology, Vol. 1386, Schmitz U and Wolkenhauer O (Eds): Systems Medicine, Springer Science press., 2015. (BibTeX | Tags: Machine Learning, Signal Annotation and Interpretation) Bayer A. O. and Riccardi G. Instance-Based On-Line Language Model Adaptation (Conference) 2013. (Abstract | Links | BibTeX | Tags: Machine Learning, Speech Processing) Riccardi G., Ghosh A., Chowdhury S. A. and Bayer A. O. Motivational Feedback in Crowdsourcing: a Case Study in Speech Transcriptions (Conference) 2013. (Abstract | Links | BibTeX | Tags: Affective Computing, Conversational and Interactive Systems , Machine Learning, Signal Annotation and Interpretation) Bayer A. O. and Riccardi G. On-line Adaptation of Semantic Models for Spoken Language Understanding (Conference) 2013. (Abstract | Links | BibTeX | Tags: Machine Learning, Speech Processing) Ghosh S., Johansson R., Riccardi G. and Tonelli S. Improving the Recall of a Discourse Parser by Constraint-Based Postprocessing (Conference) 2012. (BibTeX | Tags: Discourse, Machine Learning, Natural Language Processing) Ghosh S., Riccardi G. and Johansson R. Global Features for Shallow Discourse Parsing (Conference) 2012. (Abstract | Links | BibTeX | Tags: Machine Learning, Natural Language Processing) Bayer A. O. and Riccardi G. Joint Language Models for Automatic Speech Recognition and Understanding (Conference) 2012. (Abstract | Links | BibTeX | Tags: Machine Learning, Speech Processing) Garcia F., Hurtado L. F., Segarra E., Sanchis E. and Riccardi G. Combining Machine Translation Systems for Spoken Language Understanding Portability (Conference) 2012. (Abstract | Links | BibTeX | Tags: Machine Learning, Signal Annotation and Interpretation, Speech Processing, Statistical Machine Translation) Ghosh S., Tonelli S., Riccardi G. and Johansson R. End-to-End Discourse Parser Evaluation (Conference) 2011. (Abstract | Links | BibTeX | Tags: Discourse, Machine Learning, Natural Language Processing, Signal Annotation and Interpretation) Moschitti A., Chu-Carroll J., Patwardhan S., Fan J. and Riccardi G. Using Syntacting and Semantic Structural Kernels for Classifying Definition Questions in Jeopardy! (Conference) 2011. (Abstract | Links | BibTeX | Tags: Conversational and Interactive Systems , Machine Learning, Natural Language Processing, Signal Annotation and Interpretation) Stepanov E. and Riccardi G. Detecting General Opinions from Customer Surveys (Conference) 2011. (Abstract | Links | BibTeX | Tags: Conversational and Interactive Systems , Machine Learning, Natural Language Processing) Nguyen T. T., Moschitti A. and Riccardi G. Kernel-based Reranking for Named-Entity Extraction (Conference) 2010. (Abstract | Links | BibTeX | Tags: Machine Learning, Natural Language Processing, Signal Annotation and Interpretation) Nguyen T. T., Moschitti A. and Riccardi G. Convolution Kernels on Constituent, Dependency and Sequential Structures for Relation Extraction (Conference) 2009. (Abstract | Links | BibTeX | Tags: Machine Learning, Natural Language Processing) Baggia P., Cutugno F., Danieli M., Pieraccini R. The Multisite 2009 EVALITA Spoken Dialog System Evaluation (Conference) 2009. (Abstract | Links | BibTeX | Tags: Machine Learning, Natural Language Processing, Signal Annotation and Interpretation) Rodríguez K. J., Dipper S., Götze M., Poesio P., Riccardi G., Raymond C., Wisniewska J. Standoff Coordination for Multi-Tool Annotation in a Dialogue Corpus (Conference) 2009. (BibTeX | Tags: Machine Learning, Natural Language Processing, Signal Annotation and Interpretation) Quarteroni S., Dinarelli M. and Riccardi G. Ontology-Based Grounding of Spoken Language Understanding (Conference) 2009. (Abstract | Links | BibTeX | Tags: Machine Learning, Natural Language Processing, Speech Processing) Dinarelli M., Moschitti A. and Riccardi G. Re-Ranking Models For Spoken Language Understanding (Conference) 2009. (Abstract | Links | BibTeX | Tags: Machine Learning, Natural Language Processing, Speech Processing) Dinarelli M., Moschitti A. and Riccardi G. Concept Segmentation and Labeling for Conversational Speech (Conference) 2009. (Abstract | Links | BibTeX | Tags: Machine Learning, Natural Language Processing, Speech Processing) Varges S., Riccardi G., Quarteroni S., Ivanov A. V. The Exploration/Exploitation Trade-Off in Reinforcement Learning for Dialogue Management (Conference) 2009. (Abstract | Links | BibTeX | Tags: Conversational and Interactive Systems , Machine Learning, Speech Processing) Varges S., Riccardi G., Quarteroni S., Ivanov A. V. and Roberti P. Leveraging POMDPs trained with User Simulations and Rule-Based Dialog Management in a SDS (Conference) 2009. (Abstract | Links | BibTeX | Tags: Conversational and Interactive Systems , Machine Learning, Speech Processing) Griol D., Riccardi G. and Sanchis E. Learning the Structure of Human-Computer and Human-Human Spoken Conversations (Conference) 2009. (Abstract | Links | BibTeX | Tags: Conversational and Interactive Systems , Machine Learning, Speech Processing) Sporka A. J., Jakub F. and Riccardi G. Can Machines Call People?- User Experience While Answering Telephone Calls Initiated by Machine (Conference) 2009. (Abstract | Links | BibTeX | Tags: Conversational and Interactive Systems , Machine Learning, Speech Processing) Varges S., Riccardi G., Quarteroni S., Ivanov A. V. and Roberti P. On-Line Strategy Computation in Spoken Dialog Systems (Conference) 2009. (Links | BibTeX | Tags: Conversational and Interactive Systems , Machine Learning, Speech Processing) Riccardi G., Mosca N., Roberti P. and Baggia P. The Voice Multimodal Application Framework (Conference) 2009. (BibTeX | Tags: Conversational and Interactive Systems , Machine Learning, Speech Processing) Riccardi G., Baggia P. and Roberti P. Spoken Dialog Systems: From Theory to Technology (Conference) 2009. (BibTeX | Tags: Conversational and Interactive Systems , Machine Learning, Speech Processing) Quarteroni S., Riccardi G. and Dinarelli M. What's in an Ontology for Spoken Language Understanding (Conference) 2009. (Abstract | Links | BibTeX | Tags: Machine Learning, Natural Language Processing, Speech Processing) Dinarelli M., Moschitti A. and Riccardi G. Re-Ranking Models Based on Small Training Data for Spoken Language Understanding (Conference) 2009. (Abstract | Links | BibTeX | Tags: Machine Learning, Natural Language Processing, Speech Processing) Sebastian V., Riccardi G. and Quarteroni S. Persistent Information State in a Data-Centric Architecture (Conference) 2008. (BibTeX | Tags: Conversational and Interactive Systems , Machine Learning, Speech Processing) Raymond C. and Riccardi G. Learning with Noisy Supervision for Spoken Language Understanding (Conference) 2008. (Abstract | Links | BibTeX | Tags: Machine Learning, Natural Language Processing, Speech Processing) Dinarelli M., Moschitti A., Riccardi G. Joint Generative And Discriminative Models For Spoken Language Understanding (Conference) 2008. (BibTeX | Tags: Machine Learning, Natural Language Processing, Speech Processing) Bisazza A., Dinarelli M., Quarteroni S., Tonelli S., Moschitti A., Riccardi G. Semantic Annotations For Conversational Speech: from speech transcriptions to predicate argument structures (Conference) 2008. (BibTeX | Tags: Machine Learning, Natural Language Processing) Coppola B., Moschitti A., Tonelli S., Riccardi G. Automatic FrameNet-Based Annotation of Conversational Speech (Conference) 2008. (Abstract | Links | BibTeX | Tags: Machine Learning, Natural Language Processing) De Mori R., Bechet F., Hakkani-Tur D., McTear M., Riccardi G. and Tur G. Spoken Language Understanding (Article) IEEE Signal Processing Magazine vol. 25, pp.50-58 ,2008, 2008. (BibTeX | Tags: Machine Learning, Natural Language Processing, Speech Processing) Rodríguez K., Raymond C. and Riccardi G. Active Annotation in the LUNA Italian Corpus of Spontaneous Dialogues (Conference) 2008. (BibTeX | Tags: Machine Learning, Natural Language Processing, Signal Annotation and Interpretation) Raymond C., Riccardi G. Generative and Discriminative Algorithms for Spoken Language Understanding (Conference) 2007. (BibTeX | Tags: Machine Learning, Natural Language Processing, Speech Processing) Varges S. and Riccardi G. A Data-Centric Architecture for Data-Driven Spoken Dialog Systems (Conference) 2007. (BibTeX | Tags: Conversational and Interactive Systems , Machine Learning, Speech Processing) Moschitti A., Riccardi G. and Raymond C. Spoken Language Understanding with Kernels for Syntactic/Semantic Structures (Conference) 2007. (Abstract | Links | BibTeX | Tags: Machine Learning, Natural Language Processing, Speech Processing) Raymond C., Riccardi G., Rodríguez K. J. and Wisniewska J. The LUNA Corpus: an Annotation Scheme for a Multi-domain Multi-lingual Dialogue Corpus (Conference) 2007. (BibTeX | Tags: Machine Learning, Natural Language Processing, Signal Annotation and Interpretation) Riccardi G. and Baggia P. Spoken Dialog Systems: From Theory to Technology (Article) Edizione della Normale di Pisa, 2006, 2006. (BibTeX | Tags: Conversational and Interactive Systems , Machine Learning, Speech Processing) Coppola B., Moschitti A., Riccardi G. Shallow Semantic Parsing for Spoken Language Understanding (Conference) 2006. (Abstract | Links | BibTeX | Tags: Machine Learning, Natural Language Processing, Speech Processing) Gupta N., Tur G., Hakkani-Tur D., Bangalore S., Riccardi G. and Rahim M. The AT&T Spoken Language Understanding System (Article) IEEE Trans. on Audio, Speech and Language Processing, volume 14, Issue 1, pp. 213-22, 2006, 2006. (Abstract | Links | BibTeX | Tags: Machine Learning, Natural Language Processing, Speech Processing) Hakkani-Tur D., Riccardi G. and Tur G. An Active Approach to spoken Language Processing (Article) ACM Transactions on Speech and Language Processing, Vol. 3, No. 3, pp 1-31, 2006, 2006. (Abstract | Links | BibTeX | Tags: Machine Learning, Natural Language Processing, Speech Processing) Potamianos A., Narayanan S. and Riccardi G. Adaptive Categorical Understanding for Spoken Dialogue Systems' (Article) Potamianos A., Narayanan S and Riccardi, G., 2005. (Abstract | Links | BibTeX | Tags: Conversational and Interactive Systems , Machine Learning, Speech Processing) Hakkani-Tur D., Tur G., Riccardi G. and Kim H. K. Error Prediction in Spoken Dialog: from Signal-to-Noise Ratio to Semantic Confidence Scores (Conference) 2005. (BibTeX | Tags: Conversational and Interactive Systems , Machine Learning, Speech Processing) Bechet F., Riccardi G. and Hakkani-Tur D. Mining Spoken dialogue Corpora for system Evaluation and Modeling (Conference) 2005. (BibTeX | Tags: Conversational and Interactive Systems , Machine Learning, Speech Processing) Karahan M., Hakkani-Tur D., Riccardi G. and Tur G. Combining Classifiers for Spoken Language Understanding (Conference) 2005. (BibTeX | Tags: Machine Learning, Natural Language Processing, Speech Processing) Riccardi G. and Hakkani-Tur D. Active Learning: Theory and Applications to Automatic Speech Recognition (Article) IEEE Trans. on Speech and Audio, vol. 13, n.4 , pp. 504-511, 2005, 2005. (Abstract | Links | BibTeX | Tags: Machine Learning) Tur G., Hakkani-Tur D. and Riccardi G. Extending Boosting For Call Classification Using Word Confusion Networks (Conference) 2004. (BibTeX | Tags: Machine Learning, Signal Annotation and Interpretation) Hakkani-Tur D., Tur G., Rahim M. and Riccardi G. Unsupervised and Active Learning in Automatic Speech Recognition for Call Classification (Conference) 2004. (BibTeX | Tags: Language Modeling, Machine Learning, Speech Processing) Riccardi G. and Hakkani-Tur D. Active and Unsupervised Learning for Automatic Speech Recognition (Conference) 2003. (BibTeX | Tags: Machine Learning) Bechet F., Riccardi G. and Hakkani-Tur D. Multi-channel Sentence Classification for Spoken Dialogue Modeling (Conference) 2003. (BibTeX | Tags: Machine Learning, Speech Processing) Fabbrizio G., Dutton D., Gupta N., Hollister B., Rahim M., Riccardi G., Schapire R. and Schroeter J. AT&T Help Desk (Conference) 2002. (BibTeX | Tags: Conversational and Interactive Systems , Machine Learning, Speech Processing) Rochery M., Schapire R., Rahim M., Gupta G., Riccardi G., Bangalore S., Alshawi H. and Douglas S. Combining Prior Knowledge and Boosting for Call Classification in Spoken Language Dialogue (Conference) 2002. (BibTeX | Tags: Machine Learning, Signal Annotation and Interpretation, Speech Processing) Tur D., Riccardi G. and Gorin A. L.
Active Learning for Automatic Speech Recognition (Conference) 2002. (BibTeX | Tags: Machine Learning) Gokhan T., Wright J., Gorin A., Riccardi G., Tur H. Improving Spoken Language Understanding Using Word Confusion Networks (Conference) 2002. (BibTeX | Tags: Machine Learning, Signal Annotation and Interpretation, Speech Processing) Rahim M., Riccardi G., Saul L., Wright J., Buntschuh B. and Gorin A. L. Robust Numeric Recognition in Spoken Language Dialogue (Article) Speech Communication, 34, pp. 195-212, 2001, 2001. (Abstract | Links | BibTeX | Tags: Machine Learning, Signal Annotation and Interpretation, Speech Processing) Gretter R. and Riccardi G. On-line learning of language models with word error probability distributions (Conference) 2001. (BibTeX | Tags: Machine Learning, Speech Processing) Rose R. C., Yao H., Riccardi G. and Wright J. H. Speech Communication, 34, pp. 321-331, 2001, 2001. (Abstract | Links | BibTeX | Tags: Machine Learning, Signal Annotation and Interpretation, Speech Processing) Gorin A. L., Wright J. H., Riccardi G., Abella A. and Alonso T. Semantic information processing of spoken language (Conference) 2000. (BibTeX | Tags: Machine Learning, Signal Annotation and Interpretation, Speech Processing) Petrovska-Delacretaz D., Gorin A. L., Riccardi G. and Wright J. H. Detecting Acoustic Morphemes in Lattices for Spoken Language Understanding (Conference) 2000. (BibTeX | Tags: Machine Learning, Signal Annotation and Interpretation, Speech Processing) Riccardi G. On-line Learning of Acoustic and Lexical Units for Domain-Independent ASR (Conference) 2000. (BibTeX | Tags: Machine Learning) Potamianos A., Riccardi G. and Narayanan S. Categorical understanding using statistical N-gram models (Conference) 1999. (BibTeX | Tags: Machine Learning, Signal Annotation and Interpretation, Speech Processing) Conkie A., Riccardi G. and Rose R. C. Prosody recognition from speech utterances using acoustic and linguistic based models of prosodic events (Conference) 1999. (BibTeX | Tags: Machine Learning, Speech Processing) Gorin A. L., Petrovska-Delacretaz D., Riccardi G. and Wright J. H. Learning spoken language without transcription (Conference) 1999. (BibTeX | Tags: Machine Learning) Arai K., Wright J. H., Riccardi G. and Gorin A. L. Grammar fragment acquisition using syntactic and semantic clustering (Conference) 1998. (BibTeX | Tags: Machine Learning) Rose R. C., Yao H., Riccardi G. and Wright J. Integrating multiple knowledge sources for utterance verification in a large vocabulary speech understanding system (Conference) 1997. (BibTeX | Tags: Machine Learning, Signal Annotation and Interpretation, Speech Processing) Bocchieri E., Levin E., Pieraccini R. and Riccardi G. Understanding spontaneous speech (Article) J. of the Italian Assoc. of Artificial Intelligence, Sept. 1995, 1995. (BibTeX | Tags: Machine Learning, Signal Annotation and Interpretation, Speech Processing)2022
title = {An Unsupervised Approach to Extract Life-Events from Personal Narratives in the Mental Health Domain},
author = {Mousavi M., Negro R. and Riccardi G.},
year = {2022},
date = {2022-01-27},
publisher = {Eighth Italian Conference on Computational Linguistics},
keywords = {Machine Learning, Signal Annotation and Interpretation}
}
2021
title = {Detecting Emotion Carriers By Combining Acoustic and Lexical Representations},
author = {Bayerl P. S., Tammewar A., Riedhammer K. and Riccardi G.},
year = {2021},
date = {2021-10-01},
publisher = {IEEE Automatic Speech Recognition and Understanding Conference},
keywords = {Affective Computing, Machine Learning}
}
title = {Interpretable SincNet-based Deep Learning for Emotion Recognition in Individuals with Autism},
author = {Torres M. J., Ravanelli M., Medina-Devilliers S., Lerner D. M. and Riccardi G.},
url = {https://arxiv.org/pdf/2107.10790.pdf},
year = {2021},
date = {2021-07-18},
publisher = {IEEE Conf. Engineering in Medicine and Biology, Conference},
keywords = {Affective Computing, Autism, Machine Learning, Signal Annotation and Interpretation}
}
title = {Facial emotions are accurately encoded in the brains of those with autism: A deep learning approach},
author = {Torres M. J., Clarkson T., Hauschild K., Luhmann C. C., Lerner D. M. and Riccardi G.},
url = {https://www.sciencedirect.com/science/article/abs/pii/S2451902221001075?via%3Dihub},
year = {2021},
date = {2021-04-16},
journal = {Biological Psychiatry: Cognitive Neuroscience and Neuroimaging},
keywords = {Affective Computing, Autism, Machine Learning, Signal Annotation and Interpretation}
}
2019
title = {Active Annotation: bootstrapping annotation lexicon and guidelines for supervised NLU learning },
author = {Marinelli F., Cervone A., Tortoreto G., Stepanov E. A., Di Fabbrizio G., Riccardi G.},
editor = {INTERSPEECH, Graz},
url = {https://sisl.disi.unitn.it/wp-content/uploads/2019/11/IS19-Active_Annotation.pdf},
year = {2019},
date = {2019-01-01},
keywords = {Machine Learning, Signal Annotation and Interpretation}
}
title = {Affective Behaviour Analysis of On-line User Interactions: Are On-line Support Groups more Therapeutic than Twitter? },
author = {Tortoreto G., Stepanov E. A., Cervone A., Dubiel M., Riccardi G.},
editor = {Association for Computational Linguistics Conference, Workshop on Social Media Mining for Health Applications, Florence},
url = {https://sisl.disi.unitn.it/wp-content/uploads/2019/11/ACL19-AffectiveBehaviourOSG.pdf},
year = {2019},
date = {2019-01-01},
keywords = {Discourse, Health, Language Analytics, Machine Learning}
}
title = {Distinct but Effective Neural Networks for Facial Emotion Recognition in Individuals with Autism: A Deep Learning Approach},
author = {Mayor Torres, J.M., Clarkson, T., Luhmann, C. C., Riccardi, G., Lerner, M.D.},
url = {https://disi.unitn.it/~riccardi/papers2/INSAR_JMM_2019_Deep_Learning.pdf},
year = {2019},
date = {2019-01-01},
keywords = {Affective Computing, Autism, Machine Learning}
}
2018
title = {Depression Severity Estimation from Multiple Modalities},
author = {Stepanov E. A., Lathuilierev S., Chowdhury S. A., Ghosh A., Vieriu R.D., Sebe N. and Riccardi G.},
editor = {IEEE 20th International Conference on e-Health Networking, Applications and Services (Healthcom), Ostrava},
url = {https://sisl.disi.unitn.it/wp-content/uploads/2019/11/HealthCom18-Depression.pdf},
year = {2018},
date = {2018-01-01},
note = {EXCELLENT Paper AWARD},
keywords = {Health Analytics, Machine Learning}
}
title = {Development of an Interactive Dashboard to Analyze Cognitive Workload of Surgical Teams During Complex Procedural Care},
author = {Dias R., Conboy M. H., Gabany M. J., Clarke A. L. , Osterweil J. L., Avrunin S. G., Arney D., Goldman M. J., Riccardi G., Yule J. S., Zenati A. M.},
editor = {IEEE Conf. on Cognitive and Computational Aspects of Situation Management, Boston},
url = {https://sisl.disi.unitn.it/wp-content/uploads/2019/11/COGSIMA18ContextAwareDashboardSurgicalTeam.pdf},
year = {2018},
date = {2018-01-01},
keywords = {Health Analytics, Interactive Systems, Machine Learning, Signal Annotation and Interpretation}
}
title = {Enhanced Error Decoding from Error-Related Potentials using Convolutional Neural Networks},
author = {Mayor Torres, J.M., Clarkson, T., Stepanov E. A. , Luhmann C. C., Lerner, M.D., Riccardi, G.},
editor = {IEEE Conf. Engineering in Biology and Medicine Society, Honolulu},
url = {https://sisl.disi.unitn.it/wp-content/uploads/2019/11/EMBC18-Enhanced-Error-Decoding.pdf},
year = {2018},
date = {2018-01-01},
keywords = {Affective Computing, Autism, Machine Learning}
}
title = {Intelligent Interruption Management System to Enhance Safety and Performance in Complex Surgical and Robotic Procedures},
author = {Dias RD, Conboy HM, Gabany JM, Clarke LA, Osterweil LJ, Arney D, Goldman JM, Riccardi G, Avrunin GS, Yule SJ, Zenati MA.},
editor = {Proc. Workshop on OR 2.0 , Context-Aware Operating Theaters, Computer Assisted Robotic Endoscopy, Clinical Image-Based Procedures, and Skin Image Analysis , Grenada},
year = {2018},
date = {2018-01-01},
keywords = {Interactive Systems, Machine Learning}
}
2017
title = {Roving Mind: a balancing act between open–domain and engaging dialogue systems},
author = {Cervone A., Tortoreto G., Mezza S., Gambi E. and Riccardi G},
editor = {1st Alexa Prize Conference, Las Vegas},
url = {https://sisl.disi.unitn.it/wp-content/uploads/2019/11/AMZ17Conf-RovingMIndPaper.pdf},
year = {2017},
date = {2017-01-01},
keywords = {Conversational and Interactive Systems , Interactive Systems, Machine Learning, Natural Language Processing, Speech Processing}
}
2016
title = {Predicting Brexit: Classifying Agreement is Better than Sentiment and Pollsters},
author = {Celli F., Stepanov A. E., Poesio M. and Riccardi G.},
url = {https://sisl.disi.unitn.it/wp-content/uploads/2016/11/Coling16PEOPLE-BrexitPaper.pdf},
year = {2016},
date = {2016-11-01},
publisher = {Proc. PEOPLES Workshop at COLING, Osaka 2016.},
abstract = {On June 23rd 2016, UK held the referendum which ratified the exit from the EU. While most of the traditional pollsters failed to forecast the final vote, there were online systems that hit the result with high accuracy using opinion mining techniques and big data.
Starting one month before, we collected and monitored millions of posts about the referendum from social media conversations, and exploited Natural Language Processing techniques to predict the referendum outcome. In this paper we discuss the methods used by traditional pollsters and compare it to the predictions based on different opinion mining techniques. We find that opinion mining based on agreement/disagreement classification works better than opinion mining based on polarity classification in the forecast of the referendum outcome.},
keywords = {Affective Computing, Machine Learning, Signal Annotation and Interpretation}
}
Starting one month before, we collected and monitored millions of posts about the referendum from social media conversations, and exploited Natural Language Processing techniques to predict the referendum outcome. In this paper we discuss the methods used by traditional pollsters and compare it to the predictions based on different opinion mining techniques. We find that opinion mining based on agreement/disagreement classification works better than opinion mining based on polarity classification in the forecast of the referendum outcome.2015
title = {In the mood for Sharing Contents: Emotions, personality and interaction styles in the diffusion of news},
author = {Celli F., Ghosh A., Alam F. and Riccardi G.},
url = {https://sisl.disi.unitn.it/wp-content/uploads/2015/11/IPM15-MoodSharing.pdf},
year = {2015},
date = {2015-11-01},
journal = {Information Processing and Management, Nov 2015},
abstract = {In this paper, we analyze the influence of Twitter users in sharing news articles that may affect the readers’ mood. We collected data of more than 2000 Twitter users who shared news articles from Corriere.it, a daily newspaper that provides mood metadata annotated by readers on a voluntary basis. We automatically annotated personality types and communication styles of Twitter users and analyzed the correlations between personality, communication style, Twitter metadata (such as followig and folllowers) and the type of mood associated to the articles they shared. We also run a feature selection task, to find the best predictors of positive and negative mood sharing, and a classification task. We automatically predicted positive and negative mood sharers with 61.7% F1-measure.},
keywords = {Machine Learning, Natural Language Processing, Signal Annotation and Interpretation}
}
title = {Predicting Students’ Final Exam Scores from their Course Activities},
author = {Mogessie M, Riccardi G. and Ronchetti M.},
url = {https://sisl.disi.unitn.it/wp-content/uploads/2015/11/FIE15-PredictingStudentsScores.pdf},
year = {2015},
date = {2015-10-21},
journal = {Proc. IEEE Frontiers in Education, El Paso ( USA), 2015.},
abstract = {A common approach to the problem of predicting students’ exam scores has been to base this prediction on the previous educational history of students. In this paper, we present a model that bases this prediction on students’ performance on several tasks assigned throughout the duration
of the course. In order to build our prediction model, we use data from a semi-automated peer-assessment system implemented in two undergraduate-level computer science courses, where students ask questions about topics discussed in class, answer questions from their peers, and rate answers provided by their peers. We then construct features that are used to build several multiple linear regression models. We use the Root Mean Squared Error (RMSE) of the prediction models to evaluate their performance. Our final model, which has recorded an RMSE of 2.9326 for one course and 3.4383 for another on predicting grades on a scale of 18 to 30, is built using 14 features that capture various activities of students. Our work has possible implications in the MOOC arena and in similar online course administration systems.},
keywords = {Education Analytics, Machine Learning}
}
of the course. In order to build our prediction model, we use data from a semi-automated peer-assessment system implemented in two undergraduate-level computer science courses, where students ask questions about topics discussed in class, answer questions from their peers, and rate answers provided by their peers. We then construct features that are used to build several multiple linear regression models. We use the Root Mean Squared Error (RMSE) of the prediction models to evaluate their performance. Our final model, which has recorded an RMSE of 2.9326 for one course and 3.4383 for another on predicting grades on a scale of 18 to 30, is built using 14 features that capture various activities of students. Our work has possible implications in the MOOC arena and in similar online course administration systems.
title = {The Role of Speakers and Context in Classifying Competition in Overlapping Speech},
author = {Chowdhury A,, Danieli M. and Riccardi G.},
url = {https://sisl.disi.unitn.it/wp-content/uploads/2015/11/IS15-OverlapClassification.pdf},
year = {2015},
date = {2015-09-06},
journal = {Proc. INTERSPEECH , Dresden, 2015},
abstract = {Overlapping speech is one of the most frequently occurring events in the course of human-human conversations. Understanding the dynamics of overlapping speech is crucial for conversational analysis and for modeling human-machine dialog. Overlapping speech may signal the speaker’s intention to grab the floor with a competitive vs non-competitive act. In this paper, we study the role of speakers, whether they initiate (overlapper) or not (overlappee) the overlap, and the context of the
event. The speech overlap may be explained and predicted by the dialog context, the linguistic or acoustic descriptors. Our goal is to understand whether the competitiveness of the overlap is best predicted by the overlapper, the overlappee, the context or by their combinations. For each overlap and its context we have extracted acoustic, linguistic, and psycholinguistic features and combined decisions from the best classification models. The evaluation of the classifier has been carried out over call center human-human conversations. The results show that the complete knowledge of speakers’ role and context highly contribute to the classification results when using acoustic and
psycholinguistic features. Our findings also suggest that the lexical selections of the overlapper are good indicators of speaker’s competitive or non-competitive intentions.
Index Terms: Spoken Conversation, Automatic Classification, Overlapping Speech, Discourse, Context},
keywords = {Conversational and Interactive Systems , Machine Learning, Signal Annotation and Interpretation}
}
event. The speech overlap may be explained and predicted by the dialog context, the linguistic or acoustic descriptors. Our goal is to understand whether the competitiveness of the overlap is best predicted by the overlapper, the overlappee, the context or by their combinations. For each overlap and its context we have extracted acoustic, linguistic, and psycholinguistic features and combined decisions from the best classification models. The evaluation of the classifier has been carried out over call center human-human conversations. The results show that the complete knowledge of speakers’ role and context highly contribute to the classification results when using acoustic and
psycholinguistic features. Our findings also suggest that the lexical selections of the overlapper are good indicators of speaker’s competitive or non-competitive intentions.
Index Terms: Spoken Conversation, Automatic Classification, Overlapping Speech, Discourse, Context
title = {Selection and Aggregation Techniques for Crowdsourced Semantic Annotation Task},
author = {Chowdhury A, Calvo M., Ghosh A., Stepanov A. E., Bayer A. O., Riccardi G., Garcia F. and Sanchis E.},
url = {https://sisl.disi.unitn.it/wp-content/uploads/2015/11/IS15-crowdsourcingSelectionAggregation.pdf},
year = {2015},
date = {2015-09-06},
journal = {Proc. INTERSPEECH , Dresden, 2015},
abstract = {Crowdsourcing is an accessible and cost-effective alternative to traditional methods of collecting and annotating data. The application of crowdsourcing to simple tasks has been well investigated.
However, complex tasks like semantic annotation transfer require workers to take simultaneous decisions on chunk segmentation and labeling while acquiring on-the-go domainspecific knowledge. The increased task complexity may generate low judgment agreement and/or poor performance. The goal of this paper is to cope with these crowdsourcing requirements with semantic priming and unsupervised quality control mechanisms. We aim at an automatic quality control that takes into account different levels of workers’ expertise and annotation task performance. We investigate the judgment selection and aggregation techniques on the task of cross-language semantic annotation
transfer. We propose stochastic modeling techniques to estimate the task performance of a worker on a particular judgment with respect to the whole worker group. These estimates are used for the selection of the best judgments as well as weighted consensus-based annotation aggregation. We demonstrate that the technique is useful for increasing the quality of collected annotations.
Index Terms: Crowdsourcing, Annotation, Cross-language porting},
keywords = {Machine Learning, Signal Annotation and Interpretation, Statistical Machine Translation}
}
However, complex tasks like semantic annotation transfer require workers to take simultaneous decisions on chunk segmentation and labeling while acquiring on-the-go domainspecific knowledge. The increased task complexity may generate low judgment agreement and/or poor performance. The goal of this paper is to cope with these crowdsourcing requirements with semantic priming and unsupervised quality control mechanisms. We aim at an automatic quality control that takes into account different levels of workers’ expertise and annotation task performance. We investigate the judgment selection and aggregation techniques on the task of cross-language semantic annotation
transfer. We propose stochastic modeling techniques to estimate the task performance of a worker on a particular judgment with respect to the whole worker group. These estimates are used for the selection of the best judgments as well as weighted consensus-based annotation aggregation. We demonstrate that the technique is useful for increasing the quality of collected annotations.
Index Terms: Crowdsourcing, Annotation, Cross-language porting
title = {Annotation and Prediction of Stress and Workload from Physiological and Inertial Signals},
author = {Ghosh A, Danieli M. and Riccardi G.},
url = {https://sisl.disi.unitn.it/wp-content/uploads/2015/11/EMBC15-StressMonitoringPrediction.pdf},
year = {2015},
date = {2015-08-25},
journal = {Proc. EMBC, IEEE Conf. Engineering in Biology and Medicine Society, Milan, 2015.},
abstract = { Continuous daily stress and high workload can have negative effects on individuals’ physical and mental wellbeing. It has been shown that physiological signals may support the prediction of stress and workload. However, previous research is limited by the low diversity of signals concurring to such predictive tasks and controlled experimental design. In this paper we present 1) a pipeline for continuous and real-life acquisition of physiological and inertial signals 2) a mobile agent application for on-the-go event annotation and 3) an end-to-end signal processing and classification system for
stress and workload from diverse signal streams. We study physiological signals such as Galvanic Skin Response (GSR), Skin Temperature (ST), Inter Beat Interval (IBI) and Blood Volume Pulse (BVP) collected using a non-invasive wearable device; and inertial signals collected from accelerometer and
gyroscope sensors. We combine them with subjects’ inputs (e.g. event tagging) acquired using the agent application, and their emotion regulation scores. In our experiments we explore signal combination and selection techniques for stress and workload prediction from subjects whose signals have been recorded continuously during their daily life. The end-toend classification system is described for feature extraction, signal artifact removal, and classification. We show that a combination of physiological, inertial and user event signals provides accurate prediction of stress for real-life users and signals.},
keywords = {Health Analytics, Machine Learning, Signal Annotation and Interpretation}
}
stress and workload from diverse signal streams. We study physiological signals such as Galvanic Skin Response (GSR), Skin Temperature (ST), Inter Beat Interval (IBI) and Blood Volume Pulse (BVP) collected using a non-invasive wearable device; and inertial signals collected from accelerometer and
gyroscope sensors. We combine them with subjects’ inputs (e.g. event tagging) acquired using the agent application, and their emotion regulation scores. In our experiments we explore signal combination and selection techniques for stress and workload prediction from subjects whose signals have been recorded continuously during their daily life. The end-toend classification system is described for feature extraction, signal artifact removal, and classification. We show that a combination of physiological, inertial and user event signals provides accurate prediction of stress for real-life users and signals.
title = {Detection of Essential Hypertension with Physiological Signals from Wearable Devices},
author = {Ghosh A, Mayor Torres J.M., Danieli M. and Riccardi G.},
url = {https://sisl.disi.unitn.it/wp-content/uploads/2015/11/EMBC15-HypertensionMonitoringPrediction.pdf},
year = {2015},
date = {2015-08-25},
journal = { Proc. EMBC, IEEE Conf. Engineering in Biology and Medicine Society, Milan, 2015.},
abstract = {Early detection of essential hypertension can support the prevention of cardiovascular disease, a leading cause of death. The traditional method of identification of hypertension involves periodic blood pressure measurement using brachial cuff-based measurement devices. While these devices are noninvasive, they require manual setup for each measurement and they are not suitable for continuous monitoring. Research has shown that physiological signals such as Heart Rate Variability,
which is a measure of the cardiac autonomic activity, is correlated with blood pressure. Wearable devices capable of measuring physiological signals such as Heart Rate, Galvanic Skin Response, Skin Temperature have recently become ubiquitous. However, these signals are not accurate and are prone to noise due to different artifacts. In this paper a) we present a data collection protocol for continuous non-invasive monitoring of physiological signals from wearable devices; b) we implement
signal processing techniques for signal estimation; c) we explore how the continuous monitoring of these physiological signals can be used to identify hypertensive patients; d) We conduct a pilot study with a group of normotensive and hypertensive patients to test our techniques. We show that physiological signals extracted from wearable devices can distinguish between these two groups with high accuracy.},
keywords = {Health Analytics, Machine Learning, Signal Annotation and Interpretation}
}
which is a measure of the cardiac autonomic activity, is correlated with blood pressure. Wearable devices capable of measuring physiological signals such as Heart Rate, Galvanic Skin Response, Skin Temperature have recently become ubiquitous. However, these signals are not accurate and are prone to noise due to different artifacts. In this paper a) we present a data collection protocol for continuous non-invasive monitoring of physiological signals from wearable devices; b) we implement
signal processing techniques for signal estimation; c) we explore how the continuous monitoring of these physiological signals can be used to identify hypertensive patients; d) We conduct a pilot study with a group of normotensive and hypertensive patients to test our techniques. We show that physiological signals extracted from wearable devices can distinguish between these two groups with high accuracy.
title = {The UniTN Discourse Parser in CoNLL 2015 Shared Task},
author = {Stepanov E., Bayer A. O., Riccardi G.,},
url = {https://sisl.disi.unitn.it/wp-content/uploads/2015/11/CoNLL15-UNITNDiscourseParser.pdf},
year = {2015},
date = {2015-07-30},
journal = {Proc. CoNLL, Bejiing, 2015. Runner-up Discurse Parsing Shared-Task},
abstract = {Penn Discourse Treebank style discourse parsing is a composite task of identifying discourse relations (explicit or nonexplicit), their connective and argument spans, and assigning a sense to these relations from the hierarchy of senses. In this paper we describe University of Trento parser submitted to CoNLL 2015 Shared Task on Shallow Discourse Parsing. The span detection tasks for explicit relations are cast as token-level sequence labeling.
The argument span decisions are conditioned on relations’ being intra- or intersentential.
Non-explicit relation detection and sense assignment tasks are cast as classification. In the end-to-end closedtrack evaluation, the parser ranked second with a global F-measure of 0.2184},
keywords = {Discourse, Machine Learning, Natural Language Processing}
}
The argument span decisions are conditioned on relations’ being intra- or intersentential.
Non-explicit relation detection and sense assignment tasks are cast as classification. In the end-to-end closedtrack evaluation, the parser ranked second with a global F-measure of 0.2184
title = {Open Challenges in Modelling, Analysis and Synthesis of Human Behaviour in Human–Human and Human–Machine Interactions},
author = {Vinciarelli A., Esposito A., Andre’ E., Bonin F., Chetouani M., Cohn F. J., Cristani M., Fuhrmann F., Gilmartin E., Hammal Z., Heylen D., Kaiser R., Koutsombogera M., Potamianos A., Renals S., Riccardi G., Salah A. G.},
url = {https://sisl.disi.unitn.it/wp-content/uploads/2015/11/CogniComp15-ChallengesHHHM-Review.pdf},
year = {2015},
date = {2015-04-01},
journal = {Cognitive Computation, pp. 1-17, April 2015},
abstract = {Modelling, analysis and synthesis of behaviour are the subject of major efforts in computing science,
especially when it comes to technologies that make sense of human–human and human–machine interactions. This article outlines some of the most important issues that still need to be addressed to ensure substantial progress in the field, namely (1) development and adoption of virtuous data collection and sharing practices, (2) shift in the focus of interest from individuals to dyads and groups, (3) endowment of artificial agents with internal representations of users and context, (4) modelling of cognitive and semantic processes underlying social behaviour and (5) identification of application domains and strategies for moving from laboratory to the real-world products.},
keywords = {Conversational and Interactive Systems , Machine Learning, Speech Processing}
}
especially when it comes to technologies that make sense of human–human and human–machine interactions. This article outlines some of the most important issues that still need to be addressed to ensure substantial progress in the field, namely (1) development and adoption of virtuous data collection and sharing practices, (2) shift in the focus of interest from individuals to dyads and groups, (3) endowment of artificial agents with internal representations of users and context, (4) modelling of cognitive and semantic processes underlying social behaviour and (5) identification of application domains and strategies for moving from laboratory to the real-world products.
title = {Systems Medicine in Oncology: Signaling-networks modeling and new generation decision-support systems},
author = {Parodi S, Riccardi G, Castagnino N, Tortolina L, Maffei M, Zoppoli G, Nencioni A, Ballestrero A, Patrone F.},
year = {2015},
date = {2015-01-01},
publisher = {Methods Molecular Biology, Vol. 1386, Schmitz U and Wolkenhauer O (Eds): Systems Medicine, Springer Science press.},
keywords = {Machine Learning, Signal Annotation and Interpretation}
}
2013
title = {Instance-Based On-Line Language Model Adaptation},
author = {Bayer A. O. and Riccardi G.},
url = {https://sisl.disi.unitn.it/wp-content/uploads/2014/11/IS13-InstanceBaseLearning.pdf},
year = {2013},
date = {2013-01-01},
journal = {INTERSPEECH, Lyon, 2013},
abstract = {Language model (LM) adaptation is needed to improve the performance of language-based interaction systems. There are two important issues regarding LM adaptation; the selection of the target data set and the mathematical adaptation model. In the literature, usually statistics are drawn from the target data set (e.g. cache model) to augment (e.g. linearly) background statistical language models, as in the case of automatic speech recognition (ASR). Such models are relatively inexpensive to train, however they do not provide the necessary high-dimensional language context description needed for language-based interaction. Instance-based learning provides high-dimensional description of the lexical, semantic, or dialog context. In this paper, we present an instance-based approach to LM adaptation. We show that by retrieving similar instances from the training data and adapting the model with these instances, we can improve the performance of LMs. We propose two different similarity metrics for instance retrieval, edit distance and n-gram match score. We have performed instance-based adaptation on feed forward neural network LMs (NNLMs) to re-score n-best lists for ASR on the LUNA corpus, which includes conversational speech. We have achieved significant improvements in word error rate (WER) by using instance-based on-line LM adaptation on feed forward NNLMs.},
keywords = {Machine Learning, Speech Processing}
}
title = {Motivational Feedback in Crowdsourcing: a Case Study in Speech Transcriptions},
author = {Riccardi G., Ghosh A., Chowdhury S. A. and Bayer A. O.},
url = {https://sisl.disi.unitn.it/wp-content/uploads/2014/11/IS13-Crowdsourcing.pdf},
year = {2013},
date = {2013-01-01},
journal = {INTERSPEECH, Lyon, 2013},
abstract = {A widely used strategy in human and machine performance enhancement is achieved through feedback. In this paper we investigate the effect of live motivational feedback on motivating crowds and improving the performance of the crowdsourcing computational model. The provided feedback allows workers to react in real-time and review past actions (e.g. word deletions); thus, to improve their performance on the current and future (sub) tasks. The feedback signal can be controlled via clean (e.g. expert) supervision or noisy supervision in order to trade-off between cost and target performance of the crowd-sourced task. The feedback signal is designed to enable crowd workers to improve their performance at the (sub) task level. The type and performance of feedback signal is evaluated in the context of a speech transcription task. Amazon Mechanical Turk (AMT) platform is used to transcribe speech utterances from different corpora. We show that in both clean (expert) and noisy (worker/turker) real-time feedback conditions the crowd workers are able to provide significantly more accurate transcriptions in a shorter time.},
keywords = {Affective Computing, Conversational and Interactive Systems , Machine Learning, Signal Annotation and Interpretation}
}
title = {On-line Adaptation of Semantic Models for Spoken Language Understanding},
author = {Bayer A. O. and Riccardi G.},
url = {https://sisl.disi.unitn.it/wp-content/uploads/2014/11/ASRU13-OnLineSLUAdapt.pdf},
year = {2013},
date = {2013-01-01},
journal = {IEEE Workshop on Automatic Speech Recognition and Understanding, Olomouc, 2013},
abstract = {Spoken language understanding (SLU) systems extract semantic information from speech signals, which is usually mapped onto concept sequences. The distribution of concepts in dialogues are usually sparse. Therefore, general models may fail to model the concept distribution for a dialogue and semantic models can benefit from adaptation. In this paper, we present an instance-based approach for on-line adaptation of semantic models. We show that we can improve the performance of an SLU system on an utterance, by retrieving relevant instances from the training data and using them for on-line adapting the semantic models. The instancebased adaptation scheme uses two different similarity metrics edit distance and n-gram match score on three different tokenizations; word-concept pairs, words, and concepts. We have achieved a significant improvement (6% relative) in the understanding performance by conducting re-scoring experiments on the n-best lists that the SLU outputs. We have also applied a two-level adaptation scheme, where adaptation is first applied to the automatic speech recognizer (ASR) and then to the SLU.},
keywords = {Machine Learning, Speech Processing}
}
2012
title = {Improving the Recall of a Discourse Parser by Constraint-Based Postprocessing},
author = {Ghosh S., Johansson R., Riccardi G. and Tonelli S.},
year = {2012},
date = {2012-01-01},
journal = {LREC Istanbul, 2012},
keywords = {Discourse, Machine Learning, Natural Language Processing}
}
title = {Global Features for Shallow Discourse Parsing},
author = {Ghosh S., Riccardi G. and Johansson R.},
url = {https://sisl.disi.unitn.it/wp-content/uploads/2014/11/SigDial12-DParsing.pdf},
year = {2012},
date = {2012-01-01},
journal = {SIGDial, Seoul, 2012},
abstract = {A coherently related group of sentences may be referred to as a discourse. In this paper we address the problem of parsing coherence relations as defined in the Penn Discourse Tree Bank (PDTB). A good model for discourse structure analysis needs to account both for local dependencies at the token-level and for global dependencies and statistics. We present techniques on using inter-sentential or sentence-level (global), data-driven, nongrammatical features in the task of parsing discourse. The parser model follows up previous approach based on using tokenlevel (local) features with conditional random fields for shallow discourse parsing, which is lacking in structural knowledge of discourse. The parser adopts a twostage approach where first the local constraints are applied and then global constraints are used on a reduced weighted search space (n-best). In the latter stage we experiment with different rerankers trained on the first stage n-best parses, which are generated using lexico-syntactic local features. The two-stage parser yields significant improvements over the best performing model of discourse parser on the PDTB corpus.},
keywords = {Machine Learning, Natural Language Processing}
}
title = {Joint Language Models for Automatic Speech Recognition and Understanding},
author = {Bayer A. O. and Riccardi G.},
url = {https://sisl.disi.unitn.it/wp-content/uploads/2014/11/SLT12-NNLMSLU.pdf},
year = {2012},
date = {2012-01-01},
journal = {IEEE/ACL Workshop on Spoken Language Technology, Miami, 2012},
abstract = {Language models (LMs) are one of the main knowledge sources used by automatic speech recognition (ASR) and Spoken Language Understanding (SLU) systems. In ASR systems they are optimized to decode words from speech for a transcription task. In SLU systems they are optimized to map words into concept constructs or interpretation representations. Performance optimization is generally designed independently for ASR and SLU models in terms of word accuracy and concept accuracy respectively. However, the best word accuracy performance does not always yield the best understanding performance. In this paper we investigate how LMs originally trained to maximize word accuracy can be parametrized to account for speech understanding constraints and maximize concept accuracy. Incremental reduction in concept error rate is observed when a LM is trained on word-to-concept mappings. We show how to optimize the joint transcription and understanding task performance in the lexical-semantic relation space.},
keywords = {Machine Learning, Speech Processing}
}
title = {Combining Machine Translation Systems for Spoken Language Understanding Portability},
author = {Garcia F., Hurtado L. F., Segarra E., Sanchis E. and Riccardi G.},
url = {https://sisl.disi.unitn.it/wp-content/uploads/2014/11/SLT12-MTPortSLU.pdf},
year = {2012},
date = {2012-01-01},
journal = {IEEE/ACL Workshop on Spoken Language Technology, Miami, 2012},
abstract = {We are interested in the problem of learning Spoken Language Understanding (SLU) models for multiple target languages. Learning such models requires annotated corpora, and porting to different languages would require corpora with parallel text translation and semantic annotations. In this paper we investigate how to learn a SLU model in a target language starting from no target text and no semantic annotation. Our proposed algorithm is based on the idea of exploiting the diversity (with regard to performance and coverage) of multiple translation systems to transfer statistically stable word-toconcept mappings in the case of the romance language pair, French and Spanish. Each translation system performs differently at the lexical level (wrt BLEU). The best translation system performances for the semantic task are gained from their combination at different stages of the portability methodology. We have evaluated the portability algorithms on the French MEDIA corpus, using French as the source language and Spanish as the target language. The experiments show the effectiveness of the proposed methods with respect to the source language SLU baseline.},
keywords = {Machine Learning, Signal Annotation and Interpretation, Speech Processing, Statistical Machine Translation}
}
2011
title = {End-to-End Discourse Parser Evaluation},
author = {Ghosh S., Tonelli S., Riccardi G. and Johansson R.},
url = {https://ieeexplore.ieee.org/xpl/login.jsp?tp=&arnumber=6061347&url=http%3A%2F%2Fieeexplore.ieee.org%2Fxpls%2Fabs_all.jsp%3Farnumber%3D6061347},
year = {2011},
date = {2011-01-01},
journal = {IEEE International Conference on Semantic Computing, Menlo Park, USA, 2011},
abstract = {We are interested in the problem of discourse parsing of textual documents. We present a novel end-to-end discourse parser that, given a plain text document in input, identifies the discourse relations in the text, assigns them a semantic label and detects discourse arguments spans. The parsing architecture is based on a cascade of decisions supported by Conditional Random Fields (CRF). We train and evaluate three different parsers using the PDTB corpus. The three system versions are compared to evaluate their robustness with respect to deep/shallow and automatically extracted syntactic features.},
keywords = {Discourse, Machine Learning, Natural Language Processing, Signal Annotation and Interpretation}
}
title = {Using Syntacting and Semantic Structural Kernels for Classifying Definition Questions in Jeopardy!},
author = {Moschitti A., Chu-Carroll J., Patwardhan S., Fan J. and Riccardi G.},
url = {https://sisl.disi.unitn.it/wp-content/uploads/2014/11/EMNLP11-QuesClassJeopardy.pdf},
year = {2011},
date = {2011-01-01},
journal = {EMNLP, Edinburgh, 2011},
abstract = {The last decade has seen many interesting applications of Question Answering (QA) technology. The Jeopardy! quiz show is certainly one of the most fascinating, from the viewpoints of both its broad domain and the complexity of its language. In this paper, we study kernel methods applied to syntactic/semantic structures for accurate classification of Jeopardy! definition questions. Our extensive empirical analysis shows that our classification models largely improve on classifiers based on word-language models. Such classifiers are also used in the state-of-the-art QA pipeline constituting Watson, the IBM Jeopardy! system. Our experiments measuring their impact on Watson show enhancements in QA accuracy and a consequent increase in the amount of money earned in game-based evaluation.},
keywords = {Conversational and Interactive Systems , Machine Learning, Natural Language Processing, Signal Annotation and Interpretation}
}
title = {Detecting General Opinions from Customer Surveys},
author = {Stepanov E. and Riccardi G.},
url = {https://sisl.disi.unitn.it/wp-content/uploads/2014/11/ICDM11-OpinionDetection.pdf},
year = {2011},
date = {2011-01-01},
journal = {IEEE International Conference on Data Mining, SENTIRE Workshop, Vancouver, 2011},
abstract = {Questionnaire-based surveys and on-line product reviews resemble each other in that they both have user comments and satisfaction ratings. Since a comment might be a general opinion about a product or only one or a set of its attributes, in which case the text might not reflect the rating; surveys and reviews share the problem of pairing freetext comments with these ratings. To train accurate models for automatic evaluation of products from free-text, it is important to distinguish these two kinds of opinions. In this paper we present experiments on detecting general opinions that target a product as a whole; thus, reflect the user sentiments better. The task is different from subjectivity detection, since the goal is to detect generality of an opinion regardless of the rest of the documents being opinionated or not. The task complements feature-based opinion analysis and opinion polarity classification, since it can be applied as a preceding step to both tasks. We show that when used as a classification feature user ratings are not useful in the general opinion detection task. However, they are effective in predicting the polarity of a comment once it is identified as a general opinion.},
keywords = {Conversational and Interactive Systems , Machine Learning, Natural Language Processing}
}
2010
title = {Kernel-based Reranking for Named-Entity Extraction},
author = {Nguyen T. T., Moschitti A. and Riccardi G.},
url = {https://sisl.disi.unitn.it/wp-content/uploads/2014/11/Coling10-KernelNE.pdf},
year = {2010},
date = {2010-01-01},
journal = {COLING, Bejing, 2010},
abstract = {We present novel kernels based on structured and unstructured features for reranking the N-best hypotheses of conditional random fields (CRFs) applied to entity extraction. The former features are generated by a polynomial kernel encoding entity features whereas tree kernels are used to model dependencies amongst tagged candidate examples. The experiments on two standard corpora in two languages, i.e. the Italian EVALITA 2009 and the English CoNLL 2003 datasets, show a large improvement on CRFs in F-measure, i.e. from 80.34% to 84.33% and from 84.86% to 88.16%, respectively. Our analysis reveals that both kernels provide a comparable improvement over the CRFs baseline. Additionally, their combination improves CRFs much more than the sum of the individual contributions, suggesting an interesting kernel synergy.},
keywords = {Machine Learning, Natural Language Processing, Signal Annotation and Interpretation}
}
2009
title = {Convolution Kernels on Constituent, Dependency and Sequential Structures for Relation Extraction},
author = {Nguyen T. T., Moschitti A. and Riccardi G.},
url = {https://sisl.disi.unitn.it/wp-content/uploads/2014/11/EMNLP09-ConKernel.pdf},
year = {2009},
date = {2009-01-01},
journal = {EMNLP, Singapore, 2009},
abstract = {This paper explores the use of innovative kernels based on syntactic and semantic structures for a target relation extraction task. Syntax is derived from constituent and dependency parse trees whereas semantics concerns to entity types and lexical sequences. We investigate the effectiveness of such representations in the automated relation extraction from texts. We process the above data by means of Support Vector Machines along with the syntactic tree, the partial tree and the word sequence kernels. Our study on the ACE 2004 corpus illustrates that the combination of the above kernels achieves high effectiveness and significantly improves the current state-of-the-art.},
keywords = {Machine Learning, Natural Language Processing}
}
title = {The Multisite 2009 EVALITA Spoken Dialog System Evaluation},
author = {Baggia P., Cutugno F., Danieli M., Pieraccini R.},
url = {https://sisl.disi.unitn.it/wp-content/uploads/2014/11/EVALITA09-MultisiteSDSEvaluation.pdf},
year = {2009},
date = {2009-01-01},
journal = {AI*IA EVALITA Workshop, Brescia, 2009},
abstract = {This document presents the coordination and the evaluation procedures for the Spoken Dialogue System Task in EVALITA 2009. Three institutions participated into the competition, University of Trento, University of Naples and Loquendo. EVALITA participants were asked to develop a SDS application operating in the sales force domain, they were provided with a preliminary list of scenarios indicating system accounting modalities and a possible list of subtasks that should made possible. The three systems were hosted on a server at Trento, 19 volunteers called all of them. The calls have been recorded, transcribed and annotated. The evaluation work, based on scripting run on the annotations, has been mainly focused on assessing performance at the dialogue, task, and concept levels. Detailed results indicating the systems performances are reported in the paper. This document presents the coordination and the evaluation procedures for the Spoken Dialogue System Task in EVALITA 2009. Three institutions participated into the competition, University of Trento, University of Naples and Loquendo SpA.},
keywords = {Machine Learning, Natural Language Processing, Signal Annotation and Interpretation}
}
title = {Standoff Coordination for Multi-Tool Annotation in a Dialogue Corpus},
author = {Rodríguez K. J., Dipper S., Götze M., Poesio P., Riccardi G., Raymond C., Wisniewska J.},
year = {2009},
date = {2009-01-01},
journal = {ACL LAW Workshop, Prague, 2007},
keywords = {Machine Learning, Natural Language Processing, Signal Annotation and Interpretation}
}
title = {Ontology-Based Grounding of Spoken Language Understanding},
author = {Quarteroni S., Dinarelli M. and Riccardi G.},
url = {https://sisl.disi.unitn.it/wp-content/uploads/2014/11/ASRU09-OntologyGrounding.pdf},
year = {2009},
date = {2009-01-01},
journal = {IEEE Workshop on Automatic Speech Recognition and Understanding, Merano, 2009},
abstract = {Current Spoken Language Understanding models rely on either hand-written semantic grammars or flat attributevalue sequence labeling. In most cases, no relations between concepts are modeled, and both concepts and relations are domain-specific, making it difficult to expand or port the domain model. In contrast, we expand our previous work on a domain model based on an ontology where concepts follow the predicateargument semantics and domain-independent classical relations are defined on such concepts. We conduct a thorough study on a spoken dialog corpus collected within a customer care problemsolving domain, and we evaluate the coverage and impact of the ontology for the interpretation, grounding and},
keywords = {Machine Learning, Natural Language Processing, Speech Processing}
}
title = {Re-Ranking Models For Spoken Language Understanding},
author = {Dinarelli M., Moschitti A. and Riccardi G.},
url = {https://sisl.disi.unitn.it/wp-content/uploads/2014/11/EACL09-RR.pdf},
year = {2009},
date = {2009-01-01},
journal = {EACL Conference, Athens, 2009},
abstract = {Spoken Language Understanding aims at mapping a natural language spoken sentence into a semantic representation. In the last decade two main approaches have been pursued: generative and discriminative models. The former is more robust to overfitting whereas the latter is more robust to many irrelevant features. Additionally, the way in which these approaches encode prior knowledge is very different and their relative performance changes based on the task. In this paper we describe a machine learning framework where both models are used: a generative model produces a list of ranked hypotheses whereas a discriminative model based on structure kernels and Support Vector Machines, re-ranks such list. We tested our approach on the MEDIA corpus (human-machine dialogs) and on a new corpus (human-machine and humanhuman dialogs) produced in the European LUNA project. The results show a large improvement on the state-of-the-art in concept segmentation and labeling.},
keywords = {Machine Learning, Natural Language Processing, Speech Processing}
}
title = {Concept Segmentation and Labeling for Conversational Speech},
author = {Dinarelli M., Moschitti A. and Riccardi G.},
url = {https://sisl.disi.unitn.it/wp-content/uploads/2014/11/IS09-RR.pdf},
year = {2009},
date = {2009-01-01},
journal = {INTERSPEECH, Brighton, 2009},
abstract = {Spoken Language Understanding performs automatic concept labeling and segmentation of speech utterances. For this task, many approaches have been proposed based on both generative and discriminative models. While all these methods have shown remarkable accuracy on manual transcription of spoken utterances, robustness to noisy automatic transcription is still an open issue. In this paper we study algorithms for Spoken Language Understanding combining complementary learning models: Stochastic Finite State Transducers produce a list of hypotheses, which are re-ranked using a discriminative algorithm based on kernel methods. Our experiments on two different spoken dialog corpora, MEDIA and LUNA, show that the combined generative-discriminative model reaches the state-ofthe-art such as Conditional Random Fields (CRF) on manual transcriptions, and it is robust to noisy automatic transcriptions, outperforming, in some cases, the state-of-the-art.},
keywords = {Machine Learning, Natural Language Processing, Speech Processing}
}
title = {The Exploration/Exploitation Trade-Off in Reinforcement Learning for Dialogue Management},
author = {Varges S., Riccardi G., Quarteroni S., Ivanov A. V.},
url = {https://sisl.disi.unitn.it/wp-content/uploads/2014/11/ASRU09-ExploitationExplorationTradeoffSDS.pdf},
year = {2009},
date = {2009-01-01},
journal = {IEEE Workshop on Automatic Speech Recognition and Understanding, Merano, 2009.},
abstract = {Conversational systems use deterministic rules that trigger actions such as requests for confirmation or clarification. More recently, Reinforcement Learning and (Partially Observable) Markov Decision Processes have been proposed for this task. In this paper, we investigate action selection strategies for dialogue management, in particular the exploration/exploitation trade-off and its impact on final reward (i.e. the session reward after optimization has ended) and lifetime reward (i.e. the overall reward accumulated over the learner’s lifetime). We propose to use interleaved exploitation sessions as a learning methodology to assess the reward obtained from the current policy. The experiments show a statistically significant difference in final reward of exploitation-only sessions between a system that optimizes lifetime reward and one that maximizes the reward of the final policy.},
keywords = {Conversational and Interactive Systems , Machine Learning, Speech Processing}
}
title = {Leveraging POMDPs trained with User Simulations and Rule-Based Dialog Management in a SDS},
author = {Varges S., Riccardi G., Quarteroni S., Ivanov A. V. and Roberti P.},
url = {https://sisl.disi.unitn.it/wp-content/uploads/2014/11/SIGDial09-demo.pdf},
year = {2009},
date = {2009-01-01},
journal = {SIGDIAL, Demo Session, London, 2009},
abstract = {We have developed a complete spoken dialogue framework that includes rule-based and trainable dialogue managers, speech recognition, spoken language understanding and generation modules, and a comprehensive web visualization interface. We present a spoken dialogue system based on Reinforcement Learning that goes beyond standard rule based models and computes on-line decisions of the best dialogue moves. Bridging the gap between handcrafted (e.g. rule-based) and adaptive (e.g. based on Partially Observable Markov Decision Processes - POMDP) dialogue models, this prototype is able to learn high rewarding policies in a number of dialogue situations.},
keywords = {Conversational and Interactive Systems , Machine Learning, Speech Processing}
}
title = {Learning the Structure of Human-Computer and Human-Human Spoken Conversations},
author = {Griol D., Riccardi G. and Sanchis E.},
url = {https://sisl.disi.unitn.it/wp-content/uploads/2014/11/IS09-HHConvStructure.pdf},
year = {2009},
date = {2009-01-01},
journal = {INTERSPEECH, Brighton, 2009},
abstract = {We are interested in the problem of understanding human conversation structure in the context of human-machine and human-human interaction. We present a statistical methodology for detecting the structure of spoken dialogs based on a generative model learned using decision trees. To evaluate our approach we have used the LUNA corpora, collected from real users engaged in problem solving tasks. The results of the evaluation show that automatic segmentation of spoken dialogs is very effective not only with models built using separately human-machine dialogs or human-human dialogs, but it is also possible to infer the task-related structure of human-human dialogs with a model learned using only human-machine dialogs.},
keywords = {Conversational and Interactive Systems , Machine Learning, Speech Processing}
}
title = {Can Machines Call People?- User Experience While Answering Telephone Calls Initiated by Machine},
author = {Sporka A. J., Jakub F. and Riccardi G.},
url = {https://sisl.disi.unitn.it/wp-content/uploads/2014/11/IS09-RR1.pdf},
year = {2009},
date = {2009-01-01},
journal = {CHI, Boston, 2009},
abstract = {Spoken Language Understanding performs automatic concept labeling and segmentation of speech utterances. For this task, many approaches have been proposed based on both generative and discriminative models. While all these methods have shown remarkable accuracy on manual transcription of spoken utterances, robustness to noisy automatic transcription is still an open issue. In this paper we study algorithms for Spoken Language Understanding combining complementary learning models: Stochastic Finite State Transducers produce a list of hypotheses, which are re-ranked using a discriminative algorithm based on kernel methods. Our experiments on two different spoken dialog corpora, MEDIA and LUNA, show that the combined generative-discriminative model reaches the state-ofthe-art such as Conditional Random Fields (CRF) on manual transcriptions, and it is robust to noisy automatic transcriptions, outperforming, in some cases, the state-of-the-art.},
keywords = {Conversational and Interactive Systems , Machine Learning, Speech Processing}
}
title = {On-Line Strategy Computation in Spoken Dialog Systems},
author = {Varges S., Riccardi G., Quarteroni S., Ivanov A. V. and Roberti P.},
url = {https://sisl.disi.unitn.it/wp-content/uploads/2014/11/ICASSP09-POMDPs.pdf},
year = {2009},
date = {2009-01-01},
journal = {ICASSP, Demo Session, Singapore, 2009. VIDEO},
keywords = {Conversational and Interactive Systems , Machine Learning, Speech Processing}
}
title = {The Voice Multimodal Application Framework},
author = {Riccardi G., Mosca N., Roberti P. and Baggia P.},
year = {2009},
date = {2009-01-01},
journal = {AVIOS, San Diego, 2009. VIDEO},
keywords = {Conversational and Interactive Systems , Machine Learning, Speech Processing}
}
title = {Spoken Dialog Systems: From Theory to Technology},
author = {Riccardi G., Baggia P. and Roberti P.},
year = {2009},
date = {2009-01-01},
journal = {Proc. Work. Toni Mian, Padua, 2007},
keywords = {Conversational and Interactive Systems , Machine Learning, Speech Processing}
}
title = {What's in an Ontology for Spoken Language Understanding},
author = {Quarteroni S., Riccardi G. and Dinarelli M.},
url = {https://sisl.disi.unitn.it/wp-content/uploads/2014/11/IS09-Ontology.pdf},
year = {2009},
date = {2009-01-01},
journal = {INTERSPEECH, Brighton, 2009},
abstract = {Current Spoken Language Understanding systems rely either on hand-written semantic grammars or on flat attribute-value sequence labeling. In both approaches, concepts and their relations (when modeled at all) are domain-specific, thus making it difficult to expand or port the domain model. To address this issue, we introduce: 1) a domain model based on an ontology where concepts are classified into either predicative or argumentative; 2) the modeling of relations between such concept classes in terms of classical relations as defined in lexical semantics. We study and analyze our approach on a corpus of customer care data, where we evaluate the coverage and relevance of the ontology for the interpretation of speech utterances.},
keywords = {Machine Learning, Natural Language Processing, Speech Processing}
}
title = {Re-Ranking Models Based on Small Training Data for Spoken Language Understanding},
author = {Dinarelli M., Moschitti A. and Riccardi G.},
url = {https://sisl.disi.unitn.it/wp-content/uploads/2014/11/IS09-RR3.pdf},
year = {2009},
date = {2009-01-01},
journal = {EMNLP, Singapore, 2009},
abstract = {Spoken Language Understanding performs automatic concept labeling and segmentation of speech utterances. For this task, many approaches have been proposed based on both generative and discriminative models. While all these methods have shown remarkable accuracy on manual transcription of spoken utterances, robustness to noisy automatic transcription is still an open issue. In this paper we study algorithms for Spoken Language Understanding combining complementary learning models: Stochastic Finite State Transducers produce a list of hypotheses, which are re-ranked using a discriminative algorithm based on kernel methods. Our experiments on two different spoken dialog corpora, MEDIA and LUNA, show that the combined generative-discriminative model reaches the state-ofthe-art such as Conditional Random Fields (CRF) on manual transcriptions, and it is robust to noisy automatic transcriptions, outperforming, in some cases, the state-of-the-art.},
keywords = {Machine Learning, Natural Language Processing, Speech Processing}
}
2008
title = {Persistent Information State in a Data-Centric Architecture},
author = {Sebastian V., Riccardi G. and Quarteroni S.},
year = {2008},
date = {2008-01-01},
journal = {SIGdial Workshop on Discourse and Dialogue, Columbus, 2008},
keywords = {Conversational and Interactive Systems , Machine Learning, Speech Processing}
}
title = {Learning with Noisy Supervision for Spoken Language Understanding},
author = {Raymond C. and Riccardi G.},
url = {https://sisl.disi.unitn.it/wp-content/uploads/2014/11/ICASSP08-NoisySupervisionSLU.pdf},
year = {2008},
date = {2008-01-01},
journal = {Proc. IEEE ICASSP, Las Vegas,2008},
abstract = {Data-driven Spoken Language Understanding (SLU) systems need semantically annotated data which are expensive, time consuming and prone to human errors. Active learning has been successfully applied to automatic speech recognition and utterance classification. In general, corpora annotation for SLU involves such tasks as sentence segmentation, chunking or frame labeling and predicate-argument annotation. In such cases human annotations are subject to errors increasing with the annotation complexity. We investigate two alternative noise-robust active learning strategies that are either data-intensive or supervision-intensive. The strategies detect likely erroneous examples and improve significantly the SLU performance for a given labeling cost. We apply uncertainty based active learning with conditional random fields on the concept segmentation task for SLU. We perform annotation experiments on two databases, namely ATIS (English) and Media (French). We show that our noise-robust algorithm could improve the accuracy up to 6% (absolute) depending on the noise level and the labeling cost.},
keywords = {Machine Learning, Natural Language Processing, Speech Processing}
}
title = {Joint Generative And Discriminative Models For Spoken Language Understanding},
author = {Dinarelli M., Moschitti A., Riccardi G.},
year = {2008},
date = {2008-01-01},
journal = {IEEE/ACL Workshop on Spoken Language Technology, Goa, 2008},
keywords = {Machine Learning, Natural Language Processing, Speech Processing}
}
title = {Semantic Annotations For Conversational Speech: from speech transcriptions to predicate argument structures},
author = {Bisazza A., Dinarelli M., Quarteroni S., Tonelli S., Moschitti A., Riccardi G.},
year = {2008},
date = {2008-01-01},
journal = {IEEE/ACL Workshop on Spoken Language Technology, Goa, 2008},
keywords = {Machine Learning, Natural Language Processing}
}
title = {Automatic FrameNet-Based Annotation of Conversational Speech},
author = {Coppola B., Moschitti A., Tonelli S., Riccardi G.},
url = {https://sisl.disi.unitn.it/wp-content/uploads/2014/11/SLT08-FramenetParser.pdf},
year = {2008},
date = {2008-01-01},
journal = {IEEE/ACL Workshop on Spoken Language Technology, Goa, 2008},
abstract = {Current Spoken Language Understanding technology is based on a simple concept annotation of word sequences, where the interdependencies between concepts and their compositional semantics are neglected. This prevents an effective handling of language phenomena, with a consequential limitation on the design of more complex dialog systems. In this paper, we argue that shallow semantic representation as formulated in the Berkeley FrameNet Project may be useful to improve the capability of managing more complex dialogs. To prove this, the first step is to show that a FrameNet parser of sufficient accuracy can be designed for conversational speech. We show that exploiting a small set of FrameNetbased manual annotations, it is possible to design an effective semantic parser. Our experiments on an Italian spoken dialog corpus, created within the LUNA project, show that our approach is able to automatically annotate unseen dialog turns with a high accuracy.},
keywords = {Machine Learning, Natural Language Processing}
}
title = {Spoken Language Understanding},
author = {De Mori R., Bechet F., Hakkani-Tur D., McTear M., Riccardi G. and Tur G.},
year = {2008},
date = {2008-01-01},
journal = {IEEE Signal Processing Magazine vol. 25, pp.50-58 ,2008},
keywords = {Machine Learning, Natural Language Processing, Speech Processing}
}
title = {Active Annotation in the LUNA Italian Corpus of Spontaneous Dialogues},
author = {Rodríguez K., Raymond C. and Riccardi G.},
year = {2008},
date = {2008-01-01},
journal = {Proc. Language Resources and Evaluation (LREC), Marrakech, 2008},
keywords = {Machine Learning, Natural Language Processing, Signal Annotation and Interpretation}
}
2007
title = {Generative and Discriminative Algorithms for Spoken Language Understanding},
author = {Raymond C., Riccardi G.},
year = {2007},
date = {2007-01-01},
journal = {INTERSPEECH, Antwerp, 2007},
keywords = {Machine Learning, Natural Language Processing, Speech Processing}
}
title = {A Data-Centric Architecture for Data-Driven Spoken Dialog Systems},
author = {Varges S. and Riccardi G.},
year = {2007},
date = {2007-01-01},
journal = {IEEE Workshop on Automatic Speech Recognition and Understanding, Kyoto, 2007},
keywords = {Conversational and Interactive Systems , Machine Learning, Speech Processing}
}
title = {Spoken Language Understanding with Kernels for Syntactic/Semantic Structures},
author = {Moschitti A., Riccardi G. and Raymond C.},
url = {https://sisl.disi.unitn.it/wp-content/uploads/2014/11/ASRU07-SLUKernels.pdf},
year = {2007},
date = {2007-01-01},
journal = {IEEE Workshop on Automatic Speech Recognition and Understanding, Kyoto, 2007},
abstract = {Automatic concept segmentation and labeling are the fundamental problems of Spoken Language Understanding in dialog systems. Such tasks are usually approached by using generative or discriminative models based on n-grams. As the uncertainty or ambiguity of the spoken input to dialog system increase, we expect to need dependencies beyond n-gram statistics. In this paper, a general purpose statistical syntactic parser is used to detect syntactic/semantic dependencies between concepts in order to increase the accuracy of sentence segmentation and concept labeling. The main novelty of the approach is the use of new tree kernel functions which encode syntactic/semantic structures in discriminative learning models. We experimented with Support Vector Machines and the above kernels on the standard ATIS dataset. The proposed algorithm automatically parses natural language text with offthe-shelf statistical parser and labels the syntactic (sub)trees with concept labels. The results show that the proposed model is very accurate and competitive with respect to state-of-theart models when combined with n-gram based models.},
keywords = {Machine Learning, Natural Language Processing, Speech Processing}
}
title = {The LUNA Corpus: an Annotation Scheme for a Multi-domain Multi-lingual Dialogue Corpus},
author = {Raymond C., Riccardi G., Rodríguez K. J. and Wisniewska J.},
year = {2007},
date = {2007-01-01},
journal = {11th Workshop on the Semantics and Pragmatics of Dialogue (DECALOG'07), Rovereto, 2007},
keywords = {Machine Learning, Natural Language Processing, Signal Annotation and Interpretation}
}
2006
title = {Spoken Dialog Systems: From Theory to Technology},
author = {Riccardi G. and Baggia P.},
year = {2006},
date = {2006-01-01},
journal = {Edizione della Normale di Pisa, 2006},
keywords = {Conversational and Interactive Systems , Machine Learning, Speech Processing}
}
title = {Shallow Semantic Parsing for Spoken Language Understanding},
author = {Coppola B., Moschitti A., Riccardi G.},
url = {https://sisl.disi.unitn.it/wp-content/uploads/2014/11/NAACL09-ShallowSemanticParsingFrameNet.pdf},
year = {2006},
date = {2006-01-01},
journal = {NAACL, Boulder, Colorado, 2009},
abstract = {Most Spoken Dialog Systems are based on speech grammars and frame/slot semantics. The semantic descriptions of input utterances are usually defined ad-hoc with no ability to generalize beyond the target application domain or to learn from annotated corpora. The approach we propose in this paper exploits machine learning of frame semantics, borrowing its theoretical model from computational linguistics. While traditional automatic Semantic Role Labeling approaches on written texts may not perform as well on spoken dialogs, we show successful experiments on such porting. Hence, we design and evaluate automatic FrameNet-based parsers both for English written texts and for Italian dialog utterances. The results show that disfluencies of dialog data do not severely hurt performance. Also, a small set of FrameNet-like manual annotations is enough for realizing accurate Semantic Role Labeling on the target domains of typical Dialog Systems.},
keywords = {Machine Learning, Natural Language Processing, Speech Processing}
}
title = {The AT&T Spoken Language Understanding System},
author = {Gupta N., Tur G., Hakkani-Tur D., Bangalore S., Riccardi G. and Rahim M.},
url = {https://sisl.disi.unitn.it/wp-content/uploads/2014/11/IEEE-SAP-2005-SLU.pdf},
year = {2006},
date = {2006-01-01},
journal = {IEEE Trans. on Audio, Speech and Language Processing, volume 14, Issue 1, pp. 213-22, 2006},
abstract = {Spoken language understanding (SLU) aims at extracting meaning from natural language speech. Over the past decade, a variety of practical goal-oriented spoken dialog systems have been built for limited domains. SLU in these systems ranges from understanding predetermined phrases through fixed grammars, extracting some predefined named entities, extracting users’ intents for call classification, to combinations of users’ intents and named entities. In this paper, we present the SLU system of VoiceTone ® (a service provided by AT&T where AT&T develops, deploys and hosts spoken dialog applications for enterprise customers). The SLU system includes extracting both intents and the named entities from the users’ utterances. For intent determination, we use statistical classifiers trained from labeled data, and for named entity extraction we use rule-based fixed grammars. The focus of our work is to exploit data and to use machine learning techniques to create scalable SLU systems which can be quickly deployed for new domains with minimal human intervention. These objectives are achieved by 1) using the predicate-argument representation of semantic content of an utterance; 2) extending statistical classifiers to seamlessly integrate hand crafted classification rules with the rules learned from data; and 3) developing an active learning framework to minimize the human labeling effort for quickly building the classifier models and adapting them to changes. We present an evaluation of this system using two deployed applications of VoiceTone},
keywords = {Machine Learning, Natural Language Processing, Speech Processing}
}
title = {An Active Approach to spoken Language Processing},
author = {Hakkani-Tur D., Riccardi G. and Tur G.},
url = {https://sisl.disi.unitn.it/wp-content/uploads/2014/11/acm-tslp-06.pdf},
year = {2006},
date = {2006-01-01},
journal = {ACM Transactions on Speech and Language Processing, Vol. 3, No. 3, pp 1-31, 2006},
abstract = {State of the art data-driven speech and language processing systems require a large amount of human intervention ranging from data annotation to system prototyping. In the traditional supervised passive approach, the system is trained on a given number of annotated data samples and evaluated using a separate test set. Then more data is collected arbitrarily, annotated, and the whole cycle is repeated. In this article, we propose the active approach where the system itself selects its own training data, evaluates itself and re-trains when necessary. We first employ active learning which aims to automatically select the examples that are likely to be the most informative for a given task. We use active learning for both selecting the examples to label and the examples to re-label in order to correct labeling errors. Furthermore, the system automatically evaluates itself using active evaluation to keep track of the unexpected events and decides on-demand to label more examples. The active approach enables dynamic adaptation of spoken language processing systems to unseen or unexpected events for nonstationary input while reducing the manual annotation effort significantly. We have evaluated the active approach with the AT&T spoken dialog system used for customer care applications. In this article, we present our results for both automatic speech recognition and spoken language understanding. Categories and Subject Descriptors: I.2.7 [Artificial Intelligence]: Natural Language Processing—Speech recognition and synthesis; I.5.1 [Pattern Recognition]: Models—Statistical General Terms: Algorithms, Languages, Performance Additional Key Words and Phrases: Passive learning, active learning, adaptive learning, unsupervised learning, active evaluation, spoken language understanding, automatic speech recognition, spoken dialog systems, speech and language processing},
keywords = {Machine Learning, Natural Language Processing, Speech Processing}
}
2005
title = {Adaptive Categorical Understanding for Spoken Dialogue Systems'},
author = {Potamianos A., Narayanan S. and Riccardi G.},
url = {https://sisl.disi.unitn.it/wp-content/uploads/2014/11/ieee_adapt-categ-05.pdf},
year = {2005},
date = {2005-01-01},
journal = {Potamianos A., Narayanan S and Riccardi, G.},
abstract = {IEEE Trans. on Speech and Audio, vol. 13, n.3 , pp. 321-329, 2005},
keywords = {Conversational and Interactive Systems , Machine Learning, Speech Processing}
}
title = {Error Prediction in Spoken Dialog: from Signal-to-Noise Ratio to Semantic Confidence Scores},
author = {Hakkani-Tur D., Tur G., Riccardi G. and Kim H. K.},
year = {2005},
date = {2005-01-01},
journal = {IEEE ICASSP, Philadelphia, March 2005},
keywords = {Conversational and Interactive Systems , Machine Learning, Speech Processing}
}
title = {Mining Spoken dialogue Corpora for system Evaluation and Modeling},
author = {Bechet F., Riccardi G. and Hakkani-Tur D.},
year = {2005},
date = {2005-01-01},
journal = {EMNLP Conference, Barcelona, 2004},
keywords = {Conversational and Interactive Systems , Machine Learning, Speech Processing}
}
title = {Combining Classifiers for Spoken Language Understanding},
author = {Karahan M., Hakkani-Tur D., Riccardi G. and Tur G.},
year = {2005},
date = {2005-01-01},
journal = {IEEE ASRU, U.S. Virgin Islands, Dec., 2003},
keywords = {Machine Learning, Natural Language Processing, Speech Processing}
}
title = {Active Learning: Theory and Applications to Automatic Speech Recognition},
author = {Riccardi G. and Hakkani-Tur D.},
url = {https://sisl.disi.unitn.it/wp-content/uploads/2014/11/ieee-al-05.pdf},
year = {2005},
date = {2005-01-01},
journal = {IEEE Trans. on Speech and Audio, vol. 13, n.4 , pp. 504-511, 2005},
abstract = {We are interested in the problem of adaptive learning in the context of automatic speech recognition (ASR). In this paper, we propose an active learning algorithm for ASR. Automatic speech recognition systems are trained using human supervision to provide transcriptions of speech utterances. The goal of Active Learning is to minimize the human supervision for training acoustic and language models and to maximize the performance given the transcribed and untranscribed data. Active learning aims at reducing the number of training examples to be labeled by automatically processing the unlabeled examples, and then selecting the most informative ones with respect to a given cost function for a human to label. In this paper we describe how to estimate the confidence score for each utterance through an on-line algorithm using the lattice output of a speech recognizer. The utterance scores are filtered through the informativeness function and an optimal subset of training samples is selected. The active learning algorithm has been applied to both batch and on-line learning scheme and we have experimented with different selective sampling algorithms. Our experiments show that by using active learning the amount of labeled data needed for a given word accuracy can be reduced by more than 60% with respect to random sampling.},
keywords = {Machine Learning}
}
2004
title = {Extending Boosting For Call Classification Using Word Confusion Networks},
author = {Tur G., Hakkani-Tur D. and Riccardi G.},
year = {2004},
date = {2004-01-01},
journal = {IEEE ICASSP, Montreal, May 2004},
keywords = {Machine Learning, Signal Annotation and Interpretation}
}
title = {Unsupervised and Active Learning in Automatic Speech Recognition for Call Classification},
author = {Hakkani-Tur D., Tur G., Rahim M. and Riccardi G.},
year = {2004},
date = {2004-01-01},
journal = {ICASSP, Montreal, May 2004},
keywords = {Language Modeling, Machine Learning, Speech Processing}
}
2003
title = {Active and Unsupervised Learning for Automatic Speech Recognition},
author = {Riccardi G. and Hakkani-Tur D.},
year = {2003},
date = {2003-01-01},
journal = {EUROSPEECH, Geneve, Switzerland, Sept. 2003},
keywords = {Machine Learning}
}
title = {Multi-channel Sentence Classification for Spoken Dialogue Modeling},
author = {Bechet F., Riccardi G. and Hakkani-Tur D.},
year = {2003},
date = {2003-01-01},
journal = {EUROSPEECH, Geneve, Switzerland, Sept. 2003},
keywords = {Machine Learning, Speech Processing}
}
2002
title = {AT&T Help Desk},
author = {Fabbrizio G., Dutton D., Gupta N., Hollister B., Rahim M., Riccardi G., Schapire R. and Schroeter J.},
year = {2002},
date = {2002-01-01},
journal = {Proc. ICSLP, Denver, 2002},
keywords = {Conversational and Interactive Systems , Machine Learning, Speech Processing}
}
title = {Combining Prior Knowledge and Boosting for Call Classification in Spoken Language Dialogue},
author = {Rochery M., Schapire R., Rahim M., Gupta G., Riccardi G., Bangalore S., Alshawi H. and Douglas S.},
year = {2002},
date = {2002-01-01},
journal = {Proc. IEEE ICASSP, Orlando, 2002},
keywords = {Machine Learning, Signal Annotation and Interpretation, Speech Processing}
}
title = {Active Learning for Automatic Speech Recognition},
author = {Tur D., Riccardi G. and Gorin A. L.
},
year = {2002},
date = {2002-01-01},
journal = {Proc. IEEE ICASSP, Orlando, 2002},
keywords = {Machine Learning}
}
title = {Improving Spoken Language Understanding Using Word Confusion Networks},
author = {Gokhan T., Wright J., Gorin A., Riccardi G., Tur H.},
year = {2002},
date = {2002-01-01},
journal = {ICSLP, Denver, 2002},
keywords = {Machine Learning, Signal Annotation and Interpretation, Speech Processing}
}
2001
title = {Robust Numeric Recognition in Spoken Language Dialogue},
author = {Rahim M., Riccardi G., Saul L., Wright J., Buntschuh B. and Gorin A. L.},
url = {https://sisl.disi.unitn.it/wp-content/uploads/2014/11/numericlang-speechcomm-2001.pdf},
year = {2001},
date = {2001-11-01},
journal = {Speech Communication, 34, pp. 195-212, 2001},
abstract = {This paper addresses the problem of automatic numeric recognition and understanding in spoken language dialogue. We show that accurate numeric understanding in ̄uent unconstrained speech demands maintaining robustness at several dierent levels of system design, including acoustic, language, understanding and dialogue. We describe a robust system for numeric recognition and present algorithms for feature extraction, acoustic and language modeling, discriminative training, utterance veri®cation and numeric understanding and validation. Experimental results from a ®eld-trial of a spoken dialogue system are presented that include customers\' responses to credit card and telephone number requests. Ó 2001 Elsevier Science B.V. All rights reserved.},
keywords = {Machine Learning, Signal Annotation and Interpretation, Speech Processing}
}
title = {On-line learning of language models with word error probability distributions},
author = {Gretter R. and Riccardi G.},
year = {2001},
date = {2001-05-07},
journal = {Proc. IEEE ICASSP 2001, Salt Lake City, Utah, 7-11 May 2001},
keywords = {Machine Learning, Speech Processing}
}
title = {Integration of Utterance Verification with Statistical Language Modeling and Spoken Language Understanding},
author = {Rose R. C., Yao H., Riccardi G. and Wright J. H.},
url = {https://sisl.disi.unitn.it/wp-content/uploads/2014/11/uv-speechcomm-2001.pdf},
year = {2001},
date = {2001-01-01},
journal = {Speech Communication, 34, pp. 321-331, 2001},
abstract = {Methods for utterance veri®cation (UV) and their integration into statistical language modeling and understanding formalisms for a large vocabulary spoken understanding system are presented. The paper consists of three parts. First, a set of acoustic likelihood ratio (LR) based UV techniques are described and applied to the problem of rejecting portions of a hypothesized word string that may have been incorrectly decoded by a large vocabulary continuous speech recognizer. Second, a procedure for integrating the acoustic level con®dence measures with the statistical language model is described. Finally, the eect of integrating acoustic level con®dence into the spoken language understanding unit (SLU) in a call-type classi®cation task is discussed. These techniques were evaluated on utterances collected from a highly unconstrained call routing task performed over the telephone network. They have been evaluated in terms of their ability to classify utterances into a set of 15 call-types that are accepted by the application. Ó 2001 Elsevier Science B.V. All rights reserved.},
keywords = {Machine Learning, Signal Annotation and Interpretation, Speech Processing}
}
2000
title = {Semantic information processing of spoken language},
author = {Gorin A. L., Wright J. H., Riccardi G., Abella A. and Alonso T.},
year = {2000},
date = {2000-10-01},
journal = {ATR Workshop on Multi-Lingual Speech Communication, Oct. 2000},
keywords = {Machine Learning, Signal Annotation and Interpretation, Speech Processing}
}
title = {Detecting Acoustic Morphemes in Lattices for Spoken Language Understanding},
author = {Petrovska-Delacretaz D., Gorin A. L., Riccardi G. and Wright J. H.},
year = {2000},
date = {2000-10-01},
journal = {Proc. ICSLP, Beijing, Oct. 2000},
keywords = {Machine Learning, Signal Annotation and Interpretation, Speech Processing}
}
title = {On-line Learning of Acoustic and Lexical Units for Domain-Independent ASR},
author = {Riccardi G.},
year = {2000},
date = {2000-10-01},
journal = {Proc. ICSLP, Beijing, Oct. 2000},
keywords = {Machine Learning}
}
1999
title = {Categorical understanding using statistical N-gram models},
author = {Potamianos A., Riccardi G. and Narayanan S.},
year = {1999},
date = {1999-09-01},
journal = {Proc. EUROSPEECH, Budapest, Hungary, Sept. 1999},
keywords = {Machine Learning, Signal Annotation and Interpretation, Speech Processing}
}
title = {Prosody recognition from speech utterances using acoustic and linguistic based models of prosodic events},
author = {Conkie A., Riccardi G. and Rose R. C.},
year = {1999},
date = {1999-09-01},
journal = {Proc. EUROSPEECH, Budapest, Hungary, Sept. 1999},
keywords = {Machine Learning, Speech Processing}
}
title = {Learning spoken language without transcription},
author = {Gorin A. L., Petrovska-Delacretaz D., Riccardi G. and Wright J. H.},
year = {1999},
date = {1999-01-01},
journal = {Proc.IEEE ASRU Workshop, Colorado, 1999},
keywords = {Machine Learning}
}
1998
title = {Grammar fragment acquisition using syntactic and semantic clustering},
author = {Arai K., Wright J. H., Riccardi G. and Gorin A. L.},
year = {1998},
date = {1998-11-01},
journal = {Proc. ICSLP, Sydney, Nov. 1998},
keywords = {Machine Learning}
}
1997
title = {Integrating multiple knowledge sources for utterance verification in a large vocabulary speech understanding system},
author = {Rose R. C., Yao H., Riccardi G. and Wright J.},
year = {1997},
date = {1997-01-01},
journal = {Proc. IEEE ASR Workshop Proc., Santa Barbara, 1997},
keywords = {Machine Learning, Signal Annotation and Interpretation, Speech Processing}
}
1995
title = {Understanding spontaneous speech},
author = {Bocchieri E., Levin E., Pieraccini R. and Riccardi G.},
year = {1995},
date = {1995-01-01},
journal = {J. of the Italian Assoc. of Artificial Intelligence, Sept. 1995},
keywords = {Machine Learning, Signal Annotation and Interpretation, Speech Processing}
}