This book constitutes the thoroughly refereed post-workshop proceedings of the Third IAPR TC3 Workshop on Pattern Recognition of Social Signals in Human-Computer-Interaction, MPRSS 2014, held in Stockholm, Sweden, in August 2014, as a satellite event of the International Conference on Pattern Recognition, ICPR 2014. The 14 revised papers presented focus on pattern recognition, machine learning and information fusion methods with applications in social signal processing, including multimodal emotion recognition, user identification, and recognition of human activities.
This book constitutes the refereed post-workshop proceedings of the 5th IAPR TC9 Workshop on Pattern Recognition of Social Signals in Human-Computer-Interaction, MPRSS 2018, held in Beijing, China, in August 2018. The 10 revised papers presented in this book focus on pattern recognition, machine learning and information fusion methods with applications in social signal processing, including multimodal emotion recognition and pain intensity estimation, especially the question how to distinguish between human emotions from pain or stress induced by pain is discussed.
This book constitutes the thoroughly refereed post-workshop proceedings of the First IAPR TC3 Workshop on Pattern Recognition of Social Signals in Human-Computer-Interaction (MPRSS2012), held in Tsukuba, Japan in November 2012, in collaboration with the NLGD Festival of Games. The 21 revised papers presented during the workshop cover topics on facial expression recognition, audiovisual emotion recognition, multimodal Information fusion architectures, learning from unlabeled and partially labeled data, learning of time series, companion technologies and robotics.
This book constitutes the thoroughly refereed post-workshop proceedings of the Fourth IAPR TC9 Workshop on Pattern Recognition of Social Signals in Human-Computer-Interaction, MPRSS 2016, held in Cancun, Mexico, in December 2016. The 13 revised papers presented focus on pattern recognition, machine learning and information fusion methods with applications in social signal processing, including multimodal emotion recognition, user identification, and recognition of human activities.
The Handbook of Multimodal-Multisensor Interfaces provides the first authoritative resource on what has become the dominant paradigm for new computer interfaces: user input involving new media (speech, multi-touch, hand and body gestures, facial expressions, writing) embedded in multimodal-multisensor interfaces that often include biosignals. This edited collection is written by international experts and pioneers in the field. It provides a textbook, reference, and technology roadmap for professionals working in this and related areas. This second volume of the handbook begins with multimodal signal processing, architectures, and machine learning. It includes recent deep learning approaches for processing multisensorial and multimodal user data and interaction, as well as context-sensitivity. A further highlight is processing of information about users' states and traits, an exciting emerging capability in next-generation user interfaces. These chapters discuss real-time multimodal analysis of emotion and social signals from various modalities, and perception of affective expression by users. Further chapters discuss multimodal processing of cognitive state using behavioral and physiological signals to detect cognitive load, domain expertise, deception, and depression. This collection of chapters provides walk-through examples of system design and processing, information on tools and practical resources for developing and evaluating new systems, and terminology and tutorial support for mastering this rapidly expanding field. In the final section of this volume, experts exchange views on the timely and controversial challenge topic of multimodal deep learning. The discussion focuses on how multimodal-multisensor interfaces are most likely to advance human performance during the next decade.
Multimodal signal processing is an important research and development field that processes signals and combines information from a variety of modalities – speech, vision, language, text – which significantly enhance the understanding, modelling, and performance of human-computer interaction devices or systems enhancing human-human communication. The overarching theme of this book is the application of signal processing and statistical machine learning techniques to problems arising in this multi-disciplinary field. It describes the capabilities and limitations of current technologies, and discusses the technical challenges that must be overcome to develop efficient and user-friendly multimodal interactive systems. With contributions from the leading experts in the field, the present book should serve as a reference in multimodal signal processing for signal processing researchers, graduate students, R&D engineers, and computer engineers who are interested in this emerging field. - Presents state-of-art methods for multimodal signal processing, analysis, and modeling - Contains numerous examples of systems with different modalities combined - Describes advanced applications in multimodal Human-Computer Interaction (HCI) as well as in computer-based analysis and modelling of multimodal human-human communication scenes.
Social Signal Processing is the first book to cover all aspects of the modeling, automated detection, analysis, and synthesis of nonverbal behavior in human-human and human-machine interactions. Authoritative surveys address conceptual foundations, machine analysis and synthesis of social signal processing, and applications. Foundational topics include affect perception and interpersonal coordination in communication; later chapters cover technologies for automatic detection and understanding such as computational paralinguistics and facial expression analysis and for the generation of artificial social signals such as social robots and artificial agents. The final section covers a broad spectrum of applications based on social signal processing in healthcare, deception detection, and digital cities, including detection of developmental diseases and analysis of small groups. Each chapter offers a basic introduction to its topic, accessible to students and other newcomers, and then outlines challenges and future perspectives for the benefit of experienced researchers and practitioners in the field.
With the advance of speech, image and video technology, human-computer interaction (HCI) will reach a new phase.In recent years, HCI has been extended to human-machine communication (HMC) and the perceptual user interface (PUI). The final goal in HMC is that the communication between humans and machines is similar to human-to-human communication. Moreover, the machine can support human-to-human communication (e.g. an interface for the disabled). For this reason, various aspects of human communication are to be considered in HMC. The HMC interface, called a multimodal interface, includes different types of input methods, such as natural language, gestures, face and handwriting characters.The nine papers in this book have been selected from the 92 high-quality papers constituting the proceedings of the 2nd International Conference on Multimodal Interface (ICMI '99), which was held in Hong Kong in 1999. The papers cover a wide spectrum of the multimodal interface.
Describes advanced applications in multimodal Human-Computer Interaction (HCI) as well as in computer-based analysis and modelling of multimodal human-human communication scenes. Multimodal signal processing is an important research and development field that processes signals and combines information from a variety of modalities - speech, vision, language, text - which significantly enhance the understanding, modelling, and performance of human-computer interaction devices or systems enhancing human-human communication. The overarching theme of this book is the application of signal processing and statistical machine learning techniques to problems arising in this multi-disciplinary field. It describes the capabilities and limitations of current technologies, and discusses the technical challenges that must be overcome to develop efficient and user-friendly multimodal interactive systems. With contributions from the leading experts in the field, the present book should serve as a reference in multimodal signal processing for signal processing researchers, graduate students, R&D engineers, and computer engineers who are interested in this emerging field.
The book presents a comprehensive and up-to-date review of fuzzy pattern recognition. It carefully discusses a range of methodological and algorithmic issues, as well as implementations and case studies, and identifies the best design practices, assesses business models and practices of pattern recognition in real-world applications in industry, health care, administration, and business. Since the inception of fuzzy sets, fuzzy pattern recognition with its methodology, algorithms, and applications, has offered new insights into the principles and practice of pattern classification. Computational intelligence (CI) establishes a comprehensive framework aimed at fostering the paradigm of pattern recognition. The collection of contributions included in this book offers a representative overview of the advances in the area, with timely, in-depth and comprehensive material on the conceptually appealing and practically sound methodology and practices of CI-based pattern recognition.