Toward an Intelligent Multimodal Interface for Natural Interaction

Toward an Intelligent Multimodal Interface for Natural Interaction
Author: Ying Yin (Ph. D.)
Publisher:
Total Pages: 76
Release: 2010
Genre:
ISBN:

Advances in technology are enabling novel approaches to human-computer interaction (HCI) in a wide variety of devices and settings (e.g., the Microsoft® Surface, the Nintendo® Wii, iPhone®, etc.). While many of these devices have been commercially successful, the use of multimodal interaction technology is still not well understood from a more principled system design or cognitive science perspective. The long-term goal of our research is to build an intelligent multimodal interface for natural interaction that can serve as a testbed for enabling the formulation of a more principled system design framework for multimodal HCI. This thesis focuses on the gesture input modality. Using a new hand tracking technology capable of tracking 3D hand postures in real-time, we developed a recognition system for continuous natural gestures. By nature gestures, we mean the ones encountered in spontaneous interaction, rather than a set of artificial gestures designed for the convenience of recognition. To date we have achieved 96% accuracy on isolated gesture recognition, and 74% correct rate on continuous gesture recognition with data from different users and twelve gesture classes. We are able to connect the gesture recognition system with Google Earth, enabling gestural control of a 3D map. In particular, users can do 3D tilting of the map using non touch-based gesture which is more intuitive than touch-based ones. We also did an exploratory user study to observe natural behavior under a urban search and rescue scenario with a large tabletop display. The qualitative results from the study provides us with good starting points for understanding how users naturally gesture, and how to integrate different modalities. This thesis has set the stage for further development towards our long-term goal.

Multimodal Interfaces

Multimodal Interfaces
Author: Thomas Moran
Publisher: CRC Press
Total Pages: 0
Release: 1997
Genre: Graphical user interfaces (Computer systems)
ISBN: 9780805898675

The growing emphasis on multimodal interface design is fundamentally inspired by the aim to support natural, easy to learn and use, flexible, efficient, and powerfully expressive means of human-computer interaction. Most of the articles in this special issue present work in support of challenging applications such as algebra instruction, data summaries, and interaction with complex spatial displays. A collection of emerging research ideas on next-generation multimodal interfaces, it also addresses multimodal interface design for portable devices to be used in natural field settings. Additionally, it describes implemented systems that make computing accessible to the visually impaired.

Building an Intuitive Multimodal Interface for a Smart Home

Building an Intuitive Multimodal Interface for a Smart Home
Author: John N.A Brown
Publisher: Springer
Total Pages: 86
Release: 2017-05-03
Genre: Computers
ISBN: 331956532X

This book describes an innovative approach to the interaction between humans and a smart environment; an attempt to get a smart home to understand intuitive, multi-modal, human-centred communication. State of the art smart homes, like other “smart” technology, tend to demand that the human user must adapt herself to the needs of the system. The hunt for a truly user-centred, truly intuitive system has long proven to be beyond the grasp of current technology. When humans speak with one another, we are multimodal. Our speech is supplemented with gestures, which serve as a parallel stream of information, reinforcing the meaning of our words. Drawing on well-established protocols in engineering and psychology, and with no small amount of inspiration from a particular nonsense poem, we have successfully concluded that hunt. This book describes the efforts, undertaken over several years, to design, implement, and test a model of interaction that allows untrained individuals to intuitively control a complex series of networked and embedded systems. The theoretical concepts are supported by a series of experimental studies, showing the advantages of the novel approach, and pointing towards future work that would facilitate the deployment of this concept in the real world.

Human-Computer Interaction. HCI Intelligent Multimodal Interaction Environments

Human-Computer Interaction. HCI Intelligent Multimodal Interaction Environments
Author: Julie A. Jacko
Publisher: Springer
Total Pages: 1049
Release: 2007-08-24
Genre: Computers
ISBN: 3540731105

Here is the third of a four-volume set that constitutes the refereed proceedings of the 12th International Conference on Human-Computer Interaction, HCII 2007, held in Beijing, China, in July 2007, jointly with eight other thematically similar conferences. It covers multimodality and conversational dialogue; adaptive, intelligent and emotional user interfaces; gesture and eye gaze recognition; and interactive TV and media.

Design of Multimodal Mobile Interfaces

Design of Multimodal Mobile Interfaces
Author: Nava Shaked
Publisher: Walter de Gruyter GmbH & Co KG
Total Pages: 223
Release: 2016-04-25
Genre: Technology & Engineering
ISBN: 1501502751

The “smart mobile” has become an essential and inseparable part of our lives. This powerful tool enables us to perform multi-tasks in different modalities of voice, text, gesture, etc. The user plays an important role in the mode of operation, so multimodal interaction provides the user with new complex multiple modalities of interfacing with a system, such as speech, touch, type and more. The book will discuss the new world of mobile multimodality, focusing on innovative technologies and design which create a state-of-the-art user interface. It will examine the practical challenges entailed in meeting commercial deployment goals, and offer new approaches to the designing such interfaces. A multimodal interface for mobile devices requires the integration of several recognition technologies together with sophisticated user interface and distinct tools for input and output of data. The book will address the challenge of designing devices in a synergetic fashion which does not burden the user or to create a technological overload.

Multimodal Interface for Human-machine Communication

Multimodal Interface for Human-machine Communication
Author: P. C. Yuen
Publisher: World Scientific
Total Pages: 288
Release: 2002
Genre: Computers
ISBN: 9789810245948

With the advance of speech, image and video technology, human-computer interaction (HCI) will reach a new phase.In recent years, HCI has been extended to human-machine communication (HMC) and the perceptual user interface (PUI). The final goal in HMC is that the communication between humans and machines is similar to human-to-human communication. Moreover, the machine can support human-to-human communication (e.g. an interface for the disabled). For this reason, various aspects of human communication are to be considered in HMC. The HMC interface, called a multimodal interface, includes different types of input methods, such as natural language, gestures, face and handwriting characters.The nine papers in this book have been selected from the 92 high-quality papers constituting the proceedings of the 2nd International Conference on Multimodal Interface (ICMI '99), which was held in Hong Kong in 1999. The papers cover a wide spectrum of the multimodal interface.

The Paradigm Shift to Multimodality in Contemporary Computer Interfaces

The Paradigm Shift to Multimodality in Contemporary Computer Interfaces
Author: Sharon Oviatt
Publisher: Morgan & Claypool Publishers
Total Pages: 245
Release: 2015-04-01
Genre: Computers
ISBN: 1627057528

During the last decade, cell phones with multimodal interfaces based on combined new media have become the dominant computer interface worldwide. Multimodal interfaces support mobility and expand the expressive power of human input to computers. They have shifted the fulcrum of human-computer interaction much closer to the human. This book explains the foundation of human-centered multimodal interaction and interface design, based on the cognitive and neurosciences, as well as the major benefits of multimodal interfaces for human cognition and performance. It describes the data-intensive methodologies used to envision, prototype, and evaluate new multimodal interfaces. From a system development viewpoint, this book outlines major approaches for multimodal signal processing, fusion, architectures, and techniques for robustly interpreting users' meaning. Multimodal interfaces have been commercialized extensively for field and mobile applications during the last decade. Research also is growing rapidly in areas like multimodal data analytics, affect recognition, accessible interfaces, embedded and robotic interfaces, machine learning and new hybrid processing approaches, and similar topics. The expansion of multimodal interfaces is part of the long-term evolution of more expressively powerful input to computers, a trend that will substantially improve support for human cognition and performance.

Multimodal Interaction with W3C Standards

Multimodal Interaction with W3C Standards
Author: Deborah A. Dahl
Publisher: Springer
Total Pages: 430
Release: 2016-11-17
Genre: Technology & Engineering
ISBN: 3319428160

This book presents new standards for multimodal interaction published by the W3C and other standards bodies in straightforward and accessible language, while also illustrating the standards in operation through case studies and chapters on innovative implementations. The book illustrates how, as smart technology becomes ubiquitous, and appears in more and more different shapes and sizes, vendor-specific approaches to multimodal interaction become impractical, motivating the need for standards. This book covers standards for voice, emotion, natural language understanding, dialog, and multimodal architectures. The book describes the standards in a practical manner, making them accessible to developers, students, and researchers. Comprehensive resource that explains the W3C standards for multimodal interaction clear and straightforward way; Includes case studies of the use of the standards on a wide variety of devices, including mobile devices, tablets, wearables and robots, in applications such as assisted living, language learning, and health care; Features illustrative examples of implementations that use the standards, to help spark innovative ideas for future applications.

Readings in Intelligent User Interfaces

Readings in Intelligent User Interfaces
Author: Mark Maybury
Publisher: Morgan Kaufmann
Total Pages: 670
Release: 1998-04
Genre: Computers
ISBN: 9781558604445

This is a compilation of the classic readings in intelligent user interfaces. This text focuses on intelligent, knowledge-based interfaces, combining spoken language, natural language processing, and multimedia and multimodal processing.

Multimodal User Interfaces

Multimodal User Interfaces
Author: Dimitros Tzovaras
Publisher: Springer Science & Business Media
Total Pages: 321
Release: 2008-02-27
Genre: Technology & Engineering
ISBN: 3540783458

tionship indicates how multimodal medical image processing can be unified to a large extent, e. g. multi-channel segmentation and image registration, and extend information theoretic registration to other features than image intensities. The framework is not at all restricted to medical images though and this is illustrated by applying it to multimedia sequences as well. In Chapter 4, the main results from the developments in plastic UIs and mul- modal UIs are brought together using a theoretic and conceptual perspective as a unifying approach. It is aimed at defining models useful to support UI plasticity by relying on multimodality, at introducing and discussing basic principles that can drive the development of such UIs, and at describing some techniques as proof-of-concept of the aforementioned models and principles. In Chapter 4, the authors introduce running examples that serve as illustration throughout the d- cussion of the use of multimodality to support plasticity.