Novel Methods for Robust Real-time Hand Gesture Interfaces

Novel Methods for Robust Real-time Hand Gesture Interfaces
Author: Nathaniel Sean Rossol
Publisher:
Total Pages: 110
Release: 2015
Genre: Computer vision
ISBN:

Real-time control of visual display systems via mid-air hand gestures offers many advantages over traditional interaction modalities. In medicine, for example, it allows a practitioner to adjust display values, e.g. contrast or zoom, on a medical visualization interface without the need to re-sterilize the interface. However, there are many practical challenges that make such interfaces non-robust including poor tracking due to frequent occlusion of fingers, interference from hand-held objects, and complex interfaces that are difficult for users to learn to use efficiently. In this work, various techniques are explored for improving the robustness of computer interfaces that use hand gestures. This work is focused predominately on real-time markerless Computer Vision (CV) based tracking methods with an emphasis on systems with high sampling rates. First, we explore a novel approach to increase hand pose estimation accuracy from multiple sensors at high sampling rates in real-time. This approach is achieved through an intelligent analysis of pose estimations from multiple sensors in a way that is highly scalable because raw image data is not transmitted between devices. Experimental results demonstrate that our proposed technique significantly improves the pose estimation accuracy while still maintaining the ability to capture individual hand poses at over 120 frames per second. Next, we explore techniques for improving pose estimation for the purposes of gesture recognition in situations where only a single sensor is used at high sampling rates without image data. In this situation, we demonstrate an approach where a combination of kinematic constraints and computed heuristics are used to estimate occluded keypoints to produce a partial pose estimation of a user's hand which is then used with our gestures recognition system to control a display. The results of our user study demonstrate that the proposed algorithm significantly improves the gesture recognition rate of the setup. We then explore gesture interface designs for situations where the user may (or may not) have a large portion of their hand occluded by a hand-held tool while gesturing. We address this challenge by developing a novel interface that uses a single set of gestures designed to be equally effective for fingers and hand-held tools without the need for any markers. The effectiveness of our approach is validated through a user study on a group of people given the task of adjusting parameters on a medical image display. Finally, we examine improving the efficiency of training for our interfaces by automatically assessing key user performance metrics (such as dexterity and confidence), and adapting the interface accordingly to reduce user frustration. We achieve this through a framework that uses Bayesian networks to estimate values for abstract hidden variables in our user model, based on analysis of data recorded from the user during operation of our system.

Robust Hand Gesture Recognition for Robotic Hand Control

Robust Hand Gesture Recognition for Robotic Hand Control
Author: Ankit Chaudhary
Publisher: Springer
Total Pages: 108
Release: 2017-06-05
Genre: Technology & Engineering
ISBN: 9811047987

This book focuses on light invariant bare hand gesture recognition while there is no restriction on the types of gestures. Observations and results have confirmed that this research work can be used to remotely control a robotic hand using hand gestures. The system developed here is also able to recognize hand gestures in different lighting conditions. The pre-processing is performed by developing an image-cropping algorithm that ensures only the area of interest is included in the segmented image. The segmented image is compared with a predefined gesture set which must be installed in the recognition system. These images are stored and feature vectors are extracted from them. These feature vectors are subsequently presented using an orientation histogram, which provides a view of the edges in the form of frequency. Thereby, if the same gesture is shown twice in different lighting intensities, both repetitions will map to the same gesture in the stored data. The mapping of the segmented image's orientation histogram is firstly done using the Euclidian distance method. Secondly, the supervised neural network is trained for the same, producing better recognition results. An approach to controlling electro-mechanical robotic hands using dynamic hand gestures is also presented using a robot simulator. Such robotic hands have applications in commercial, military or emergency operations where human life cannot be risked. For such applications, an artificial robotic hand is required to perform real-time operations. This robotic hand should be able to move its fingers in the same manner as a human hand. For this purpose, hand geometry parameters are obtained using a webcam and also using KINECT. The parameter detection is direction invariant in both methods. Once the hand parameters are obtained, the fingers’ angle information is obtained by performing a geometrical analysis. An artificial neural network is also implemented to calculate the angles. These two methods can be used with only one hand, either right or left. A separate method that is applicable to both hands simultaneously is also developed and fingers angles are calculated. The contents of this book will be useful for researchers and professional engineers working on robotic arm/hand systems.

Real-time 2D Static Hand Gesture Recognition and 2D Hand Tracking for Human-Computer Interaction

Real-time 2D Static Hand Gesture Recognition and 2D Hand Tracking for Human-Computer Interaction
Author: Pavel Alexandrovich Popov
Publisher:
Total Pages:
Release: 2020
Genre:
ISBN:

The topic of this thesis is Hand Gesture Recognition and Hand Tracking for user interface applications. 3 systems were produced, as well as datasets for recognition and tracking, along with UI applications to prove the concept of the technology. These represent significant contributions to resolving the hand recognition and tracking problems for 2d systems. The systems were designed to work in video only contexts, be computationally light, provide recognition and tracking of the user's hand, and operate without user driven fine tuning and calibration. Existing systems require user calibration, use depth sensors and do not work in video only contexts, or are computationally heavy requiring GPU to run in live situations. A 2-step static hand gesture recognition system was created which can recognize 3 different gestures in real-time. A detection step detects hand gestures using machine learning models. A validation step rejects false positives. The gesture recognition system was combined with hand tracking. It recognizes and then tracks a user's hand in video in an unconstrained setting. The tracking uses 2 collaborative strategies. A contour tracking strategy guides a minimization based template tracking strategy and makes it real-time, robust, and recoverable, while the template tracking provides stable input for UI applications. Lastly, an improved static gesture recognition system addresses the drawbacks due to stratified colour sampling of the detection boxes in the detection step. It uses the entire presented colour range and clusters it into constituent colour modes which are then used for segmentation, which improves the overall gesture recognition rates. One dataset was produced for static hand gesture recognition which allowed for the comparison of multiple different machine learning strategies, including deep learning. Another dataset was produced for hand tracking which provides a challenging series of user scenarios to test the gesture recognition and hand tracking system. Both datasets are significantly larger than other available datasets. The hand tracking algorithm was used to create a mouse cursor control application, a paint application for Android mobile devices, and a FPS video game controller. The latter in particular demonstrates how the collaborating hand tracking can fulfill the demanding nature of responsive aiming and movement controls.

Real-Time Vision for Human-Computer Interaction

Real-Time Vision for Human-Computer Interaction
Author: Branislav Kisacanin
Publisher: Springer Science & Business Media
Total Pages: 308
Release: 2005-12-06
Genre: Computers
ISBN: 0387278907

200Ts Vision of Vision One of my formative childhood experiences was in 1968 stepping into the Uptown Theater on Connecticut Avenue in Washington, DC, one of the few movie theaters nationwide that projected in large-screen cinerama. I was there at the urging of a friend, who said I simply must see the remarkable film whose run had started the previous week. "You won't understand it," he said, "but that doesn't matter. " All I knew was that the film was about science fiction and had great special eflPects. So I sat in the front row of the balcony, munched my popcorn, sat back, and experienced what was widely touted as "the ultimate trip:" 2001: A Space Odyssey. My friend was right: I didn't understand it. . . but in some senses that didn't matter. (Even today, after seeing the film 40 times, I continue to discover its many subtle secrets. ) I just had the sense that I had experienced a creation of the highest aesthetic order: unique, fresh, awe inspiring. Here was a film so distinctive that the first half hour had no words whatsoever; the last half hour had no words either; and nearly all the words in between were banal and irrelevant to the plot - quips about security through Voiceprint identification, how to make a phonecall from a space station, government pension plans, and so on.

Image Analysis

Image Analysis
Author: Bjarne K. Ersboll
Publisher: Springer
Total Pages: 1001
Release: 2007-07-03
Genre: Computers
ISBN: 3540730400

This book constitutes the refereed proceedings of the 15th Scandinavian Conference on Image Analysis, SCIA 2007, held in Aalborg, Denmark in June 2007. It covers computer vision, 2D and 3D reconstruction, classification and segmentation, medical and biological applications, appearance and shape modeling, face detection, tracking and recognition, motion analysis, feature extraction and object recognition.

Dual-sensor Approaches for Real-time Robust Hand Gesture Recognition

Dual-sensor Approaches for Real-time Robust Hand Gesture Recognition
Author: Kui Liu
Publisher:
Total Pages: 198
Release: 2015
Genre: Gesture
ISBN:

The use of hand gesture recognition has been steadily growing in various human-computer interaction applications. Under realistic operating conditions, it has been shown that hand gesture recognition systems exhibit recognition rate limitations when using a single sensor. Two dual-sensor approaches have thus been developed in this dissertation in order to improve the performance of hand gesture recognition under realistic operating conditions. The first approach involves the use of image pairs from a stereo camera setup by merging the image information from the left and right camera, while the second approach involves the use of a Kinect depth camera and an inertial sensor by fusing differing modality data within the framework of a hidden Markov model. The emphasis of this dissertation has been on system building and practical deployment. More specifically, the major contributions of the dissertation are: (a) improvement of hand gestures recognition rates when using a pair of images from a stereo camera compared to when using a single image by fusing the information from the left and right images in a complementary manner, and (b) improvement of hand gestures recognition rates when using a dual-modality sensor setup consisting of a Kinect depth camera and an inertial body sensor compared to the situations when each sensor is used individually on its own. Experimental results obtained indicate that the developed approaches generate higher recognition rates in different backgrounds and lighting conditions compared to the situations when an individual sensor is used. Both approaches are designed such that the entire recognition system runs in real-time on PC platform.

Markov Models for Pattern Recognition

Markov Models for Pattern Recognition
Author: Gernot A. Fink
Publisher: Springer Science & Business Media
Total Pages: 275
Release: 2014-01-14
Genre: Computers
ISBN: 1447163087

This thoroughly revised and expanded new edition now includes a more detailed treatment of the EM algorithm, a description of an efficient approximate Viterbi-training procedure, a theoretical derivation of the perplexity measure and coverage of multi-pass decoding based on n-best search. Supporting the discussion of the theoretical foundations of Markov modeling, special emphasis is also placed on practical algorithmic solutions. Features: introduces the formal framework for Markov models; covers the robust handling of probability quantities; presents methods for the configuration of hidden Markov models for specific application areas; describes important methods for efficient processing of Markov models, and the adaptation of the models to different tasks; examines algorithms for searching within the complex solution spaces that result from the joint application of Markov chain and hidden Markov models; reviews key applications of Markov models.

Human-Computer Interaction. Novel Interaction Methods and Techniques

Human-Computer Interaction. Novel Interaction Methods and Techniques
Author: Julie A. Jacko
Publisher: Springer Science & Business Media
Total Pages: 923
Release: 2009-07-14
Genre: Computers
ISBN: 3642025773

The 13th International Conference on Human–Computer Interaction, HCI Inter- tional 2009, was held in San Diego, California, USA, July 19–24, 2009, jointly with the Symposium on Human Interface (Japan) 2009, the 8th International Conference on Engineering Psychology and Cognitive Ergonomics, the 5th International Conference on Universal Access in Human–Computer Interaction, the Third International Conf- ence on Virtual and Mixed Reality, the Third International Conference on Internati- alization, Design and Global Development, the Third International Conference on Online Communities and Social Computing, the 5th International Conference on Augmented Cognition, the Second International Conference on Digital Human Mod- ing, and the First International Conference on Human Centered Design. A total of 4,348 individuals from academia, research institutes, industry and gove- mental agencies from 73 countries submitted contributions, and 1,397 papers that were judged to be of high scientific quality were included in the program. These papers - dress the latest research and development efforts and highlight the human aspects of design and use of computing systems. The papers accepted for presentation thoroughly cover the entire field of human–computer interaction, addressing major advances in the knowledge and effective use of computers in a variety of application areas.

Robust Real-time Hands-and-face Detection for Human Robot Interaction

Robust Real-time Hands-and-face Detection for Human Robot Interaction
Author: SeyedMehdi MohaimenianPour
Publisher:
Total Pages: 91
Release: 2018
Genre:
ISBN:

With recent advances, robots have become more affordable and intelligent, which expands their application domain and number of consumers. Having robots around us in our daily lives creates a demand for an interaction system for communicating humans' intentions and commands to robots. We are interested in interactions that are easy, intuitive, and do not require the human to use any additional equipment. We present a robust real-time system for visual detection of hands and faces in RGB and gray-scale images based on a Deep Convolutional Neural Network. This system is designed to meet the requirements of a hands-free interface to UAVs described below that could be used for communicating to other robots equipped with a monocular camera using only hands and face gestures without any extra instruments. This work is accompanied by a novel hands-and-faces detection dataset gathered and labelled from a wide variety of sources including our own Human-UAV interaction videos, and several third-party datasets. By training our model on all these data, we obtain qualitatively good detection results in terms of both accuracy and speed on a commodity GPU. The same detector gives state-of-the-art accuracy and speed in a hand-detection benchmark and competitive results in a face detection benchmark. To demonstrate its effectiveness for Human-Robot Interaction we describe its use as the input to a novel, simple but practical gestural Human-UAV interface for static gesture detection based on hand position relative to the face. A small vocabulary of hand gestures is used to demonstrate our end-to-end pipeline for un-instrumented human-UAV interaction useful for entertainment or industrial applications. All software, training and test data produced for this thesis is released as an Open Source contribution.