vision-based hand gesture recognition for human-computer ... · pdf filevision-based hand...

Download Vision-based Hand Gesture Recognition for Human-Computer ... · PDF fileVision-based Hand Gesture Recognition for Human-Computer ... † techniques that are based on cameras ... to

If you can't read please download the document

Upload: hadien

Post on 06-Feb-2018

221 views

Category:

Documents


2 download

TRANSCRIPT

  • Vision-based Hand Gesture Recognitionfor Human-Computer Interaction

    X. Zabulis, H. Baltzakis, A. Argyros

    Institute of Computer ScienceFoundation for Research and Technology - Hellas (FORTH)Heraklion, Crete, Greece

    Computer Science Department, University of CreteHeraklion, Crete, Grece

    {zabulis,xmpalt,argyros}@ics.forth.gr1 Introduction

    In recent years, research efforts seeking to provide more natural, human-centeredmeans of interacting with computers have gained growing interest. A particu-larly important direction is that of perceptive user interfaces, where the com-puter is endowed with perceptive capabilities that allow it to acquire both im-plicit and explicit information about the user and the environment. Vision hasthe potential of carrying a wealth of information in a non-intrusive manner andat a low cost, therefore it constitutes a very attractive sensing modality fordeveloping perceptive user interfaces. Proposed approaches for vision-driveninteractive user interfaces resort to technologies such as head tracking, face andfacial expression recognition, eye tracking and gesture recognition.

    In this paper, we focus our attention to vision-based recognition of handgestures. The first part of the paper provides an overview of the current stateof the art regarding the recognition of hand gestures as these are observed andrecorded by typical video cameras. In order to make the review of the relatedliterature tractable, this paper does not discuss:

    techniques that are based on cameras operating beyond the visible spec-trum (e.g. thermal cameras, etc),

    active techniques that require the projection of some form of structuredlight, and,

    1

  • 2 Computer Vision Techniques for Hand Gesture Recognition 2

    invasive techniques that require modifications of the environment, e.g.that the user wears gloves of particular color distribution or with particularmarkers.

    Despite these restrictions, a complete review of the computer vision-based tech-nology for hand gesture recognition remains a very challenging task. Never-theless, and despite the fact that the provided review might not be complete,an effort was made to report research results pertaining to the full cycle of vi-sual processing towards gesture recognition, covering issues from low level imageanalysis and feature extraction to higher level interpretation techniques.

    The second part of the paper presents a specific approach taken to gesturerecognition intended to support natural interaction with autonomous robotsthat guide visitors in museums and exhibition centers. The proposed gesturerecognition system builds on a probabilistic framework that allows the utiliza-tion of multiple information cues to efficiently detect image regions that belongto human hands. Tracking over time is achieved by a technique that can simul-taneously handle multiple hands that may move in complex trajectories, occludeeach other in the field of view of the robots camera and vary in number overtime. Dependable hand tracking, combined with fingertip detection, facilitatesthe definition of a small, simple, intuitive hand gestures vocabulary that can beused to support robust human robot interaction. Sample experimental resultspresented in this paper, confirm the effectiveness and the efficiency of the pro-posed approach, meeting the robustness and performance requirements of thisparticular case of human-computer interaction.

    2 Computer Vision Techniques for Hand Gesture Recognition

    Most of the complete hand interactive systems can be considered to be comprisedof three layers: detection, tracking and recognition. The detection layer isresponsible for defining and extracting visual features that can be attributed tothe presence of hands in the field of view of the camera(s). The tracking layer isresponsible for performing temporal data association between successive imageframes, so that, at each moment in time, the system may be aware of whatis where. Moreover, in model-based methods, tracking also provides a wayto maintain estimates of model parameters, variables and features that are notdirectly observable at a certain moment in time. Last, the recognition layer isresponsible for grouping the spatiotemporal data extracted in the previous layersand assigning the resulting groups with labels associated to particular classesof gestures. In this section, research on these three identified subproblems ofvision-based gesture recognition is reviewed.

    2.1 Detection

    The primary step in gesture recognition systems is the detection of hands andthe segmentation of the corresponding image regions. This segmentation iscrucial because it isolates the task-relevant data from the image background,

  • 2 Computer Vision Techniques for Hand Gesture Recognition 3

    before passing them to the subsequent tracking and recognition stages. A largenumber of methods have been proposed in the literature that utilize a severaltypes of visual features and, in many cases, their combination. Such featuresare skin color, shape, motion and anatomical models of hands. In [CPC06], acomparative study on the performance of some hand segmentation techniquescan be found.

    2.1.1 Color

    Skin color segmentation has been utilized by several approaches for hand detec-tion. A major decision towards providing a model of skin color is the selectionof the color space to be employed. Several color spaces have been proposedincluding RGB, normalized RGB, HSV, YCrCb, YUV, etc. Color spaces ef-ficiently separating the chromaticity from the luminance components of colorare typically considered preferable. This is due to the fact that by employingchromaticity-dependent components of color only, some degree of robustness toillumination changes can be achieved. Terrillon et al [TSFA00] review differentskin chromaticity models and evaluate their performance.

    To increase invariance against illumination variability some methods [MC97,Bra98, Kam98, FM99, HVD+99a, KOKS01] operate in the HSV [SF96], YCrCb[CN98], or YUV [YLW98, AL04b] colorspaces, in order to approximate thechromaticity of skin (or, in essence, its absorption spectrum) rather than itsapparent color value. They typically eliminate the luminance component, toremove the effect of shadows, illumination changes, as well as modulations oforientation of the skin surface relative to the light source(s). The remaining2D color vector is nearly constant for skin regions and a 2D histogram of thepixels from a region containing skin shows a strong peak at the skin color.Regions where this probability is above a threshold are detected and describedusing connected components analysis. In several cases (e.g. [AL04b]), hysteresisthresholding on the derived probabilities is also employed prior to connectedcomponents labeling. The rationale of hysteresis thresholding is that pixels withrelatively low probability of being skin-colored, should be interpreted as such incase that they are connected to pixels with high such probability.Having selecteda suitable color space, the simplest approach for defining what constitutes skincolor is to employ bounds on the coordinates of the selected space [CN98]. Thesebounds are typically selected empirically, i.e. by examining the distribution ofskin colors in a preselected set of images. Another approach is to assume thatthe probabilities of skin colors follow a distribution that can be learned eitheroff-line or by employing an on-line iterative method [SF96].

    Several methods [SF96, KK96, SWP98, DKS01, JR02, AL04b, SSA04] uti-lize precomputed color distributions extracted from statistical analysis of largedatasets. For example, in [JR02], a statistical model of skin color was obtainedfrom the analysis of thousands of photos on the Web. In contrast, methods suchas those described in [KOKS01, ZYW00] build a color model based on collectedsamples of skin color during system initialization.

    When using a histogram to represent a color distribution (as for example

  • 2 Computer Vision Techniques for Hand Gesture Recognition 4

    in [JR02, KK96, WLH00]), the color space is quantized and, thus, the level ofquantization affects the shape of the histogram. Parametric models of the colordistribution have also been used in the form of a single Gaussian distribution[KKAK98, YLW98, CG99] or a mixture of Gaussians [RMG98, RG98, SRG99,JP97, JRP97]. Maximum-likelihood estimation techniques can be thereafterutilized to infer the parameters of the probability density functions. In an-other parametric approach [WLH00], an unsupervised clustering algorithm toapproximate color distribution is based on a self-organizing map.

    The perceived color of human skin varies greatly across human races or evenbetween individuals of the same race. Additional variability may be introduceddue to changing illumination conditions and/or camera characteristics. There-fore, color-based approaches to hand detection need to employ some means forcompensating for this variability. In [YA98, SSA04], an invariant representationof skin color against changes in illumination is pursued, but still with not conclu-sive results. In [YLW98], an adaptation technique estimates the new parametersfor the mean and covariance of the multivariate Gaussian skin color distribution,based on a linear combination of previous parameters. However, most of thesemethods are still sensitive to quickly changing or mixed lighting conditions. Asimple color comparison scheme is employed in [DKS01], where the dominantcolor of a homogeneous region is tested as if occurring within a color range thatcorresponds to skin color variability. Other approaches [Bra98, KOKS01, MC97]consider skin color to be uniform across image space and extract the pursuedregions through typical region-growing and pixel-grouping techniques. More ad-vanced color segmentation techniques rel