Gesture recognition

Last updated
A child's hand location and movement being detected by a gesture recognition algorithm Gesture Recognition.jpg
A child's hand location and movement being detected by a gesture recognition algorithm

Gesture recognition is an area of research and development in computer science and language technology concerned with the recognition and interpretation of human gestures. A subdiscipline of computer vision,[ citation needed ] it employs mathematical algorithms to interpret gestures. [1]

Contents

Gesture recognition offers a path for computers to begin to better understand and interpret human body language, previously not possible through text or unenhanced graphical (GUI) user interfaces.

Gestures can originate from any bodily motion or state, but commonly originate from the face or hand. One area of the field is emotion recognition derived from facial expressions and hand gestures. Users can make simple gestures to control or interact with devices without physically touching them.

Many approaches have been made using cameras and computer vision algorithms to interpret sign language, however, the identification and recognition of posture, gait, proxemics, and human behaviors is also the subject of gesture recognition techniques. [2]

Overview

Middleware usually processes gesture recognition, then sends the results to the user. Linux kernel and gaming input-output latency.svg
Middleware usually processes gesture recognition, then sends the results to the user.

Gesture recognition has application in such areas as:[ when? ]

Gesture recognition can be conducted with techniques from computer vision and image processing. [5]

The literature includes ongoing work in the computer vision field on capturing gestures or more general human pose and movements by cameras connected to a computer. [6] [7] [8] [9]

The term "gesture recognition" has been used to refer more narrowly to non-text-input handwriting symbols, such as inking on a graphics tablet, multi-touch gestures, and mouse gesture recognition. This is computer interaction through the drawing of symbols with a pointing device cursor. [10] [11] [12] Pen computing expands digital gesture recognition beyond traditional input devices such as keyboards and mice, and reduces the hardware impact of a system.[ how? ]

Gesture types

In computer interfaces, two types of gestures are distinguished: [13] We consider online gestures, which can also be regarded as direct manipulations like scaling and rotating, and in contrast, offline gestures are usually processed after the interaction is finished; e. g. a circle is drawn to activate a context menu.

Touchless interface

A touchless user interface (TUI) is an emerging type of technology wherein a device is controlled via body motion and gestures without touching a keyboard, mouse, or screen. [14]

Types of touchless technology

There are several devices utilizing this type of interface such as smartphones, laptops, games, TVs, and music equipment.

One type of touchless interface uses the Bluetooth connectivity of a smartphone to activate a company's visitor management system. This eliminates having to touch an interface, for convenience or to avoid a potential source of contamination as during the COVID-19 pandemic. [15]

Input devices

The ability to track a person's movements and determine what gestures they may be performing can be achieved through various tools. Kinetic user interfaces (KUIs) are an emerging type of user interfaces that allow users to interact with computing devices through the motion of objects and bodies.[ citation needed ] Examples of KUIs include tangible user interfaces and motion-aware games such as Wii and Microsoft's Kinect, and other interactive projects. [16]

Although there is a large amount of research done in image/video-based gesture recognition, there is some variation in the tools and environments used between implementations.

Algorithms

Some alternative methods of tracking and analyzing gestures, and their respective relationships BigDiagram2.jpg
Some alternative methods of tracking and analyzing gestures, and their respective relationships

Depending on the type of input data, the approach for interpreting a gesture could be done in different ways. However, most of the techniques rely on key pointers represented in a 3D coordinate system. Based on the relative motion of these, the gesture can be detected with high accuracy, depending on the quality of the input and the algorithm's approach. [30]

In order to interpret movements of the body, one has to classify them according to common properties and the message the movements may express. For example, in sign language, each gesture represents a word or phrase.

Some literature differentiates 2 different approaches in gesture recognition: a 3D model-based and an appearance-based. [31] The foremost method makes use of 3D information on key elements of the body parts in order to obtain several important parameters, like palm position or joint angles. Approaches derived from it such as the volumetric models have proven to be very intensive in terms of computational power and require further technological developments in order to be implemented for real-time analysis. Alternately, appearance-based systems use images or videos for direct interpretation. Such models are easier to process, but usually lack the generality required for human-computer interaction.

3D model-based algorithms

A real hand (left) is interpreted as a collection of vertices and lines in the 3D mesh version (right), and the software uses their relative position and interaction in order to infer the gesture. Volumetric-hands.jpg
A real hand (left) is interpreted as a collection of vertices and lines in the 3D mesh version (right), and the software uses their relative position and interaction in order to infer the gesture.

The 3D model approach can use volumetric or skeletal models or even a combination of the two. Volumetric approaches have been heavily used in the computer animation industry and for computer vision purposes. The models are generally created from complicated 3D surfaces, like NURBS or polygon meshes.

The drawback of this method is that it is very computationally intensive, and systems for real-time analysis are still to be developed. For the moment, a more interesting approach would be to map simple primitive objects to the person's most important body parts (for example cylinders for the arms and neck, sphere for the head) and analyze the way these interact with each other. Furthermore, some abstract structures like super-quadrics and generalized cylinders maybe even more suitable for approximating the body parts.

Skeletal-based algorithms

The skeletal version (right) is effectively modeling the hand (left). This has fewer parameters than the volumetric version and it's easier to compute, making it suitable for real-time gesture analysis systems. Skeletal-hand.jpg
The skeletal version (right) is effectively modeling the hand (left). This has fewer parameters than the volumetric version and it's easier to compute, making it suitable for real-time gesture analysis systems.

Instead of using intensive processing of the 3D models and dealing with a lot of parameters, one can just use a simplified version of joint angle parameters along with segment lengths. This is known as a skeletal representation of the body, where a virtual skeleton of the person is computed and parts of the body are mapped to certain segments. The analysis here is done using the position and orientation of these segments and the relation between each one of them( for example the angle between the joints and the relative position or orientation)

Advantages of using skeletal models:

Appearance-based models

These binary silhouette(left) or contour(right) images represent typical input for appearance-based algorithms. They are compared with different hand templates and if they match, the correspondent gesture is inferred. Appearance hands.jpg
These binary silhouette(left) or contour(right) images represent typical input for appearance-based algorithms. They are compared with different hand templates and if they match, the correspondent gesture is inferred.

Appearance-based models no longer use a spatial representation of the body, instead deriving their parameters directly from the images or videos using a template database. Some are based on the deformable 2D templates of the human parts of the body, particularly the hands. Deformable templates are sets of points on the outline of an object, used as interpolation nodes for the object's outline approximation. One of the simplest interpolation functions is linear, which performs an average shape from point sets, point variability parameters, and external deformation. These template-based models are mostly used for hand-tracking, but could also be used for simple gesture classification.

The second approach in gesture detection using appearance-based models uses image sequences as gesture templates. Parameters for this method are either the images themselves, or certain features derived from these. Most of the time, only one (monoscopic) or two (stereoscopic) views are used.

Electromyography-based models

Electromyography (EMG) concerns the study of electrical signals produced by muscles in the body. Through classification of data received from the arm muscles, it is possible to classify the action and thus input the gesture to external software. [1] Consumer EMG devices allow for non-invasive approaches such as an arm or leg band and connect via Bluetooth. Due to this, EMG has an advantage over visual methods since the user does not need to face a camera to give input, enabling more freedom of movement.

Challenges

There are many challenges associated with the accuracy and usefulness of gesture recognition and software designed to implement it. For image-based gesture recognition, there are limitations on the equipment used and image noise. Images or video may not be under consistent lighting, or in the same location. Items in the background or distinct features of the users may make recognition more difficult.

The variety of implementations for image-based gesture recognition may also cause issues with the viability of the technology for general usage. For example, an algorithm calibrated for one camera may not work for a different camera. The amount of background noise also causes tracking and recognition difficulties, especially when occlusions (partial and full) occur. Furthermore, the distance from the camera, and the camera's resolution and quality, also cause variations in recognition accuracy.

In order to capture human gestures by visual sensors robust computer vision methods are also required, for example for hand tracking and hand posture recognition [32] [33] [34] [35] [36] [37] [38] [39] [40] or for capturing movements of the head, facial expressions or gaze direction.

Social acceptability

One significant challenge to the adoption of gesture interfaces on consumer mobile devices such as smartphones and smartwatches stems from the social acceptability implications of gestural input. While gestures can facilitate fast and accurate input on many novel form-factor computers, their adoption and usefulness are often limited by social factors rather than technical ones. To this end, designers of gesture input methods may seek to balance both technical considerations and user willingness to perform gestures in different social contexts. [41] In addition, different device hardware and sensing mechanisms support different kinds of recognizable gestures.

Mobile device

Gesture interfaces on mobile and small form-factor devices are often supported by the presence of motion sensors such as inertial measurement units (IMUs). On these devices, gesture sensing relies on users performing movement-based gestures capable of being recognized by these motion sensors. This can potentially make capturing signals from subtle or low-motion gestures challenging, as they may become difficult to distinguish from natural movements or noise. Through a survey and study of gesture usability, researchers found that gestures that incorporate subtle movement, which appear similar to existing technology, look or feel similar to every action, and are enjoyable were more likely to be accepted by users, while gestures that look strange, are uncomfortable to perform, interfere with communication, or involve uncommon movement caused users more likely to reject their usage. [41] The social acceptability of mobile device gestures relies heavily on the naturalness of the gesture and social context.

On-body and wearable computers

Wearable computers typically differ from traditional mobile devices in that their usage and interaction location takes place on the user's body. In these contexts, gesture interfaces may become preferred over traditional input methods, as their small size renders touch-screens or keyboards less appealing. Nevertheless, they share many of the same social acceptability obstacles as mobile devices when it comes to gestural interaction. However, the possibility of wearable computers being hidden from sight or integrated into other everyday objects, such as clothing, allow gesture input to mimic common clothing interactions, such as adjusting a shirt collar or rubbing one's front pant pocket. [42] [43] A major consideration for wearable computer interaction is the location for device placement and interaction. A study exploring third-party attitudes towards wearable device interaction conducted across the United States and South Korea found differences in the perception of wearable computing use of males and females, in part due to different areas of the body considered socially sensitive. [43] Another study investigating the social acceptability of on-body projected interfaces found similar results, with both studies labelling areas around the waist, groin, and upper body (for women) to be least acceptable while areas around the forearm and wrist to be most acceptable. [44]

Public installations

Public Installations, such as interactive public displays, allow access to information and displays interactive media in public settings such as museums, galleries, and theaters. [45] While touch screens are a frequent form of input for public displays, gesture interfaces provide additional benefits such as improved hygiene, interaction from a distance, and improved discoverability, and may favor performative interaction. [42] An important consideration for gestural interaction with public displays is the high probability or expectation of a spectator audience. [45]

Fatigue

Arm fatigue was a side-effect of vertically oriented touch-screen or light-pen use. In periods of prolonged use, users' arms began to feel fatigued and/or discomfort. This effect contributed to the decline of touch-screen input despite its initial popularity in the 1980s. [46] [47]

In order to measure arm fatigue side effect, researchers developed a technique called Consumed Endurance. [48] [49]

See also

Related Research Articles

<span class="mw-page-title-main">Pointing device</span> Human interface device for computers

A pointing device is a human interface device that allows a user to input spatial data to a computer. CAD systems and graphical user interfaces (GUI) allow the user to control and provide data to the computer using physical gestures by moving a hand-held mouse or similar device across the surface of the physical desktop and activating switches on the mouse. Movements of the pointing device are echoed on the screen by movements of the pointer and other visual changes. Common gestures are point and click and drag and drop.

<span class="mw-page-title-main">Augmented reality</span> View of the real world with computer-generated supplementary features

Augmented reality (AR) is an interactive experience that combines the real world and computer-generated content. The content can span multiple sensory modalities, including visual, auditory, haptic, somatosensory and olfactory. AR can be defined as a system that incorporates three basic features: a combination of real and virtual worlds, real-time interaction, and accurate 3D registration of virtual and real objects. The overlaid sensory information can be constructive, or destructive. This experience is seamlessly interwoven with the physical world such that it is perceived as an immersive aspect of the real environment. In this way, augmented reality alters one's ongoing perception of a real-world environment, whereas virtual reality completely replaces the user's real-world environment with a simulated one.

<span class="mw-page-title-main">Tangible user interface</span>

A tangible user interface (TUI) is a user interface in which a person interacts with digital information through the physical environment. The initial name was Graspable User Interface, which is no longer used. The purpose of TUI development is to empower collaboration, learning, and design by giving physical forms to digital information, thus taking advantage of the human ability to grasp and manipulate physical objects and materials.

Multimodal interaction provides the user with multiple modes of interacting with a system. A multimodal interface provides several distinct tools for input and output of data.

<span class="mw-page-title-main">Wired glove</span> Input device for human–computer interaction

A wired glove is an input device for human–computer interaction worn like a glove.

<span class="mw-page-title-main">Multi-touch</span> Technology

In computing, multi-touch is technology that enables a surface to recognize the presence of more than one point of contact with the surface at the same time. The origins of multitouch began at CERN, MIT, University of Toronto, Carnegie Mellon University and Bell Labs in the 1970s. CERN started using multi-touch screens as early as 1976 for the controls of the Super Proton Synchrotron. Capacitive multi-touch displays were popularized by Apple's iPhone in 2007. Multi-touch may be used to implement additional functionality, such as pinch to zoom or to activate certain subroutines attached to predefined gestures using gesture recognition.

In computing, post-WIMP comprises work on user interfaces, mostly graphical user interfaces, which attempt to go beyond the paradigm of windows, icons, menus and a pointing device, i.e. WIMP interfaces.

Activity recognition aims to recognize the actions and goals of one or more agents from a series of observations on the agents' actions and the environmental conditions. Since the 1980s, this research field has captured the attention of several computer science communities due to its strength in providing personalized support for many different applications and its connection to many different fields of study such as medicine, human-computer interaction, or sociology.

<span class="mw-page-title-main">3D reconstruction</span> Process of capturing the shape and appearance of real objects

In computer vision and computer graphics, 3D reconstruction is the process of capturing the shape and appearance of real objects. This process can be accomplished either by active or passive methods. If the model is allowed to change its shape in time, this is referred to as non-rigid or spatio-temporal reconstruction.

<span class="mw-page-title-main">Interaction technique</span>

An interaction technique, user interface technique or input technique is a combination of hardware and software elements that provides a way for computer users to accomplish a single task. For example, one can go back to the previously visited page on a Web browser by either clicking a button, pressing a key, performing a mouse gesture or uttering a speech command. It is a widely used term in human-computer interaction. In particular, the term "new interaction technique" is frequently used to introduce a novel user interface design idea.

<span class="mw-page-title-main">Organic user interface</span> Type of user interface

In human–computer interaction, an organic user interface (OUI) is defined as a user interface with a non-flat display. After Engelbart and Sutherland's graphical user interface (GUI), which was based on the cathode ray tube (CRT), and Kay and Weiser's ubiquitous computing, which is based on the flat panel liquid-crystal display (LCD), OUI represents one possible third wave of display interaction paradigms, pertaining to multi-shaped and flexible displays. In an OUI, the display surface is always the focus of interaction, and may actively or passively change shape upon analog inputs. These inputs are provided through direct physical gestures, rather than through indirect point-and-click control. Note that the term "Organic" in OUI was derived from organic architecture, referring to the adoption of natural form to design a better fit with human ecology. The term also alludes to the use of organic electronics for this purpose.

In computing, 3D interaction is a form of human-machine interaction where users are able to move and perform interaction in 3D space. Both human and machine process information where the physical position of elements in the 3D space is relevant.

In computing, a natural user interface (NUI) or natural interface is a user interface that is effectively invisible, and remains invisible as the user continuously learns increasingly complex interactions. The word "natural" is used because most computer interfaces use artificial control devices whose operation has to be learned. Examples include voice assistants, such as Alexa and Siri, touch and multitouch interactions on today's mobile phones and tablets, but also touch interfaces invisibly integrated into the textiles furnitures.

<span class="mw-page-title-main">Human–computer interaction</span> Academic discipline studying the relationship between computer systems and their users

Human–computer interaction (HCI) is research in the design and the use of computer technology, which focuses on the interfaces between people (users) and computers. HCI researchers observe the ways humans interact with computers and design technologies that allow humans to interact with computers in novel ways. A device that allows interaction between human being and a computer is known as a "Human-computer Interface (HCI)".

<span class="mw-page-title-main">Finger tracking</span> High-resolution technique in gesture recognition and image processing

In the field of gesture recognition and image processing, finger tracking is a high-resolution technique developed in 1969 that is employed to know the consecutive position of the fingers of the user and hence represent objects in 3D. In addition to that, the finger tracking technique is used as a tool of the computer, acting as an external device in our computer, similar to a keyboard and a mouse.

<span class="mw-page-title-main">Omek Interactive</span> Former technology company

Omek Interactive was a venture-backed technology company developing advanced motion-sensing software for human-computer interaction. Omek was co-founded in 2007 by Janine Kutliroff and Gershom Kutliroff.

Project Digits is a Microsoft Research Project under Microsoft's computer science laboratory at the University of Cambridge; researchers from Newcastle University and University of Crete are also involved in this project. Project is led by David Kim a Microsoft Research PhD and also a PhD Student in computer science at Newcastle University. Digits is an input device which can be mounted on the wrist of human hand and it captures and displays a complete 3D graphical representation of the user's hand on screen without using any external sensing device or hand covering material like data gloves. This project aims to make gesture controlled interfaces completely hands free with greater mobility and accuracy. It allows user to interact with whatever hardware while moving from room to room or walking down the street without any line of sight connection with the hardware.

<span class="mw-page-title-main">Nadine Social Robot</span> Social Humanoid Robot

Nadine is a gynoid humanoid social robot that is modelled on Professor Nadia Magnenat Thalmann. The robot has a strong human-likeness with a natural-looking skin and hair and realistic hands. Nadine is a socially intelligent robot which returns a greeting, makes eye contact, and can remember all the conversations had with it. It is able to answer questions autonomously in several languages, simulate emotions both in gestures and facially, depending on the content of the interaction with the user. Nadine can recognise persons it has previously seen, and engage in flowing conversation. Nadine has been programmed with a "personality", in that its demeanour can change according to what is said to it. Nadine has a total of 27 degrees of freedom for facial expressions and upper body movements. With persons it has previously encountered, it remembers facts and events related to each person. It can assist people with special needs by reading stories, showing images, put on Skype sessions, send emails, and communicate with other members of the family. It can play the role of a receptionist in an office or be dedicated to be a personal coach.

<span class="mw-page-title-main">Shumin Zhai</span> Human–computer interaction research scientist

Shumin Zhai is a Chinese-born American Canadian Human–computer interaction (HCI) research scientist and inventor. He is known for his research specifically on input devices and interaction methods, swipe-gesture-based touchscreen keyboards, eye-tracking interfaces, and models of human performance in human-computer interaction. His studies have contributed to both foundational models and understandings of HCI and practical user interface designs and flagship products. He previously worked at IBM where he invented the ShapeWriter text entry method for smartphones, which is a predecessor to the modern Swype keyboard. Dr. Zhai's publications have won the ACM UIST Lasting Impact Award and the IEEE Computer Society Best Paper Award, among others, and he is most known for his research specifically on input devices and interaction methods, swipe-gesture-based touchscreen keyboards, eye-tracking interfaces, and models of human performance in human-computer interaction. Dr. Zhai is currently a Principal Scientist at Google where he leads and directs research, design, and development of human-device input methods and haptics systems.

Joseph J. LaViola Jr. is an American computer scientist, author, consultant, and academic. He holds the Charles N. Millican Professorship in Computer Science and leads the Interactive Computing Experiences Research Cluster at the University of Central Florida (UCF). He also serves as a visiting scholar in the Computer Science Department at Brown University, Consultant at JJL Interface Consultants as well as co-founder of Fluidity Software.

References

  1. 1 2 Kobylarz, Jhonatan; Bird, Jordan J.; Faria, Diego R.; Ribeiro, Eduardo Parente; Ekárt, Anikó (2020-03-07). "Thumbs up, thumbs down: non-verbal human-robot interaction through real-time EMG classification via inductive and supervised transductive transfer learning" (PDF). Journal of Ambient Intelligence and Humanized Computing. Springer Science and Business Media LLC. 11 (12): 6021–6031. doi: 10.1007/s12652-020-01852-z . ISSN   1868-5137.
  2. Matthias Rehm, Nikolaus Bee, Elisabeth André, Wave Like an Egyptian – Accelerometer Based Gesture Recognition for Culture Specific Interactions, British Computer Society, 2007
  3. "Patent Landscape Report Hand Gesture Recognition PatSeer Pro". PatSeer. Archived from the original on 2019-10-20. Retrieved 2017-11-02.
  4. Chai, Xiujuan, et al. "Sign language recognition and translation with kinect Archived 2021-01-10 at the Wayback Machine ." IEEE Conf. on AFGR. Vol. 655. 2013.
  5. Sultana A, Rajapuspha T (2012), "Vision Based Gesture Recognition for Alphabetical Hand Gestures Using the SVM Classifier" [ permanent dead link ], International Journal of Computer Science & Engineering Technology (IJCSET)., 2012
  6. Pavlovic, V., Sharma, R. & Huang, T. (1997), "Visual interpretation of hand gestures for human-computer interaction: A review", IEEE Transactions on Pattern Analysis and Machine Intelligence, July, 1997. Vol. 19(7), pp. 677 -695.
  7. R. Cipolla and A. Pentland, Computer Vision for Human-Machine Interaction, Cambridge University Press, 1998, ISBN   978-0-521-62253-0
  8. Ying Wu and Thomas S. Huang, "Vision-Based Gesture Recognition: A Review" Archived 2011-08-25 at the Wayback Machine , In: Gesture-Based Communication in Human-Computer Interaction, Volume 1739 of Springer Lecture Notes in Computer Science, pages 103-115, 1999, ISBN   978-3-540-66935-7, doi : 10.1007/3-540-46616-9
  9. Alejandro Jaimes and Nicu Sebe, Multimodal human–computer interaction: A survey Archived 2011-06-06 at the Wayback Machine , Computer Vision and Image Understanding Volume 108, Issues 1-2, October–November 2007, Pages 116-134 Special Issue on Vision for Human-Computer Interaction, doi : 10.1016/j.cviu.2006.10.019
  10. Dopertchouk, Oleg; "Recognition of Handwriting Gestures", gamedev.net, January 9, 2004
  11. Chen, Shijie; "Gesture Recognition Techniques in Handwriting Recognition Application", Frontiers in Handwriting Recognition p 142-147 November 2010
  12. Balaji, R; Deepu, V; Madhvanath, Sriganesh; Prabhakaran, Jayasree "Handwritten Gesture Recognition for Gesture Keyboard" Archived 2008-09-06 at the Wayback Machine , Hewlett-Packard Laboratories
  13. Dietrich Kammer, Mandy Keck, Georg Freitag, Markus Wacker, Taxonomy and Overview of Multi-touch Frameworks: Architecture, Scope, and Features Archived 2011-01-25 at the Wayback Machine
  14. "touchless user interface Definition from PC Magazine Encyclopedia". pcmag.com. Retrieved 2017-07-28.
  15. "The emerging need for touchless interaction technologies". ResearchGate. Retrieved 2021-06-30.
  16. S. Benford; H. Schnadelbach; B. Koleva; B. Gaver; A. Schmidt; A. Boucher; A. Steed; R. Anastasi; C. Greenhalgh; T. Rodden; H. Gellersen (2003). "Sensible, sensable and desirable: a framework for designing physical interfaces" (PDF). CiteSeerX   10.1.1.190.2504 . Archived from the original (PDF) on January 26, 2006.{{cite journal}}: Cite journal requires |journal= (help)
  17. Thomas G. Zimmerman, Jaron Lanier, Chuck Blanchard, Steve Bryson, and Young Harvill. http://portal.acm.org. "A HAND GESTURE INTERFACE DEVICE." http://portal.acm.org.
  18. Yang Liu, Yunde Jia, A Robust Hand Tracking and Gesture Recognition Method for Wearable Visual Interfaces and Its Applications, Proceedings of the Third International Conference on Image and Graphics (ICIG'04), 2004
  19. Kue-Bum Lee, Jung-Hyun Kim, Kwang-Seok Hong, An Implementation of Multi-Modal Game Interface Based on PDAs, Fifth International Conference on Software Engineering Research, Management and Applications, 2007
  20. "Gestigon Gesture Tracking - TechCrunch Disrupt". TechCrunch. Retrieved 11 October 2016.
  21. Matney, Lucas (29 August 2016). "uSens shows off new tracking sensors that aim to deliver richer experiences for mobile VR". TechCrunch. Retrieved 29 August 2016.
  22. Khalili, Abdullah; Soliman, Abdel-Hamid; Asaduzzaman, Md; Griffiths, Alison (March 2020). "Wi-Fi sensing: applications and challenges". The Journal of Engineering. 2020 (3): 87–97. arXiv: 1901.00715 . doi: 10.1049/joe.2019.0790 . ISSN   2051-3305.
  23. Per Malmestig, Sofie Sundberg, SignWiiver – implementation of sign language technology Archived 2008-12-25 at the Wayback Machine
  24. Thomas Schlomer, Benjamin Poppinga, Niels Henze, Susanne Boll, Gesture Recognition with a Wii Controller Archived 2013-07-27 at the Wayback Machine , Proceedings of the 2nd international Conference on Tangible and Embedded interaction, 2008
  25. AiLive Inc., LiveMove White Paper Archived 2007-07-13 at the Wayback Machine , 2006
  26. Electronic Design September 8, 2011. William Wong. Natural User Interface Employs Sensor Integration.
  27. Cable & Satellite International September/October, 2011. Stephen Cousins. A view to a thrill. Archived 2012-01-19 at the Wayback Machine
  28. TechJournal South January 7, 2008. Hillcrest Labs rings up $25M D round.
  29. Percussa AudioCubes Blog October 4, 2012. Gestural Control in Sound Synthesis. Archived 2015-09-10 at the Wayback Machine
  30. Mamtaz Alam; Dileep Kumar Tiwari (2016). "Gesture Recognization & its Applications". doi:10.13140/RG.2.2.28139.54563.{{cite journal}}: Cite journal requires |journal= (help)
  31. Vladimir I. Pavlovic, Rajeev Sharma, Thomas S. Huang, Visual Interpretation of Hand Gestures for Human-Computer Interaction; A Review, IEEE Transactions on Pattern Analysis and Machine Intelligence, 1997
  32. Ivan Laptev and Tony Lindeberg "Tracking of Multi-state Hand Models Using Particle Filtering and a Hierarchy of Multi-scale Image Features", Proceedings Scale-Space and Morphology in Computer Vision, Volume 2106 of Springer Lecture Notes in Computer Science, pages 63-74, Vancouver, BC, 1999. ISBN   978-3-540-42317-1, doi : 10.1007/3-540-47778-0
  33. von Hardenberg, Christian; Bérard, François (2001). "Bare-hand human-computer interaction". Proceedings of the 2001 workshop on Perceptive user interfaces. ACM International Conference Proceeding Series. Vol. 15 archive. Orlando, Florida. pp. 1–8. CiteSeerX   10.1.1.23.4541 .
  34. Lars Bretzner, Ivan Laptev, Tony Lindeberg "Hand gesture recognition using multi-scale colour features, hierarchical models and particle filtering", Proceedings of the Fifth IEEE International Conference on Automatic Face and Gesture Recognition, Washington, DC, USA, 21–21 May 2002, pages 423-428. ISBN   0-7695-1602-5, doi : 10.1109/AFGR.2002.1004190
  35. Domitilla Del Vecchio, Richard M. Murray Pietro Perona, "Decomposition of human motion into dynamics-based primitives with application to drawing tasks" Archived 2010-02-02 at the Wayback Machine , Automatica Volume 39, Issue 12, December 2003, Pages 2085–2098, doi : 10.1016/S0005-1098(03)00250-4.
  36. Thomas B. Moeslund and Lau Nørgaard, "A Brief Overview of Hand Gestures used in Wearable Human Computer Interfaces" Archived 2011-07-19 at the Wayback Machine , Technical report: CVMT 03-02, ISSN   1601-3646, Laboratory of Computer Vision and Media Technology, Aalborg University, Denmark.
  37. M. Kolsch and M. Turk "Fast 2D Hand Tracking with Flocks of Features and Multi-Cue Integration" Archived 2008-08-21 at the Wayback Machine , CVPRW '04. Proceedings Computer Vision and Pattern Recognition Workshop, May 27-June 2, 2004, doi : 10.1109/CVPR.2004.71
  38. Xia Liu Fujimura, K., "Hand gesture recognition using depth data", Proceedings of the Sixth IEEE International Conference on Automatic Face and Gesture Recognition, May 17–19, 2004 pages 529- 534, ISBN   0-7695-2122-3, doi : 10.1109/AFGR.2004.1301587.
  39. Stenger B, Thayananthan A, Torr PH, Cipolla R: "Model-based hand tracking using a hierarchical Bayesian filter", IEEE Transactions on IEEE Transactions on Pattern Analysis and Machine Intelligence, 28(9):1372-84, Sep 2006.
  40. A Erol, G Bebis, M Nicolescu, RD Boyle, X Twombly, "Vision-based hand pose estimation: A review", Computer Vision and Image Understanding Volume 108, Issues 1-2, October–November 2007, Pages 52-73 Special Issue on Vision for Human-Computer Interaction, doi : 10.1016/j.cviu.2006.10.012.
  41. 1 2 Rico, Julie; Brewster, Stephen (2010). "Usable gestures for mobile interfaces". Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. CHI '10. New York, NY, USA: ACM. pp. 887–896. doi:10.1145/1753326.1753458. ISBN   9781605589299. S2CID   16118067.
  42. 1 2 Walter, Robert; Bailly, Gilles; Müller, Jörg (2013). "StrikeAPose". Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. New York, New York, USA: ACM Press. pp. 841–850. doi:10.1145/2470654.2470774. ISBN   9781450318990. S2CID   2041073.
  43. 1 2 Profita, Halley P.; Clawson, James; Gilliland, Scott; Zeagler, Clint; Starner, Thad; Budd, Jim; Do, Ellen Yi-Luen (2013). "Don't mind me touching my wrist". Proceedings of the 2013 International Symposium on Wearable Computers. ISWC '13. New York, NY, USA: ACM. pp. 89–96. doi:10.1145/2493988.2494331. ISBN   9781450321273. S2CID   3236927.
  44. Harrison, Chris; Faste, Haakon (2014). "Implications of location and touch for on-body projected interfaces". Proceedings of the 2014 conference on Designing interactive systems. DIS '14. New York, NY, USA: ACM. pp. 543–552. doi:10.1145/2598510.2598587. ISBN   9781450329026. S2CID   1121501.
  45. 1 2 Reeves, Stuart; Benford, Steve; O'Malley, Claire; Fraser, Mike (2005). "Designing the spectator experience" (PDF). Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (PDF). New York, New York, USA: ACM Press. pp. 741–750. doi:10.1145/1054972.1055074. ISBN   978-1581139983. S2CID   5739231.
  46. Rupert Goodwins. "Windows 7? No arm in it". ZDNet.
  47. "gorilla arm". catb.org.
  48. Hincapié-Ramos, J.D., Guo, X., Moghadasian, P. and Irani. P. 2014. "Consumed Endurance: A Metric to Quantify Arm Fatigue of Mid-Air Interactions". In Proceedings of the 32nd annual ACM conference on Human factors in computing systems (CHI '14). ACM, New York, NY, USA, 1063–1072. DOI=10.1145/2556288.2557130
  49. Hincapié-Ramos, J.D., Guo, X., and Irani, P. 2014. "The Consumed Endurance Workbench: A Tool to Assess Arm Fatigue During Mid-Air Interactions". In Proceedings of the 2014 companion publication on Designing interactive systems (DIS Companion '14). ACM, New York, NY, USA, 109-112. DOI=10.1145/2598784.2602795