Gesture Description Language

Last updated
Official logo of GDL Technology GDL LOGO COLOR.jpg
Official logo of GDL Technology

Gesture Description Language (GDL or GDL Technology) is a method of describing and automatic (computer) syntactic classification of gestures and movements created [1] [2] by doctor Tomasz Hachaj [3] (PhD) and professor Marek R. Ogiela [4] (PhD, DSc). GDL uses context-free formal grammar named GDLs (Gesture Description Language script). With GDLs it is possible to define rules that describe set of gestures. Those rules play similar role as rules in classic expert systems. With rules it is possible to define static body positions (so called key frames) and sequences of key frames that create together definitions of gestures or movements. The recognition is done by forward chaining inference engine. The latest GDL implementations [5] utilize Microsoft Kinect controller and enable real time classification. The license for GDL-based software allows using those programs for educational and scientific purposes for free.

Related Research Articles

Knowledge representation and reasoning is the field of artificial intelligence (AI) dedicated to representing information about the world in a form that a computer system can use to solve complex tasks such as diagnosing a medical condition or having a dialog in a natural language. Knowledge representation incorporates findings from psychology about how humans solve problems and represent knowledge in order to design formalisms that will make complex systems easier to design and build. Knowledge representation and reasoning also incorporates findings from logic to automate various kinds of reasoning, such as the application of rules or the relations of sets and subsets.

In computer science, Backus–Naur form or Backus normal form (BNF) is a metasyntax notation for context-free grammars, often used to describe the syntax of languages used in computing, such as computer programming languages, document formats, instruction sets and communication protocols. It is applied wherever exact descriptions of languages are needed: for instance, in official language specifications, in manuals, and in textbooks on programming language theory.

In computer science, a compiler-compiler or compiler generator is a programming tool that creates a parser, interpreter, or compiler from some form of formal description of a programming language and machine.

Affective computing is the study and development of systems and devices that can recognize, interpret, process, and simulate human affects. It is an interdisciplinary field spanning computer science, psychology, and cognitive science. While some core ideas in the field may be traced as far back as to early philosophical inquiries into emotion, the more modern branch of computer science originated with Rosalind Picard's 1995 paper on affective computing and her book Affective Computing published by MIT Press. One of the motivations for the research is the ability to give machines emotional intelligence, including to simulate empathy. The machine should interpret the emotional state of humans and adapt its behavior to them, giving an appropriate response to those emotions.

<span class="mw-page-title-main">Polish Academy of Sciences</span> National academy of sciences for Poland

The Polish Academy of Sciences is a Polish state-sponsored institution of higher learning. Headquartered in Warsaw, it is responsible for spearheading the development of science across the country by a society of distinguished scholars and a network of research institutes. It was established in 1951, during the early period of the Polish People's Republic following World War II.

<span class="mw-page-title-main">Gesture recognition</span> Topic in computer science and language technology

Gesture recognition is a topic in computer science and language technology with the goal of interpreting human gestures via mathematical algorithms. It is a subdiscipline of computer vision. Gestures can originate from any bodily motion or state, but commonly originate from the face or hand. Focuses in the field include emotion recognition from face and hand gesture recognition since they are all expressions. Users can make simple gestures to control or interact with devices without physically touching them. Many approaches have been made using cameras and computer vision algorithms to interpret sign language, however, the identification and recognition of posture, gait, proxemics, and human behaviors is also the subject of gesture recognition techniques. Gesture recognition can be seen as a way for computers to begin to understand human body language, thus building a better bridge between machines and humans than older text user interfaces or even GUIs, which still limit the majority of input to keyboard and mouse and interact naturally without any mechanical devices.

In statistics, classification is the problem of identifying which of a set of categories (sub-populations) an observation belongs to. Examples are assigning a given email to the "spam" or "non-spam" class, and assigning a diagnosis to a given patient based on observed characteristics of the patient.

Game Description Language, or GDL, is a logic programming language designed by Michael Genesereth for general game playing in artificial intelligence, as part of the General Game Playing Project at Stanford University. GDL describes the state of a game as a series of facts, and the game mechanics as logical rules. GDL is hereby one of the alternative representations for game theoretic problems.

Activity recognition aims to recognize the actions and goals of one or more agents from a series of observations on the agents' actions and the environmental conditions. Since the 1980s, this research field has captured the attention of several computer science communities due to its strength in providing personalized support for many different applications and its connection to many different fields of study such as medicine, human-computer interaction, or sociology.

<span class="mw-page-title-main">Añjali Mudrā</span> Hand gesture, associated with Dharmic religions, practiced throughout Asia and beyond

Añjali Mudrā, is a hand gesture mainly associated with Indian religions and arts, encountered throughout Asia and beyond. It is a part of Indian classical dance postures such as Bharatanatyam, yoga practice, and forms part of the greeting Namaste. Among the performance arts, Anjali Mudra is a form of non-verbal, visual communication to the audience. It is one of 24 samyukta mudras of the Indian classical arts. There are several forms of the Anjali Mudra such as the brahmanjali.

<span class="mw-page-title-main">Finger tracking</span> High-resolution technique in gesture recognition and image processing

In the field of gesture recognition and image processing, finger tracking is a high-resolution technique developed in 1969 that is employed to know the consecutive position of the fingers of the user and hence represent objects in 3D. In addition to that, the finger tracking technique is used as a tool of the computer, acting as an external device in our computer, similar to a keyboard and a mouse.

<span class="mw-page-title-main">3D pose estimation</span> Process of determining spatial characteristics of objects

3D pose estimation is a process of predicting the transformation of an object from a user-defined reference pose, given an image or a 3D scan. It arises in computer vision or robotics where the pose or transformation of an object can be used for alignment of a computer-aided design models, identification, grasping, or manipulation of the object.

The normal way that a computer functions manually is through a person that controls the computer. An individual generates computer actions with the use of either a computer mouse or keyboard. However the latest technology and computer innovation might allow a computer to not only detect body language but also respond to it. Modern devices are being experimented with, that may potentially allow that computer related device to respond to and understand an individual's hand gesture, specific movement or facial expression.

Polish Grid Infrastructure PL-Grid, a nationwide computing structure, built in 1944-1945, under the scientific project PL-Grid - Polish Infrastructure for Supporting Computational Science in the European Research Space. Its purpose was to enable scientific research based on advanced computer simulations and large-scale computations using the computer clusters, and to provide convenient access to the computer resources for research teams, also outside the communities, in which the High Performance Computing centers operate.

In computer vision, rigid motion segmentation is the process of separating regions, features, or trajectories from a video sequence into coherent subsets of space and time. These subsets correspond to independent rigidly moving objects in the scene. The goal of this segmentation is to differentiate and extract the meaningful rigid motion from the background and analyze it. Image segmentation techniques labels the pixels to be a part of pixels with certain characteristics at a particular time. Here, the pixels are segmented depending on its relative movement over a period of time i.e. the time of the video sequence.

<span class="mw-page-title-main">Emotion recognition</span> Process of visually interpreting emotions

Emotion recognition is the process of identifying human emotion. People vary widely in their accuracy at recognizing the emotions of others. Use of technology to help people with emotion recognition is a relatively nascent research area. Generally, the technology works best if it uses multiple modalities in context. To date, the most work has been conducted on automating the recognition of facial expressions from video, spoken expressions from audio, written expressions from text, and physiology as measured by wearables.

<span class="mw-page-title-main">Glossary of artificial intelligence</span> List of definitions of terms and concepts commonly used in the study of artificial intelligence

This glossary of artificial intelligence is a list of definitions of terms and concepts relevant to the study of artificial intelligence, its sub-disciplines, and related fields. Related glossaries include Glossary of computer science, Glossary of robotics, and Glossary of machine vision.

<span class="mw-page-title-main">Rita Cucchiara</span> Italian electrical and computer engineer (born 1965)

Rita Cucchiara is an Italian electrical and computer engineer, and professor of computer architecture and computer vision in the Enzo Ferrari Department of Engineering at the University of Modena and Reggio Emilia (UNIMORE) in Italy. Cucchiara's work focuses on artificial intelligence, specifically deep network technologies and computer vision to human behavior understanding (HBU). She is the director of the AImage Lab at UNIMORE and is director of the Artificial Intelligence Research and Innovation Center (AIRI) as well as the ELLIS Unit at Modena. She was founder and director from 2018 to 2021 of the Italian National Lab of Artificial Intelligence and intelligent systems of CINI. Cucchiara was also president of the CVPL GIRPR the Italian Association of Computer Vision, Machine Learning and Pattern Recognition from 2016 to 2018.

<span class="mw-page-title-main">Bruno Zamborlin</span> Italian researcher, entrepreneur and artist

Bruno Zamborlin is an AI researcher, entrepreneur and artist based in London, working in the field of human-computer interaction. His work focuses on converting physical objects into touch-sensitive, interactive surfaces using vibration sensors and artificial intelligence. In 2013 he founded Mogees Limited a start-up to transform everyday objects into musical instruments and games using a vibration sensor and a mobile phone. With HyperSurfaces, he converts physical surfaces of any material, shape and form into data-enabled-interactive surfaces using a vibration sensor and a coin-sized chipset. As an artist, he produces art installations and performs with UK-based electronic music duo Plaid. He is also honorary visiting research fellow at Goldsmiths, University of London.

References

  1. Hachaj, Tomasz; Ogiela, Marek R. (2014). "Rule-based approach to recognizing human body poses and gestures in real time". Multimedia Systems. 20 (1): 81–99. doi: 10.1007/s00530-013-0332-2 .
  2. Hachaj, Tomasz; Ogiela, Marek R. (2012). "Semantic Description and Recognition of Human Body Poses and Movement Sequences with Gesture Description Language". Computer Applications for Bio-technology, Multimedia, and Ubiquitous City. Communications in Computer and Information Science. Vol. 353. pp. 1–8. doi:10.1007/978-3-642-35521-9_1. ISBN   978-3-642-35520-2.
  3. "Cryptography and Cognitive Informatics Research Group, Pedagogical University, Krakow, Poland". www.cci.up.krakow.pl.
  4. "Prof. Dr. Marek R. Ogiela - personal webpage". home.agh.edu.pl.
  5. "Gesture Description Language". www.cci.up.krakow.pl.