Scratch input

Last updated
Scratch input using geophones (contact microphones) bonded to each fingernail as a musical instrument. The awful sounds generated by running fingernails down the chalkboard are separately pitch-corrected, resulting in a pleasant-sounding musical instrument. ScratchInput SteveMann self portrait.jpg
Scratch input using geophones (contact microphones) bonded to each fingernail as a musical instrument. The awful sounds generated by running fingernails down the chalkboard are separately pitch-corrected, resulting in a pleasant-sounding musical instrument.

In computing, scratch input is an acoustic-based method of Human-Computer Interaction (HCI) that takes advantage of the characteristic sound produced when a finger nail or other object is dragged over a surface, such as a table or wall. The technique is not limited to fingers; a stick or writing implements (e.g. chalk, or a pen) can also be used. The sound is often inaudible to the naked ear (i.e., silent). However, specialized microphones can digitize the sounds for interactive purposes. Scratch input was invented by Mann et al. in 2007, [1] [2] [3] though the term was first used by Chris Harrison et al. [4]

Contents

History

A natural interface for musical expression operating on scratch input principles was first published and presented in June 2007. [1] Later that year, it was extended to an implementation on a smartphone and also a wearable computer system. [2]

In 2008, the Scratch Input project [4] demonstrated a mobile device input system utilizing scratch input, simultaneously popularizing the term. [5] This system captured audio transmitted through a surface on which a mobile phone was placed, enabling the entire surface to be used as an input device.

Uses

Scratch input is an enabling input technique that is used in multitude of applications. The earliest application was a highly expressive musical instrument (Mann et al.) for use with mobile devices on natural objects, surfaces, or the like, as a non-synthesizing (i.e. idiophonic) musical instrument. Harrison et al. [4] proposed it to create large, ad hoc gestural input areas when mobile devices are rested on tables.

Scratching a stick on bricks and mortar, with vision-assisted gesture recognition. This gives better results than scratch input alone. (Public performance in Toronto, Ontario; image from the 2007 paper.) Scratch Input musical cyborg.jpg
Scratching a stick on bricks and mortar, with vision-assisted gesture recognition. This gives better results than scratch input alone. (Public performance in Toronto, Ontario; image from the 2007 paper.)
Scratch input with ice skates is called a pagophone. Hyperacoustic Mann etal.jpg
Scratch input with ice skates is called a pagophone.

Commercial potential

Microsoft has expressed interest in Scratch Input. [5]

With this we can start to think of every flat surface as an potential input area. If mass produced this sensor could cost less than a dollar....Despite the limitations, the technology holds enough promise to make it into the hands of consumers. It is exciting because it is so low cost. This idea has the potential to go beyond just a research project.

Daniel Wigdor, User experience architect at Microsoft and curator of the emerging technology demos at SIGGRAPH

See also

Related Research Articles

Pointing device Human interface device for computers

A pointing device is a human interface device that allows a user to input spatial data to a computer. CAD systems and graphical user interfaces (GUI) allow the user to control and provide data to the computer using physical gestures by moving a hand-held mouse or similar device across the surface of the physical desktop and activating switches on the mouse. Movements of the pointing device are echoed on the screen by movements of the pointer and other visual changes. Common gestures are point and click and drag and drop.

Steve Mann (inventor) Professor and wearable computing researcher

William Stephen George Mann is a Canadian engineer, professor, and inventor who works in augmented reality, computational photography, particularly wearable computing, and high-dynamic-range imaging. Mann is sometimes labeled the "Father of Wearable Computing" for early inventions and continuing contributions to the field. He cofounded InteraXon, makers of the Muse brain-sensing headband, and is also a founding member of the IEEE Council on Extended Intelligence (CXI). Mann is currently CTO and cofounder at Blueberry X Technologies and Chairman of MannLab. Mann was born in Canada, and currently lives in Toronto, Canada, with his wife and two children.

Computer-mediated reality Ability to manipulate ones perception of reality through the use of a computer

Computer-mediated reality refers to the ability to add to, subtract information from, or otherwise manipulate one's perception of reality through the use of a wearable computer or hand-held device such as a smartphone.

Touchscreen Input and output device

A touchscreen or touch screen is the assembly of both an input and output ('display') device. The touch panel is normally layered on the top of an electronic visual display of an information processing system. The display is often an LCD, AMOLED or OLED display while the system is usually a laptop, tablet, or smartphone. A user can give input or control the information processing system through simple or multi-touch gestures by touching the screen with a special stylus or one or more fingers. Some touchscreens use ordinary or specially coated gloves to work while others may only work using a special stylus or pen. The user can use the touchscreen to react to what is displayed and, if the software allows, to control how it is displayed; for example, zooming to increase the text size.

Handheld projector Image projector in a handheld device

A handheld projector is an image projector in a handheld device. It was developed as a computer display device for compact portable devices such as mobile phones, personal digital assistants, and digital cameras, which have sufficient storage capacity to handle presentation materials but are too small to accommodate a display screen that an audience can see easily. Handheld projectors involve miniaturized hardware, and software that can project digital images onto a nearby viewing surface.

Gesture recognition Topic in computer science and language technology

Gesture recognition is a topic in computer science and language technology with the goal of interpreting human gestures via mathematical algorithms. It is a subdiscipline of computer vision. Gestures can originate from any bodily motion or state but commonly originate from the face or hand. Current focuses in the field include emotion recognition from face and hand gesture recognition. Users can use simple gestures to control or interact with devices without physically touching them. Many approaches have been made using cameras and computer vision algorithms to interpret sign language. However, the identification and recognition of posture, gait, proxemics, and human behaviors is also the subject of gesture recognition techniques. Gesture recognition can be seen as a way for computers to begin to understand human body language, thus building a richer bridge between machines and humans than primitive text user interfaces or even GUIs, which still limit the majority of input to keyboard and mouse and interact naturally without any mechanical devices.

New Interfaces for Musical Expression International conference

New Interfaces for Musical Expression, also known as NIME, is an international conference dedicated to scientific research on the development of new technologies and their role in musical expression and artistic performance.

A voice-user interface (VUI) makes spoken human interaction with computers possible, using speech recognition to understand spoken commands and answer questions, and typically text to speech to play a reply. A voice command device (VCD) is a device controlled with a voice user interface.

Multi-touch Technology

In computing, multi-touch is technology that enables a surface to recognize the presence of more than one point of contact with the surface at the same time. The origins of multitouch began at CERN, MIT, University of Toronto, Carnegie Mellon University and Bell Labs in the 1970s. CERN started using multi-touch screens as early as 1976 for the controls of the Super Proton Synchrotron. Capacitive multi-touch displays were popularized by Apple's iPhone in 2007. Plural-point awareness may be used to implement additional functionality, such as pinch to zoom or to activate certain subroutines attached to predefined gestures.

A text entry interface or text entry device is an interface that is used to enter text information an electronic device. A commonly used device is a mechanical computer keyboard. Most laptop computers have an integrated mechanical keyboard, and desktop computers are usually operated primarily using a keyboard and mouse. Devices such as smartphones and tablets mean that interfaces such as virtual keyboards and voice recognition are becoming more popular as text entry systems.

Hydraulophone Hydraulic musical instrument

A hydraulophone is a tonal acoustic musical instrument played by direct physical contact with water where sound is generated or affected hydraulically. The hydraulophone was described and named by Steve Mann in 2005, and patented in 2011. Typically, sound is produced by the same hydraulic fluid in contact with the player's fingers. It has been used as a sensory exploration device for low-vision individuals.

In electrical engineering, capacitive sensing is a technology, based on capacitive coupling, that can detect and measure anything that is conductive or has a dielectric different from air. Many types of sensors use capacitive sensing, including sensors to detect and measure proximity, pressure, position and displacement, force, humidity, fluid level, and acceleration. Human interface devices based on capacitive sensing, such as touchpads, can replace the computer mouse. Digital audio players, mobile phones, and tablet computers use capacitive sensing touchscreens as input devices. Capacitive sensors can also replace mechanical buttons.

SixthSense Gesture-based wearable computer system

SixthSense is a gesture-based wearable computer system developed at MIT Media Lab by Steve Mann in 1994 and 1997, and 1998, and further developed by Pranav Mistry, in 2009, both of whom developed both hardware and software for both headworn and neckworn versions of it. It comprises a headworn or neck-worn pendant that contains both a data projector and camera. Headworn versions were built at MIT Media Lab in 1997 that combined cameras and illumination systems for interactive photographic art, and also included gesture recognition.

In computing, a natural user interface, or NUI, or natural interface is a user interface that is effectively invisible, and remains invisible as the user continuously learns increasingly complex interactions. The word "natural" is used because most computer interfaces use artificial control devices whose operation has to be learned. Examples include voice assistants, such as Alexa and Siri, touch and multitouch interactions on today's mobile phones and tablets, but also touch interfaces invisibly integrated into the textiles furnitures.

Skinput Input technology

Skinput is an input technology that uses bio-acoustic sensing to localize finger taps on the skin. When augmented with a pico-projector, the device can provide a direct manipulation, graphical user interface on the body. The technology was developed by Chris Harrison, Desney Tan, and Dan Morris, at Microsoft Research's Computational User Experiences Group. Skinput represents one way to decouple input from electronic devices with the aim of allowing devices to become smaller without simultaneously shrinking the surface area on which input can be performed. While other systems, like SixthSense have attempted this with computer vision, Skinput employs acoustics, which take advantage of the human body's natural sound conductive properties. This allows the body to be annexed as an input surface without the need for the skin to be invasively instrumented with sensors, tracking markers, or other items.

Sergi Jordà is a Catalan innovator, installation artist, digital musician and Associate Professor at the Music Technology Group, Universitat Pompeu Fabra in Barcelona. He is best known for directing the team that invented the Reactable. He is also a trained Physicist.

An immersive virtual musical instrument, or immersive virtual environment for music and sound, represents sound processes and their parameters as 3D entities of a virtual reality so that they can be perceived not only through auditory feedback but also visually in 3D and possibly through tactile as well as haptic feedback, using 3D interface metaphors consisting of interaction techniques such as navigation, selection and manipulation (NSM). It builds on the trend in electronic musical instruments to develop new ways to control sound and perform music such as explored in conferences like NIME.

Absement Measure of sustained displacement of an object from its initial position

In kinematics, absement is a measure of sustained displacement of an object from its initial position, i.e. a measure of how far away and for how long. The word absement is a portmanteau of the words absence and displacement. Similarly, absition is a portmanteau of the words absence and position.

Chris Harrison (computer scientist)

Chris Harrison is a British-born, American computer scientist and entrepreneur, working in the fields of human–computer interaction, machine learning and sensor-driven interactive systems. He is a professor at Carnegie Mellon University and director of the Future Interfaces Group within the Human–Computer Interaction Institute. He has previously conducted research at AT&T Labs, Microsoft Research, IBM Research and Disney Research. He is also the CTO and co-founder of Qeexo, a machine learning and interaction technology startup.

Bruno Zamborlin Italian researcher, entrepreneur and artist

Bruno Zamborlin is an AI researcher, entrepreneur and artist based in London, working in the field of human-computer interaction. His work focuses on converting physical objects into touch-sensitive, interactive surfaces using vibration sensors and artificial intelligence. In 2013 he founded Mogees Limited a start-up to transform everyday objects into musical instruments and games using a vibration sensor and a mobile phone. In 2017 he founded HyperSurfaces, his second startup, that converts physical surfaces of any material, shape and form into data-enabled-interactive surfaces using a vibration sensor and a coin-sized chipset. As an artist, he produces art installations and performs with UK-based electronic music duo Plaid. He is also honorary visiting research fellow at Goldsmiths, University of London.

References

  1. 1 2 Steve Mann, "Natural Interfaces for Musical Expression: Physiphones and a physics-based organology", Proceedings of the 2007 Conference on New Interfaces for Musical Expression (NIME07), June 6–10, New York, NY, USA.
  2. 1 2 Steve Mann, Ryan Janzen, Raymond Lo, and Chris Aimone, "Inventing new instruments based on a computational `hack' to make a badly tuned or unpitched instrument play in perfect harmony", Proceedings of the 2007 International Computer Music Conference (ICMC2007), August 27–31, Copenhagen.
  3. 1 2 Steve Mann, Ryan Janzen and Raymond Lo, "Hyperacoustic instruments: Computer-controlled instruments that are not electrophones", Proc. International IEEE conference (ICME 2008), Hannover, Germany, June 23–26, 2008.
  4. 1 2 3 Harrison, Chris and Hudson, Scott E. "Scratch Input: Creating Large, Inexpensive, Unpowered and Mobile finger Input Surfaces". In Proceedings of the 21st Annual ACM Symposium on User interface Software and Technology. UIST '08. ACM, New York, NY, 205-208
  5. 1 2 Priya Ganapati (7 August 2009). "To Answer the Phone, Scratch Your Jeans". WIRED.