![]() | This article may contain an excessive number of citations .(November 2023) |
Arnav Kapur is a computer scientist and engineer at MIT. He is known for his work and advocacy in developing AI systems that enhance and complement human abilities.
In 2016, Kapur worked at the Biomedical Cybernetics Laboratory at Harvard Medical School. While at Harvard, he showed that gene expression data in microarray and RNA-Seq experiments, could be considered approximately low-rank, which could then be used to reliably predict the data. [1] [2] [3] [4]
After this, Kapur attended Massachusetts Institute of Technology, working at the MIT Media Lab. [5] [6] [7] While at MIT, he developed a peripheral neural interface, dubbed AlterEgo, that recorded neuromuscular signals, sent from the brain to the various muscles of the speech system through subtle and voluntary stimulation, and transcribed them into basic speech commands. [5] [8] [9] [10] [11] This created a silent speech interface, having applications in facilitating speech for individuals who had lost the ability to communicate verbally. [12] [13] [14] [15] [16] [17] In 2018, Kapur in an interview with 60 Minutes, detailed the inner workings of the prototype and showcased a live demonstration. [18] [19] [20] [21] In 2019, Kapur gave a talk at TED, and demonstrated an updated version of the system. [22] [23] [24] [25] [26] He talked about how ethics can inform design and engineering as a principle [24] [27] [28] [29] [30] and advocated for technologies such as AI to be designed in a way that extended human capabilities. [29] [31] [32] [33]
Kapur has also exhibited artwork and AI tools that collaborate with human artists. [34] His work has been exhibited at alt.ai New York, Design Museum, Art Center Nabi. [35] [36] [37] [38] [39]
In 2020, Kapur featured on TIME magazine's 100 Best Inventions of 2020. [40] He has been awarded the Lemelson-MIT graduate prize. [41]
Speech synthesis is the artificial production of human speech. A computer system used for this purpose is called a speech synthesizer, and can be implemented in software or hardware products. A text-to-speech (TTS) system converts normal language text into speech; other systems render symbolic linguistic representations like phonetic transcriptions into speech. The reverse process is speech recognition.
The Lemelson–MIT Program awards several prizes yearly to inventors in the United States. The largest is the Lemelson–MIT Prize which was endowed in 1994 by Jerome H. Lemelson, funded by the Lemelson Foundation, and is administered through the School of Engineering at the Massachusetts Institute of Technology. The winner receives $500,000, making it the largest cash prize for invention in the U.S.
The MIT Media Lab is a research laboratory at the Massachusetts Institute of Technology, growing out of MIT's Architecture Machine Group in the School of Architecture. Its research does not restrict to fixed academic disciplines, but draws from technology, media, science, art, and design. As of 2014, Media lab's research groups include neurobiology, biologically inspired fabrication, socially engaging robots, emotive computing, bionics, and hyperinstruments.
An alter ego is an alternate personality or persona.
Subvocal recognition (SVR) is the process of taking subvocalization and converting the detected results to a digital output, aural or text-based. A silent speech interface is a device that allows speech communication without using the sound made when people vocalize their speech sounds. It works by the computer identifying the phonemes that an individual pronounces from nonauditory sources of information about their speech movements. These are then used to recreate the speech using speech synthesis.
Intelligence amplification (IA) refers to the effective use of information technology in augmenting human intelligence. The idea was first proposed in the 1950s and 1960s by cybernetics and early computer pioneers.
Rosalind Wright Picard is an American scholar and inventor who is Professor of Media Arts and Sciences at MIT, founder and director of the Affective Computing Research Group at the MIT Media Lab, and co-founder of the startups Affectiva and Empatica.
NeuroSky, Inc. is a manufacturer of brain-computer interface (BCI) technologies for consumer product applications, which was founded in 2004 in Silicon Valley, California. The company adapts electroencephalography (EEG) and electromyography (EMG) technology to fit a consumer market within a number of fields such as entertainment, education, automotive, and health.
Gary Fred Marcus is an American psychologist, cognitive scientist, and author, known for his research on the intersection of cognitive psychology, neuroscience, and artificial intelligence (AI).
Wearable technology is any technology that is designed to be used while worn. Common types of wearable technology include smartwatches and smartglasses. Wearable electronic devices are often close to or on the surface of the skin, where they detect, analyze, and transmit information such as vital signs, and/or ambient data and which allow in some cases immediate biofeedback to the wearer.
Neurowear is a gadget project organization in Japan founded on the concept of the "Augmented Human Body". The group's first project, known as Necomimi is a headband with a brain wave sensor and motorized cat shaped ears programmed to turn up or down based on the wearer's electroencephalogram influenced by "thoughts and emotions". neurowear collaborated with Qosmo and Daito Manabe on "unboxxx" exhibition in July 2012 at Gallery KATA Ebisu.
Ramesh Raskar is a Massachusetts Institute of Technology associate professor and head of the MIT Media Lab's Camera Culture research group. Previously he worked as a senior research scientist at Mitsubishi Electric Research Laboratories (MERL) during 2002 to 2008. He holds 132 patents in computer vision, computational health, sensors and imaging. He received the $500K Lemelson–MIT Prize in 2016. The prize money will be used for launching REDX.io, a group platform for co-innovation in Artificial Intelligence. He is well known for inventing EyeNetra, EyeCatra and EyeSelfie, Femto-photography and his TED talk for cameras to see around corners.
Google Brain was a deep learning artificial intelligence research team that served as the sole AI branch of Google before being incorporated under the newer umbrella of Google AI, a research division at Google dedicated to artificial intelligence. Formed in 2011, it combined open-ended machine learning research with information systems and large-scale computing resources. It created tools such as TensorFlow, which allow neural networks to be used by the public, and multiple internal AI research projects, and aimed to create research opportunities in machine learning and natural language processing. It was merged into former Google sister company DeepMind to form Google DeepMind in April 2023.
Windows Mixed Reality (WMR) is a discontinued platform by Microsoft which provides augmented reality and virtual reality experiences with compatible head-mounted displays.
Catalin Voss is a German-born inventor and entrepreneur. Voss is considered a pioneer in applying artificial intelligence and machine learning for societal impact in areas such as childhood literacy, Autism, financial inclusion in emerging economies, and the criminal justice system.
The machine translation of sign languages has been possible, albeit in a limited fashion, since 1977. When a research project successfully matched English letters from a keyboard to ASL manual alphabet letters which were simulated on a robotic hand. These technologies translate signed languages into written or spoken language, and written or spoken language to sign language, without the use of a human interpreter. Sign languages possess different phonological features than spoken languages, which has created obstacles for developers. Developers use computer vision and machine learning to recognize specific phonological parameters and epentheses unique to sign languages, and speech recognition and natural language processing allow interactive communication between hearing and deaf people.
AlterEgo is a wearable silent speech output-input device developed by MIT Media Lab. The device is attached around the head, neck, and jawline and translates your brain speech center impulse input into words on a computer, without vocalization.
The first-generation Oculus Quest is a discontinued virtual reality headset developed by Oculus, a brand of Facebook Inc., and released on May 21, 2019. Similar to its predecessor, Oculus Go, it is a standalone device, that can run games and software wirelessly under an Android-based operating system. It supports positional tracking with six degrees of freedom, using internal sensors and an array of cameras in the front of the headset rather than external sensors. The cameras are also used as part of the safety feature "Passthrough", which shows a view from the cameras when the user exits their designated boundary area known as "Guardian". A later software update added "Oculus Link", a feature that allows the Quest to be connected to a computer via USB, enabling use with Oculus Rift-compatible software and games.
This article needs additional or more specific categories .(November 2023) |