Virtual assistant

Last updated

Google Assistant running on a Pixel XL smartphone Android Assistant on the Google Pixel XL smartphone (29526761674).jpg
Google Assistant running on a Pixel XL smartphone

A virtual assistant (VA) is a software agent that can perform a range of tasks or services for a user based on user input such as commands or questions, including verbal ones. Such technologies often incorporate chatbot capabilities to simulate human conversation, such as via online chat, to facilitate interaction with their users. The interaction may be via text, graphical interface, or voice - as some virtual assistants are able to interpret human speech and respond via synthesized voices.

Contents

In many cases users can ask their virtual assistants questions, control home automation devices and media playback, and manage other basic tasks such as email, to-do lists, and calendars - all with verbal commands. [1] In recent years, prominent virtual assistants for direct consumer use have included Apple's Siri, Amazon Alexa, Google Assistant, and Samsung's Bixby. [2] Also, companies in various industries often incorporate some kind of virtual assistant technology into their customer service or support. [3]

Recently, the emergence of recent artificial intelligence based chatbots, such as ChatGPT, has brought increased capability and interest to the field of virtual assistant products and services. [4] [5] [6]

History

Experimental decades: 1910s–1980s

Radio Rex was the first voice activated toy, patented in 1916 [7] and released in 1922. [8] It was a wooden toy in the shape of a dog that would come out of its house when its name is called.

In 1952, Bell Labs presented "Audrey", the Automatic Digit Recognition machine. It occupied a six- foot-high relay rack, consumed substantial power, had streams of cables and exhibited the myriad maintenance problems associated with complex vacuum-tube circuitry. It could recognize the fundamental units of speech, phonemes. It was limited to accurate recognition of digits spoken by designated talkers. It could therefore be used for voice dialing, but in most cases push-button dialing was cheaper and faster, rather than speaking the consecutive digits. [9]

Another early tool which was enabled to perform digital speech recognition was the IBM Shoebox voice-activated calculator, presented to the general public during the 1962 Seattle World's Fair after its initial market launch in 1961. This early computer, developed almost 20 years before the introduction of the first IBM Personal Computer in 1981, was able to recognize 16 spoken words and the digits 0 to 9.

The first natural language processing computer program or the chatbot ELIZA was developed by MIT professor Joseph Weizenbaum in the 1960s. It was created to "demonstrate that the communication between man and machine was superficial". [10] ELIZA used pattern matching and substitution methodology into scripted responses to simulate conversation, which gave an illusion of understanding on the part of the program.

Weizenbaum's own secretary reportedly asked Weizenbaum to leave the room so that she and ELIZA could have a real conversation. Weizenbaum was surprised by this, later writing: "I had not realized ... that extremely short exposures to a relatively simple computer program could induce powerful delusional thinking in quite normal people. [11]

This gave name to the ELIZA effect, the tendency to unconsciously assume computer behaviors are analogous to human behaviors; that is, anthropomorphisation, a phenomenon present in human interactions with virtual assistants.

The next milestone in the development of voice recognition technology was achieved in the 1970s at the Carnegie Mellon University in Pittsburgh, Pennsylvania with substantial support of the United States Department of Defense and its DARPA agency, funded five years of a Speech Understanding Research program, aiming to reach a minimum vocabulary of 1,000 words. Companies and academia including IBM, Carnegie Mellon University (CMU) and Stanford Research Institute took part in the program.

The result was "Harpy", it mastered about 1000 words, the vocabulary of a three-year-old and it could understand sentences. It could process speech that followed pre-programmed vocabulary, pronunciation, and grammar structures to determine which sequences of words made sense together, and thus reducing speech recognition errors.

In 1986 Tangora was an upgrade of the Shoebox, it was a voice recognizing typewriter. Named after the world's fastest typist at the time, it had a vocabulary of 20,000 words and used prediction to decide the most likely result based on what was said in the past. IBM's approach was based on a hidden Markov model, which adds statistics to digital signal processing techniques. The method makes it possible to predict the most likely phonemes to follow a given phoneme. Still each speaker had to individually train the typewriter to recognize his or her voice, and pause between each word.

Birth of smart virtual assistants: 1990s–2010s

In the 1990s, digital speech recognition technology became a feature of the personal computer with IBM, Philips and Lernout & Hauspie fighting for customers. Much later the market launch of the first smartphone IBM Simon in 1994 laid the foundation for smart virtual assistants as we know them today.[ citation needed ]

In 1997, Dragon's Naturally Speaking software could recognize and transcribe natural human speech without pauses between each word into a document at a rate of 100 words per minute. A version of Naturally Speaking is still available for download and it is still used today, for instance, by many doctors in the US and the UK to document their medical records.[ citation needed ]

In 2001 Colloquis publicly launched SmarterChild, on platforms like AIM and MSN Messenger. While entirely text-based SmarterChild was able to play games, check the weather, look up facts, and converse with users to an extent. [12]

The first modern digital virtual assistant installed on a smartphone was Siri, which was introduced as a feature of the iPhone 4S on 4 October 2011. [13] Apple Inc. developed Siri following the 2010 acquisition of Siri Inc., a spin-off of SRI International, which is a research institute financed by DARPA and the United States Department of Defense. [14] Its aim was to aid in tasks such as sending a text message, making phone calls, checking the weather or setting up an alarm. Over time, it has developed to provide restaurant recommendations, search the internet, and provide driving directions.[ citation needed ]

In November 2014, Amazon announced Alexa alongside the Echo. [15]

In April 2017 Amazon released a service for building conversational interfaces for any type of virtual assistant or interface.

Artificial intelligence and language models: 2020s-present

In the 2020s, artificial intelligence (AI) systems like ChatGPT have gained popularity for their ability to generate human-like responses to text-based conversations. In February 2020, Microsoft introduced its Turing Natural Language Generation (T-NLG), which was then the "largest language model ever published at 17 billion parameters." [16] On November 30, 2022, ChatGPT was launched as a prototype and quickly garnered attention for its detailed responses and articulate answers across many domains of knowledge. The advent of ChatGPT and its introduction to the wider public increased interest and competition in the space. In February 2023, Google began introducing an experimental service called "Bard" which is based on its LaMDA program to generate text responses to questions asked based on information gathered from the web.

While ChatGPT and other generalized chatbots based on the latest generative AI are capable of performing various tasks associated with virtual assistants, there are also more specialized forms of such technology that are designed to target more specific situations or needs. [17] [4]

Method of interaction

Amazon Echo Dot smart speaker running the Alexa virtual assistant Amazon Echo Dot (27716286638).jpg
Amazon Echo Dot smart speaker running the Alexa virtual assistant

Virtual assistants work via:

Many virtual assistants are accessible via multiple methods, offering versatility in how users can interact with them, whether through chat, voice commands, or other integrated technologies.

Virtual assistants use natural language processing (NLP) to match user text or voice input to executable commands. Some continually learn using artificial intelligence techniques including machine learning and ambient intelligence.

To activate a virtual assistant using the voice, a wake word might be used. This is a word or groups of words such as "Hey Siri", "OK Google" or "Hey Google", "Alexa", and "Hey Microsoft". [20] As virtual assistants become more popular, there are increasing legal risks involved. [21] :815

Devices and objects

Apple TV remote control, with which users can ask Siri the virtual assistant to find content to watch Apple tv gen 4 remote.jpeg
Apple TV remote control, with which users can ask Siri the virtual assistant to find content to watch

Virtual assistants may be integrated into many types of platforms or, like Amazon Alexa, across several of them:

Services

Virtual assistants can provide a wide variety of services. These include: [29]

Conversational commerce

Conversational commerce is e-commerce via various means of messaging, including via voice assistants [32] but also live chat on e-commerce Web sites, live chat on messaging applications such as WeChat, Facebook Messenger and WhatsApp [33] and chatbots on messaging applications or Web sites.

Customer support

A virtual assistant can work with customer support team of a business to provide 24x7 support to customers. It provides quick responses, which enhances a customer's experience.

Third-party services

Amazon enables Alexa "Skills" and Google "Actions", essentially applications that run on the assistant platforms.

Privacy

Virtual assistants have a variety of privacy concerns associated with them. Features such as activation by voice pose a threat, as such features requires the device to always be listening. [34] Modes of privacy such as the virtual security button have been proposed to create a multilayer authentication for virtual assistants. [35]

Google Assistant

The privacy policy of Google Assistant states that it does not store the audio data without the user's permission, but may store the conversation transcripts to personalise its experience. Personalisation can be turned off in settings. If a user wants Google Assistant to store audio data, they can go to Voice & Audio Activity (VAA) and turn on this feature. Audio files are sent to the cloud and used by Google to improve the performance of Google Assistant, but only if the VAA feature is turned on. [36]

Amazon Alexa

The privacy policy of Amazon's virtual assistant, Alexa, states that it only listens to conversations when its wake word (like Alexa, Amazon, Echo) is used. It starts recording the conversation after the call of a wake word, and stops recording after 8 seconds of silence. It sends the recorded conversation to the cloud. It is possible to delete the recording from the cloud by visiting 'Alexa Privacy' in 'Alexa'. [37]

Apple's Siri

Apple states that it does not record audio to improve Siri. Instead, it claims to use transcripts. Transcript data is only sent if it is deemed important for analysis. Users can opt out anytime if they don't want Siri to send the transcripts in the cloud. [38]

Consumer interest

Presumed added value as allowing a new way of interactions

Added value of the virtual assistants can come among others from the following:

  1. It is convenient: there are some sectors where voice is the only way of possible communication, and more generally, it allows to free-up both hands and vision potentially for doing another activity in parallel, or helps also disabled people.
  2. It is faster: Voice is more efficient than writing on a keyboard: we can speak up to 200 words per minute opposed to 60 in case of writing on a keyboard. It is also more natural thus requiring less effort (reading a text however can reach 700 words per minute). [39]

Perceived interest

Graphical sum up of the study capturing reasons of interest of virtual assistants for consumers Study results.jpg
Graphical sum up of the study capturing reasons of interest of virtual assistants for consumers
  1. Provided content quality has a very strong influence on perceived usefulness and a strong influence on perceived enjoyment.
  2. Visual attractiveness has a very strong influence on perceived enjoyment.
  3. Automation has a strong influence on perceived usefulness. [41]

Controversies

Artificial intelligence controversies

  1. Not intelligent because all they do is being the assistant of the human, and only by doing tasks that a human could do easily, and in a very limited specter of actions: find, class, and present information, offers or documents. Also, virtual assistants are neither able to make decisions on their own nor to anticipate things.
  2. And not artificial because they would be impossible without human labelization through micro working. [43]

Ethical implications

In 2019 Antonio A. Casilli, a French sociologist, criticized artificial intelligence and virtual assistants in particular in the following way:

At a first level the fact that the consumer provides free data for the training and improvement of the virtual assistant, often without knowing it, is ethically disturbing.

But at a second level, it might be even more ethically disturbing to know how these AIs are trained with this data.

This artificial intelligence is trained via neural networks, which require a huge amount of labelled data. However, this data needs to be labelled through a human process, which explains the rise of microwork in the last decade. That is, remotely using some people worldwide doing some repetitive and very simple tasks for a few cents, such as listening to virtual assistant speech data, and writing down what was said. Microwork has been criticized for the job insecurity it causes, and for the total lack of regulation: The average salary was 1,38 dollar/hour in 2010, [44] and it provides neither healthcare nor retirement benefits, sick pay, minimum wage. Hence, virtual assistants and their designers are controversial for spurring job insecurity, and the AIs they propose are still human in the way that they would be impossible without the microwork of millions of human workers. [43]

Privacy concerns are raised by the fact that voice commands are available to the providers of virtual assistants in unencrypted form, and can thus be shared with third parties and be processed in an unauthorized or unexpected manner. [45] Additionally to the linguistic content of recorded speech, a user's manner of expression and voice characteristics can implicitly contain information about his or her biometric identity, personality traits, body shape, physical and mental health condition, sex, gender, moods and emotions, socioeconomic status and geographical origin. [46]

Developer platforms

Notable developer platforms for virtual assistants include:

Previous generations

In previous generations of text chat-based virtual assistants, the assistant was often represented by an avatar (a.k.a. interactive online character or automated character) — this was known as an embodied agent.

Economic relevance

For individuals

Digital experiences enabled by virtual assistants are considered to be among the major recent technological advances and most promising consumer trends. Experts claim that digital experiences will achieve a status-weight comparable to 'real' experiences, if not become more sought-after and prized. [51] The trend is verified by a high number of frequent users and the substantial growth of worldwide user numbers of virtual digital assistants. In mid-2017, the number of frequent users of digital virtual assistants is estimated to be around 1 bn worldwide. [52] In addition, it can be observed that virtual digital assistant technology is no longer restricted to smartphone applications, but present across many industry sectors (incl. automotive, telecommunications, retail, healthcare and education). [53] In response to the significant R&D expenses of firms across all sectors and an increasing implementation of mobile devices, the market for speech recognition technology is predicted to grow at a CAGR of 34.9% globally over the period of 2016 to 2024 and thereby surpass a global market size of US$7.5 billion by 2024. [53] According to an Ovum study, the "native digital assistant installed base" is projected to exceed the world's population by 2021, with 7.5 billion active voice AI–capable devices. [54] According to Ovum, by that time "Google Assistant will dominate the voice AI–capable device market with 23.3% market share, followed by Samsung's Bixby (14.5%), Apple's Siri (13.1%), Amazon's Alexa (3.9%), and Microsoft's Cortana (2.3%)." [54]

Taking into consideration the regional distribution of market leaders, North American companies (e.g. Nuance Communications, IBM, eGain) are expected to dominate the industry over the next years, due to the significant impact of BYOD (Bring Your Own Device) and enterprise mobility business models. Furthermore, the increasing demand for smartphone-assisted platforms are expected to further boost the North American intelligent virtual assistant (IVA) industry growth. Despite its smaller size in comparison to the North American market, the intelligent virtual assistant industry from the Asia-Pacific region, with its main players located in India and China is predicted to grow at an annual growth rate of 40% (above global average) over the 2016–2024 period. [53]

Economic opportunity for enterprises

Virtual assistants should not be only seen as a gadget for individuals, as they could have a real economic utility for enterprises. As an example, a virtual assistant can take the role of an always available assistant with an encyclopedic knowledge. And which can organize meetings, check inventories, verify informations. Virtual assistants are all the more important that their integration in small and middle-sized enterprises often consists in an easy first step through the more global adaptation and use of Internet of Things (IoT). Indeed, IoT technologies are first perceived by small and medium-sized enterprises as technologies of critical importance, but too complicated, risky or costly to be used. [55]

Security

In May 2018, researchers from the University of California, Berkeley, published a paper that showed audio commands undetectable for the human ear could be directly embedded into music or spoken text, thereby manipulating virtual assistants into performing certain actions without the user taking note of it. [56] The researchers made small changes to audio files, which cancelled out the sound patterns that speech recognition systems are meant to detect. These were replaced with sounds that would be interpreted differently by the system and command it to dial phone numbers, open websites or even transfer money. [56] The possibility of this has been known since 2016, [56] and affects devices from Apple, Amazon and Google. [57]

In addition to unintentional actions and voice recording, another security and privacy risk associated with intelligent virtual assistants is malicious voice commands: An attacker who impersonates a user and issues malicious voice commands to, for example, unlock a smart door to gain unauthorized entry to a home or garage or order items online without the user's knowledge. Although some IVAs provide a voice-training feature to prevent such impersonation, it can be difficult for the system to distinguish between similar voices. Thus, a malicious person who is able to access an IVA-enabled device might be able to fool the system into thinking that they are the real owner and carry out criminal or mischievous acts. [58]

Comparison of notable assistants

Intelligent personal assistantDeveloper Free software Free and open-source hardware HDMI outExternal I/O IOT Chromecast integrationSmart phone appAlways onUnit to unit voice channelSkill language
Alexa (a.k.a. Echo) Amazon.com NoNoNoNoYesNoYesYes ?JavaScript
Alice Yandex NoYesNoYesYes ?
AliGenie Alibaba Group NoNoYesNoYesYes ?
Assistant Speaktoit NoNoNoYesNo ?
Bixby Samsung Electronics NoNoNoYesJavaScript
BlackBerry Assistant BlackBerry Limited NoNoNoYesNo ?
Braina BrainasoftNoNoNoYesNo ?
Clova Naver Corporation NoYesNoYesYes ?
Cortana Microsoft NoYesNoYesYes ?
Duer Baidu [59]
Evi Amazon.com and True Knowledge NoNoNoYesNo ?
Google Assistant Google NoYesYesYesYesC++
Google Now Google NoYesYesYesYes ?
Mycroft [60] Mycroft AIYesYesYesYesYesYesYesYesYesPython
SILVIA Cognitive CodeNoNoNoYesNo ?
Siri Apple Inc. NoNoYesNoYesYes ?
Viv Samsung Electronics NoYesNoYesNo ?
Xiaowei Tencent  ?
Celia Huawei NoNoYesNoYesYes?

See also

Related Research Articles

<span class="mw-page-title-main">Chatbot</span> Program that simulates conversation

A chatbot is a software application or web interface that is designed to mimic human conversation through text or voice interactions. Modern chatbots are typically online and use generative artificial intelligence systems that are capable of maintaining a conversation with a user in natural language and simulating the way a human would behave as a conversational partner. Such chatbots often use deep learning and natural language processing, but simpler chatbots have existed for decades.

A voice-user interface (VUI) enables spoken human interaction with computers, using speech recognition to understand spoken commands and answer questions, and typically text to speech to play a reply. A voice command device is a device controlled with a voice user interface.

In artificial intelligence, an embodied agent, also sometimes referred to as an interface agent, is an intelligent agent that interacts with the environment through a physical body within that environment. Agents that are represented graphically with a body, for example a human or a cartoon animal, are also called embodied agents, although they have only virtual, not physical, embodiment. A branch of artificial intelligence focuses on empowering such agents to interact autonomously with human beings and the environment. Mobile robots are one example of physically embodied agents; Ananova and Microsoft Agent are examples of graphically embodied agents. Embodied conversational agents are embodied agents that are capable of engaging in conversation with one another and with humans employing the same verbal and nonverbal means that humans do.

<span class="mw-page-title-main">Siri</span> Software-based personal assistant from Apple Inc.

Siri is the digital assistant that is part of Apple Inc.'s iOS, iPadOS, watchOS, macOS, tvOS, audioOS, and visionOS operating systems. It uses voice queries, gesture based control, focus-tracking and a natural-language user interface to answer questions, make recommendations, and perform actions by delegating requests to a set of Internet services. With continued use, it adapts to users' individual language usages, searches, and preferences, returning individualized results.

<span class="mw-page-title-main">Braina</span> Intelligent personal assistant & dictation software

Braina is a virtual assistant and speech-to-text dictation application for Microsoft Windows developed by Brainasoft. Braina uses natural language interface, speech synthesis, and speech recognition technology to interact with its users and allows them to use natural language sentences to perform various tasks on a computer in most languages of the world. The name Braina is a short form of “Brain Artificial”.

<span class="mw-page-title-main">Google Assistant</span> AI-powered digital assistant from Google

The Google Assistant is a virtual assistant software application developed by Google that is primarily available on mobile and home automation devices. Based on artificial intelligence, The Google Assistant can engage in two-way conversations, unlike the company's previous virtual assistant, Google Now.

Amazon Alexa or Alexa is a virtual assistant technology largely based on a Polish speech synthesizer named Ivona, bought by Amazon in 2013. It was first used in the Amazon Echo smart speaker and the Echo Dot, Echo Studio and Amazon Tap speakers developed by Amazon Lab126. It is capable of natural language processing (NLP) for tasks such as voice interaction, music playback, creating to-do lists, setting alarms, streaming podcasts, playing audiobooks, providing weather, traffic, sports, other real-time information and news. Alexa can also control several smart devices as a home automation system. Alexa capabilities may be extended by installing "skills" such as weather programs and audio features. It performs these tasks using automatic speech recognition, NLP, and other forms of weak AI.

A smart speaker is a type of loudspeaker and voice command device with an integrated virtual assistant that offers interactive actions and hands-free activation with the help of one "hot word". Some smart speakers can also act as a smart device that utilizes Wi-Fi and other protocol standards to extend usage beyond audio playback, such as to control home automation devices. This can include, but is not limited to, features such as compatibility across a number of services and platforms, peer-to-peer connection through mesh networking, virtual assistants, and others. Each can have its own designated interface and features in-house, usually launched or controlled via application or home automation software. Some smart speakers also include a screen to show the user a visual response.

Amazon Lex is a service for building conversational interfaces into any application using voice and text. It powers the Amazon Alexa virtual assistant. In April 2017, the platform was released to the developer community, and suggested that it could be used for conversational interfaces including Web, mobile apps, robots, toys, drones, and more. Amazon already had launched Alexa Voice Services, which developers can use to integrate Alexa into their own devices, like smart speakers, alarm clocks, etc.; however, Lex will not require that end users interact with the Alexa assistant per se, but rather any type of assistant or interface. As of February 2018, users can now define a response for Amazon Lex chatbots directly from the AWS management console.

<span class="mw-page-title-main">Witlingo</span> Software as a service company

Witlingo is a B2B Software as a Service (SaaS) company that enables businesses and organization of all sizes to use the latest innovations in Human Language Technology and Conversational AI, such Speech recognition, Natural Language Processing, IVR, Virtual Assistant apps on Smartphone platforms(iOS and Android), Chatbots, and Digital audio, to deeply engage with their communities.

Alice is a Russian intelligent personal assistant for Android, iOS and Windows operating systems and Yandex's own devices developed by Yandex. Alice was officially introduced on 10 October 2017. Aside from common tasks, such as internet search or weather forecasts, it can also run applications and chit-chat. Alice is also the virtual assistant used for the Yandex Station smart speaker.

<span class="mw-page-title-main">Haptik</span> Indian enterprise conversational AI platform

Haptik is an Indian enterprise conversational AI platform founded in August 2013, and acquired by Reliance Industries Limited in 2019. The company develops technology to enable enterprises to build conversational AI systems that allow users to converse with applications and electronic devices in free-format, natural language, using speech or text. The company has been accorded numerous accolades including the Frost & Sullivan Award, NASSCOM's Al Game Changer Award, and serves Fortune 500 brands globally in industries such as financial, insurance, healthcare, technology and communications.

A conversational user interface (CUI) is a user interface for computers that emulates a conversation with a real human. Historically, computers have relied on text-based user interfaces and graphical user interfaces (GUIs) to translate the user's desired action into commands the computer understands. While an effective mechanism of completing computing actions, there is a learning curve for the user associated with GUI. Instead, CUIs provide opportunity for the user to communicate with the computer in their natural language rather than in a syntax specific commands.

Conversational commerce is e-commerce done via various means of conversation and using technology such as: speech recognition, speaker recognition, natural language processing and artificial intelligence.

Virtual assistants are software technology that assist users complete various tasks. Well known virtual assistants include Amazon Alexa, and Siri, produced by Apple. Other companies, such as Google and Microsoft, also have virtual assistants. There are privacy issues concerning what information can go to the third party corporations that operate virtual assistants and how this data can potentially be used.

<span class="mw-page-title-main">Voice computing</span> Discipline in computing

Voice computing is the discipline that develops hardware or software to process voice inputs.

LaMDA is a family of conversational large language models developed by Google. Originally developed and introduced as Meena in 2020, the first-generation LaMDA was announced during the 2021 Google I/O keynote, while the second generation was announced the following year. In June 2022, LaMDA gained widespread attention when Google engineer Blake Lemoine made claims that the chatbot had become sentient. The scientific community has largely rejected Lemoine's claims, though it has led to conversations about the efficacy of the Turing test, which measures whether a computer can pass for a human. In February 2023, Google announced Bard, a conversational artificial intelligence chatbot powered by LaMDA, to counter the rise of OpenAI's ChatGPT.

<span class="mw-page-title-main">ChatGPT</span> Chatbot developed by OpenAI

ChatGPT is a chatbot developed by OpenAI and launched on November 30, 2022. Based on a large language model, it enables users to refine and steer a conversation towards a desired length, format, style, level of detail, and language. Successive prompts and replies, known as prompt engineering, are considered at each conversation stage as a context.

References

  1. Hoy, Matthew B. (2018). "Alexa, Siri, Cortana, and More: An Introduction to Voice Assistants". Medical Reference Services Quarterly. 37 (1): 81–88. doi:10.1080/02763869.2018.1404391. PMID   29327988. S2CID   30809087.
  2. "Siri vs Alexa vs Google Assistant vs Bixby: Which one reigns supreme?". Android Authority.
  3. "The Magic of Virtual Assistants and Their Impact on Customer Service".
  4. 1 2 "The One Thing You Should Definitely be Using AI Chatbot for". 7 April 2023.
  5. "A.I. Means everyone gets a 'white-collar' personal assistant, Bill Gates says".
  6. "Chat GPT: What is it?". uca.edu. Retrieved 8 February 2024.
  7. US 1209636,Christian Berger,"Sound-Operated Circuit Controller",issued 1916-12-19, assigned to Submarine Wireless Company
  8. Markowitz, Judith. "Toys That Have a Voice". SpeechTechMag.
  9. Moskvitch, Katia. "The machines that learned to listen". BBC. Retrieved 5 May 2020.
  10. Epstein, J; Klinkenberg, W. D (1 May 2001). "From Eliza to Internet: a brief history of computerized assessment". Computers in Human Behavior. 17 (3): 295–314. doi:10.1016/S0747-5632(01)00004-8. ISSN   0747-5632.
  11. Weizenbaum, Joseph (1976). Computer power and human reason : from judgment to calculation. Oliver Wendell Holmes Library Phillips Academy. San Francisco : W. H. Freeman.
  12. "Smartphone: your new personal assistant – Orange Pop". 10 July 2017. Archived from the original on 10 July 2017. Retrieved 5 May 2020.
  13. Murph, Darren (4 October 2011). "iPhone 4S hands-on!". Engadget.com. Retrieved 10 December 2017.
  14. "Feature: Von IBM Shoebox bis Siri: 50 Jahre Spracherkennung – WELT" [From IBM Shoebox to Siri: 50 years of speech recognition]. Die Welt (in German). Welt.de. 20 April 2012. Retrieved 10 December 2017.
  15. Kundu, Kishalaya (2023). "Amazon expands Echo lineup with new smart speaker, earbuds, and more". XDA. Retrieved 26 May 2023.
  16. Sterling, Bruce (13 February 2020). "Web Semantics: Microsoft Project Turing introduces Turing Natural Language Generation (T-NLG)". Wired. ISSN   1059-1028 . Retrieved 31 July 2020.
  17. "GPT-4 takes the world by storm - List of companies that integrated the chatbot". 21 March 2023.
  18. "Conversica Raises $31 Million in Series C Funding to Fuel Expansion of Conversational AI for Business". Bloomberg.com. 30 October 2018. Retrieved 23 October 2020.
  19. Herrera, Sebastian (26 September 2019). "Amazon Extends Alexa's Reach Into Wearables". The Wall Street Journal . Retrieved 26 September 2019.
  20. "S7617 – Developing Your Own Wake Word Engine Just Like 'Alexa' and 'OK Google'". GPU Technology Conference. Retrieved 17 July 2017.
  21. Van Loo, Rory (1 March 2019). "Digital Market Perfection". Michigan Law Review. 117 (5): 815. doi: 10.36644/mlr.117.5.digital . S2CID   86402702.
  22. La, Lynn (27 February 2017). "Everything Google Assistant can do on the Pixel". CNET. Retrieved 10 December 2017.
  23. Morrison, Maureen (5 October 2014). "Domino's Pitches Voice-Ordering App in Fast-Food First | CMO Strategy". AdAge. Retrieved 10 December 2017.
  24. O'Shea, Dan (4 January 2017). "LG introduces smart refrigerator with Amazon Alexa-enabled grocery ordering". Retail Dive. Retrieved 10 December 2017.
  25. Gibbs, Samuel (7 February 2017). "Amazon's Alexa escapes the Echo and gets into cars | Technology". The Guardian. Retrieved 10 December 2017.
  26. "What is Google Assistant, how does it work, and which devices offer it?". Pocket-lint. 6 October 2017. Retrieved 10 December 2017.
  27. "'Ask Jenn', Alaska Airlines website". Alaska Airlines. 2 January 2017. Retrieved 10 December 2017.
  28. AT&T Tech Channel (26 June 2013). "American Airlines (US Airways) – First US Airline to Deploy Natural Language Speech" (video), Nuance Enterprise on YouTube. Archived from the original on 21 December 2021. Retrieved 10 December 2017 via YouTube. YouTube title: Airline Information System, 1989 – AT&T Archives – speech recognition
  29. Martin, Taylor; Priest, David (10 September 2017). "The complete list of Alexa commands so far". CNET. Retrieved 10 December 2017.
  30. Kongthon, Alisa; Sangkeettrakarn, Chatchawal; Kongyoung, Sarawoot; Haruechaiyasak, Choochart (1 January 2009). "Implementing an online help desk system based on conversational agent". Proceedings of the International Conference on Management of Emergent Digital EcoSystems. MEDES '09. New York, NY, USA: ACM. pp. 69:450–69:451. doi:10.1145/1643823.1643908. ISBN   9781605588292. S2CID   1046438.
  31. O'Donnell, Anthony (3 June 2010). "Aetna's new "virtual online assistant"". Insurance & Technology. Archived from the original on 7 June 2010.
  32. "How to prepare your products and brand for conversational commerce". 6 March 2018.
  33. Taylor, Glenn (5 March 2018). "Retail's Big Opportunity: 87% Of U.S. Consumers Grasp The Power Of Conversational Commerce – Retail TouchPoints".
  34. Zhang, Guoming; Yan, Chen; Ji, Xiaoyu; Zhang, Tianchen; Zhang, Taimin; Xu, Wenyuan (2017). "DolphinAttack". Proceedings of the 2017 ACM SIGSAC Conference on Computer and Communications Security - CCS '17. pp. 103–117. arXiv: 1708.09537 . doi:10.1145/3133956.3134052. ISBN   9781450349468. S2CID   2419970.
  35. Lei, Xinyu; Tu, Guan-Hua; Liu, Alex X.; Li, Chi-Yu; Xie, Tian (2017). "The Insecurity of Home Digital Voice Assistants – Amazon Alexa as a Case Study". arXiv: 1712.03327 [cs.CR].
  36. "Doing more to protect your privacy with the Assistant". Google. 23 September 2019. Retrieved 27 February 2020.
  37. "Alexa, Echo Devices, and Your Privacy". Amazon.com. Retrieved 27 February 2020.
  38. "Improving Siri's privacy protections". Apple Newsroom. Retrieved 27 February 2020.
  39. Minker, W.; Néel, F. (2002). "Développement des technologies vocales". Le Travail Humain. 65 (3): 261. doi: 10.3917/th.653.0261 . ISSN   0041-1868.
  40. Wajcman, Judy (2019). "The Digital Architecture of time Management" (PDF). Science, Technology, & Human Values. 44 (2): 315–337. doi:10.1177/0162243918795041. S2CID   149648777.
  41. Yang, Heetae; Lee, Hwansoo (26 June 2018). "Understanding user behavior of virtual personal assistant devices". Information Systems and E-Business Management. 17 (1): 65–87. doi:10.1007/s10257-018-0375-1. ISSN   1617-9846. S2CID   56838915.
  42. Tisseron, Serge (2019). "La famille sous écoute". L'École des Parents. n° 632 (3): 16. doi:10.3917/epar.632.0016. ISSN   0424-2238. S2CID   199344092.{{cite journal}}: |volume= has extra text (help)
  43. 1 2 Casilli, Antonio A. (2019). En attendant les robots. Enquête sur le travail du clic. Editions Seuil. ISBN   978-2-02-140188-2. OCLC   1083583353.
  44. Horton, John Joseph; Chilton, Lydia B. (2010). "The labor economics of paid crowdsourcing". Proceedings of the 11th ACM conference on Electronic commerce. EC '10. New York, New York, USA: ACM Press. pp. 209–218. arXiv: 1001.0627 . doi:10.1145/1807342.1807376. ISBN   978-1-60558-822-3. S2CID   18237602.
  45. "Apple, Google, and Amazon May Have Violated Your Privacy by Reviewing Digital Assistant Commands". Fortune. 5 August 2019. Retrieved 13 May 2020.
  46. Kröger, Jacob Leon; Lutz, Otto Hans-Martin; Raschke, Philip (2020). "Privacy Implications of Voice and Speech Analysis – Information Disclosure by Inference". Privacy and Identity Management. Data for Better Living: AI and Privacy. IFIP Advances in Information and Communication Technology. Vol. 576. pp. 242–258. doi: 10.1007/978-3-030-42504-3_16 . ISBN   978-3-030-42503-6. ISSN   1868-4238.
  47. "Amazon Lex, the technology behind Alexa, opens up to developers". TechCrunch. 20 April 2017. Retrieved 10 December 2017.
  48. "Actions on Google | Google Developers" . Retrieved 10 December 2017.
  49. "Watson – Stories of how AI and Watson are transforming business and our world". Ibm.com. Retrieved 10 December 2017.
  50. Memeti, Suejb; Pllana, Sabri (January 2018). "PAPA: A parallel programming assistant powered by IBM Watson cognitive computing technology". Journal of Computational Science. 26: 275–284. doi:10.1016/j.jocs.2018.01.001.
  51. "5 Consumer Trends for 2017". TrendWatching. 31 October 2016. Retrieved 10 December 2017.
  52. Richter, Felix (26 August 2016). "Chart: Digital Assistants – Always at Your Service". Statista. Retrieved 10 December 2017.
  53. 1 2 3 "Virtual Assistant Industry Statistics". Global Market Insights. 30 January 2017. Retrieved 10 December 2017.
  54. 1 2 "Virtual digital assistants to overtake world population by 2021". ovum.informa.com. Retrieved 11 May 2018.
  55. Jones, Nory B.; Graham, C. Matt (February 2018). "Can the IoT Help Small Businesses?". Bulletin of Science, Technology & Society. 38 (1–2): 3–12. doi:10.1177/0270467620902365. ISSN   0270-4676. S2CID   214031256.
  56. 1 2 3 "Alexa and Siri Can Hear This Hidden Command. You Can't". The New York Times. 10 May 2018. ISSN   0362-4331 . Retrieved 11 May 2018.
  57. "As voice assistants go mainstream, researchers warn of vulnerabilities". CNET. 10 May 2018. Retrieved 11 May 2018.
  58. Chung, H.; Iorga, M.; Voas, J.; Lee, S. (2017). "Alexa, Can I Trust You?". Computer. 50 (9): 100–104. doi:10.1109/MC.2017.3571053. ISSN   0018-9162. PMC   5714311 . PMID   29213147.
  59. "Baidu unveils 3 smart speakers with its Duer digital assistant". 8 January 2018.
  60. MSV, Janakiram (20 August 2015). "Meet Mycroft, The Open Source Alternative To Amazon Echo". Forbes. Retrieved 27 October 2016.