The San Juan Daily Star
How Siri, Alexa and Google Assistant lost the AI race
By Brian X. Chen, Nico Grant and Karen Weise
On a rainy Tuesday in San Francisco, Apple executives took the stage in a crowded auditorium to unveil the fifth-generation iPhone. The phone, which looked identical to the previous version, had a new feature that the audience was soon buzzing about: Siri, a virtual assistant.
Scott Forstall, then Apple’s head of software, pushed an iPhone button to summon Siri and prodded it with questions. At his request, Siri checked the time in Paris (“8:16 p.m.,” Siri replied), defined the word “mitosis” (“Cell division in which the nucleus divides into nuclei containing the same number of chromosomes,” it said) and pulled up a list of 14 highly rated Greek restaurants, five of them in Palo Alto, California.
“I’ve been in the AI field for a long time, and this still blows me away,” Forstall said.
That was 12 years ago. Since then, people have been far from blown away by Siri and competing assistants that are powered by artificial intelligence, like Amazon’s Alexa and Google Assistant. The technology has largely remained stagnant, and the talking assistants have become the butt of jokes, including in a 2018 “Saturday Night Live” sketch featuring a smart speaker for seniors.
The tech world is now gushing over a different kind of virtual assistant: chatbots. These AI-powered bots, such as ChatGPT and the new ChatGPT Plus from the San Francisco company OpenAI, can improvise answers to questions typed into a chat box with alacrity. People have used ChatGPT to handle complex tasks like coding software, drafting business proposals and writing fiction.
And ChatGPT, which uses AI to guess what word comes next, is rapidly improving. A few months ago, it couldn’t write a proper haiku; now it can do so with gusto. On Tuesday, OpenAI unveiled its next-generation AI engine, GPT-4, which powers ChatGPT.
The excitement around chatbots illustrates how Siri, Alexa and other voice assistants — which once elicited similar enthusiasm — have squandered their lead in the AI race.
Over the past decade, the products hit roadblocks. Siri ran into technological hurdles, including clunky code that took weeks to update with basic features, said John Burkey, a former Apple engineer who worked on the assistant. Amazon and Google miscalculated how the voice assistants would be used, leading them to invest in areas with the technology that rarely paid off, former employees said. When those experiments failed, enthusiasm for the technology waned at the companies, they said.
Voice assistants are “dumb as a rock,” Satya Nadella, Microsoft’s chief executive, said in an interview this month with The Financial Times, declaring that newer AI would lead the way. Microsoft has worked closely with OpenAI, investing $13 billion in the startup and incorporating its technology into the Bing search engine, as well as other products.
Apple declined to comment on Siri. Google said it was committed to providing a great virtual assistant to help people on their phones and inside their homes and cars; the company is separately testing a chatbot called Bard. Amazon said that it saw a 30% increase in customer engagement globally with Alexa in the last year and that it was optimistic about its mission to build world-class AI.
The assistants and the chatbots are based on different flavors of AI. Chatbots are powered by what are known as large language models, which are systems trained to recognize and generate text based on enormous data sets scraped off the web. They can then suggest words to complete a sentence.
In contrast, Siri, Alexa and Google Assistant are essentially what are known as command-and-control systems. These can understand a finite list of questions and requests like “What’s the weather in New York City?” or “Turn on the bedroom lights.” If a user asks the virtual assistant to do something that is not in its code, the bot simply says it can’t help.
Siri also had a cumbersome design that made it time-consuming to add new features, said Burkey, who was given the job of improving Siri in 2014. Siri’s database contains a gigantic list of words, including the names of musical artists and locations like restaurants, in nearly two dozen languages.
That made it “one big snowball,” he said. If someone wanted to add a word to Siri’s database, he added, “it goes in one big pile.”
So seemingly simple updates, like adding some new phrases to the data set, would require rebuilding the entire database, which could take up to six weeks, Burkey said. Adding more complex features like new search tools could take nearly a year. That meant there was no path for Siri to become a creative assistant like ChatGPT, he said.
Alexa and Google Assistant relied on technology similar to Siri’s, but the companies struggled to generate meaningful revenue with the assistants, former managers at Amazon and Google said. (In contrast, Apple successfully used Siri to entice buyers to its iPhones.)
After Amazon released the Echo, a smart speaker powered by Alexa, in 2014, the company hoped the product would help it increase sales in its online store by enabling consumers to talk to Alexa to place orders, said a former Amazon leader involved with Alexa. But while people had fun playing with Alexa’s ability to answer weather prompts and set alarms, few asked Alexa to order items, he added.
Amazon may have overinvested in making new kinds of hardware, like now-discontinued alarm clocks and microwaves that worked with Alexa, which sold at or below cost, the former executive said.
The company also underinvested in creating an ecosystem for people to easily expand Alexa’s abilities, in the way that Apple had done with its App Store, which helped stoke interest in the iPhone, the person said. While Amazon offered a “skills” store to make Alexa control third-party accessories like light switches, it was difficult for people to find and set up skills for the speakers — unlike the friction-free experience of downloading mobile apps from app stores.
Amazon’s misfires with Alexa may have led Google astray, said a former manager who worked on Google Assistant. Google engineers spent years experimenting with its assistant to mimic what Alexa could do, including designing smart speakers and voice-controlled tablet screens to control home accessories like thermostats and light switches. The company later integrated ads into those home products, which did not become a major source of revenue.
Many of the big tech companies are now racing to come up with responses to ChatGPT. At Apple’s headquarters last month, the company held its annual AI summit, an internal event for employees to learn about its large language model and other AI tools, two people who were briefed on the program said. Many engineers, including members of the Siri team, have been testing language-generating concepts every week, the people said.
On Tuesday, Google also said it would soon release generative AI tools to help businesses, governments and software developers build applications with embedded chatbots, and incorporate the underlying technology into their systems.
In the future, the technologies of chatbots and voice assistants will converge, AI experts said. That means people will be able to control chatbots with speech, and those who use Apple, Amazon and Google products will be able to ask the virtual assistants to help them with their jobs, not just tasks like checking the weather.
“These products never worked in the past because we never had human-level dialogue capabilities,” said Aravind Srinivas, a founder of Perplexity, an AI startup that offers a chatbot-powered search engine. “Now we do.”