“Siri, Read My Mind”: A New Device Lets Users Think Commands

Arnav Kapur, a researcher in the Fluid Interfaces Group at MIT, wears an earlier prototype of the AlterEgo. A newer version is more discreet.

(Photo credit: Lorrie LeJeune/4.0 International)

Sometime in the near future, we won't need to type on a smartphone or computer to silently communicate our thoughts to others.

"We're moving as fast as possible to get the technology right, to get the ethics right, to get everything right."

In fact, the devices themselves will quietly understand our intentions and express them to other people. We won't even need to move our mouths.

That "sometime in the near future" is now.

At the recent TED Conference, MIT student and TED Fellow Arnav Kapur was onstage with a colleague doing the first live public demo of his new technology. He was showing how you can communicate with a computer using signals from your brain. The usually cool, erudite audience seemed a little uncomfortable.

"If you look at the history of computing, we've always treated computers as external devices that compute and act on our behalf," Kapur said. "What I want to do is I want to weave computing, AI and Internet as part of us."

His colleague started up a device called AlterEgo. Thin like a sticker, AlterEgo picks up signals in the mouth cavity. It recognizes the intended speech and processes it through the built-in AI. The device then gives feedback to the user directly through bone conduction: It vibrates your inner ear drum and gives you a response meshing with your normal hearing.

Onstage, the assistant quietly thought of a question: "What is the weather in Vancouver?" Seconds later, AlterEgo told him in his ear. "It's 50 degrees and rainy here in Vancouver," the assistant announced.

AlterEgo essentially gives you a built-in Siri.

"We don't have a deadline [to go to market], but we're moving as fast as possible to get the technology right, to get the ethics right, to get everything right," Kapur told me after the talk. "We're developing it both as a general purpose computer interface and [in specific instances] like on the clinical side or even in people's homes."

Nearly-telepathic communication actually makes sense now. About ten years ago, the Apple iPhone replaced the ubiquitous cell phone keyboard with a blank touchscreen. A few years later, Google Glass put computer screens into a simple lens. More recently, Amazon Alexa and Microsoft Cortana have dropped the screen and gone straight for voice control. Now those voices are getting closer to our minds and may even become indistinguishable in the future.

"We knew the voice market was growing, like with getting map locations, and audio is the next frontier of user interfaces," says Dr. Rupal Patel, Founder and CEO of VocalID. The startup literally gives voices to the voiceless, particularly people unable to speak because of illness or other circumstances.

"We start with [our database of] human voices, then train our deep learning technology to learn the pattern of speech… We mix voices together from our voice bank, so it's not just Damon's voice, but three or five voices. They are different enough to blend it into a voice that does not exist today – kind of like a face morph."

The VocalID customer then has a voice as unique as he or she is, mixed together like a Sauvignon blend. It is a surrogate voice for those of us who cannot speak, just as much as AlterEgo is a surrogate companion for our brains.

"I'm very skeptical keyboards or voice-based communication will be replaced any time soon."

Voice equality will become increasingly important as Siri, Alexa and voice-based interfaces become the dominant communication method.

It may feel odd to view your voice as a privilege, but as the world becomes more voice-activated, there will be a wider gap between the speakers and the voiceless. Picture going shopping without access to the Internet or trying to eat healthily when your neighborhood is a food desert. And suffering from vocal difficulties is more common than you might think. In fact, according to government statistics, around 7.5 million people in the U.S. have trouble using their voices.

While voice communication appears to be here to stay, at least for now, a more radical shift to mind-controlled communication is not necessarily inevitable. Tech futurist Wagner James Au, for one, is dubious.

"I'm very skeptical keyboards or voice-based communication will be replaced any time soon. Generation Z has grown up with smartphones and games like Fortnite, so I don't see them quickly switching to a new form factor. It's still unclear if even head-mounted AR/VR displays will see mass adoption, and mind-reading devices are a far greater physical imposition on the user."

How adopters use the newest brain impulse-reading, voice-altering technology is a much more complicated discussion. This spring, a video showed U.S. House Speaker Nancy Pelosi stammering and slurring her words at a press conference. The problem is that it didn't really happen: the video was manufactured and heavily altered from the original source material.

So-called deepfake videos use computer algorithms to capture the visual and vocal cues of an individual, and then the creator can manipulate it to say whatever it wants. Deepfakes have already created false narratives in the political and media systems – and these are only videos. Newer tech is making the barrier between tech and our brains, if not our entire identity, even thinner.

"Last year," says Patel of VocalID, "we did penetration testing with our voices on banks that use voice control – and our generation 4 system is even tricky for you and me to identify the difference (between real and fake). As a forward-thinking company, we want to prevent risk early on by watermarking voices, creating a detector of false voices, and so on." She adds, "The line will become more blurred over time."

Onstage at TED, Kapur reassured the audience about who would be in the driver's seat. "This is why we designed the system to deliberately record from the peripheral nervous system, which is why the control in all situations resides with the user."

And, like many creators, he quickly shifted back to the possibilities. "What could the implications of something like this be? Imagine perfectly memorizing things, where you perfectly record information that you silently speak, and then hear them later when you want to, internally searching for information, crunching numbers at speeds computers do, silently texting other people."

"The potential," he concluded, "could be far-reaching."

Damon Brown
Damon Brown co-founded the popular platonic connection app Cuddlr. Now he helps side hustlers, solopreneurs, and other non-traditional entrepreneurs bloom. He is author of the TED book "Our Virtual Shadow" and, most recently, the best-selling "The Bite-Sized Entrepreneur" series. Join his creative community at www.JoinDamon.me.
Get our top stories twice a month
Follow us on

Dr. David Fajgenbaum looking through a microscope at his lab.

Courtesy of Fajgenbaum

In late March, just as the COVID-19 pandemic was ramping up in the United States, David Fajgenbaum, a physician-scientist at the University of Pennsylvania, devised a 10-day challenge for his lab: they would sift through 1,000 recently published scientific papers documenting cases of the deadly virus from around the world, pluck out the names of any drugs used in an attempt to cure patients, and track the treatments and their outcomes in a database.

Before late 2019, no one had ever had to treat this exact disease before, which meant all treatments would be trial and error. Fajgenbaum, a pioneering researcher in the field of drug repurposing—which prioritizes finding novel uses for existing drugs, rather than arduously and expensively developing new ones for each new disease—knew that physicians around the world would be embarking on an experimental journey, the scale of which would be unprecedented. His intention was to briefly document the early days of this potentially illuminating free-for-all, as a sidebar to his primary field of research on a group of lymph node disorders called Castleman disease. But now, 11 months and 29,000 scientific papers later, he and his team of 22 are still going strong.

Keep Reading Keep Reading
Julia Sklar
Julia Sklar is a Boston-based independent journalist who covers science, health, and technology. You can follow her on Twitter at @jfsklar.

Leading medical and scientific experts will discuss the latest developments around the COVID-19 vaccines at our March 11th event.

Photo by Daniel Schludi on Unsplash



Thursday, March 11th, 2021 at 12:30pm - 1:45pm EST

On the one-year anniversary of the global declaration of the pandemic, this virtual event will convene leading scientific and medical experts to discuss the most pressing questions around the COVID-19 vaccines. Planned topics include the effect of the new circulating variants on the vaccines, what we know so far about transmission dynamics post-vaccination, how individuals can behave post-vaccination, the myths of "good" and "bad" vaccines as more alternatives come on board, and more. A public Q&A will follow the expert discussion.

Keep Reading Keep Reading
Kira Peikoff
Kira Peikoff is a journalist whose work has appeared in The New York Times, Newsweek, Nautilus, Popular Mechanics, The New York Academy of Sciences, and other outlets. She is also the author of four suspense novels that explore controversial issues arising from scientific innovation: Living Proof, No Time to Die, Die Again Tomorrow, and Mother Knows Best. Peikoff holds a B.A. in Journalism from New York University and an M.S. in Bioethics from Columbia University. She lives in New Jersey with her husband and son.