AI photo
SHARE

Alexa, Amazon’s virtual assistant, was laughing when it shouldn’t. You might have seen tweets about it: The weird, disembodied chuckle bothered people for reasons you can imagine, as well as because it reportedly could happen unprompted—our assistants, after all, are only supposed to listen and speak to us after they hear the wake word. We want them to tell us the weather and set kitchen timers on command, not spook us with laughter. (Amazon has both acknowledged the problem and pushed out a fix.)

We all know that virtual personas like Alexa, Siri, and the Google Assistant are not real humans. They can’t laugh the way we laugh, because they are not alive. But it does, in fact, makes sense for them to try to be social and emulate our behavior. They just need to get it right. Telling jokes, for example, is not an essential skill for an assistant like Alexa to have, but it will still cough one up if you ask.

“We’re social beings—we should build machines that are social as well,” says Timo Baumann, a systems scientist in the language technologies institute at Carnegie Mellon University. He researches the interactions that occur between people and virtual assistants, and between people and other people. “Our research shows that it’s super important to try to build a relationship with a user,” he adds. Ideally, there should be a rapport there.

At Carnegie Mellon, for example, Baumann tested a virtual agent—an animated one that participants could both speak with and see on a screen. Its job was to give movie recommendations, and there were two versions: one was programmed to be more social than the other. “They [the study participants] liked this more social system significantly more, and they liked the movie recommendations significantly more,” he says. That was true even though they hadn’t changed the part of the system that made the film suggestions—it’s just that the more social persona resulted in higher ratings.

Of course, anyone who has interacted with an assistant like Alexa or Siri knows that their social chops are works in progress. “The critical problem is that being social is super difficult,” Baumann says—the best they can try to do now is approximate it. He adds: “If you’re not correctly social, then it’s creepy.”

In fact, truly being social is different from telling a joke on command or laughing when asked, Baumann argues. Ideally, a virtual agent could do things like read the mood of the person with whom they are speaking, and adjust accordingly. That’s much harder to do than simply regurgitating a joke.

But while those more advanced moves aren’t here, it still remains important that the assistants we do speak with interact with us in ways that create rapport. “The machine is trying to build trust,” he says. “Destroying trust is so much easier than building it.” On a simpler level, that’s what happening when an assistant like Siri or Alexa responds to an inquiry correctly—or flubs it.

In the case of the weird laugh, Amazon has provided an explanation, and made an adjustment. A spokesperson said via email: “In rare circumstances, Alexa can mistakenly hear the phrase ‘Alexa, laugh.’ We are changing that phrase to be ‘Alexa, can you laugh?’ which is less likely to have false positives, and we are disabling the short utterance ‘Alexa, laugh.’ We are also changing Alexa’s response from simply laughter to ‘Sure, I can laugh’ followed by laughter.”

If you’re curious, the new laughter sound in this case is “teehee.” And while the fix doesn’t exactly mean Alexa is now suddenly perfectly social, at least it’s no longer being antisocial in this particular way. But if a virtual assistant could read your mood and maybe even laugh at the right time, the way a friend would, that might not be creepy. It might be friendly.