Even virtual assistants are sexually harassed

Microsoft: When Cortana launched in 2014, many early queries were about her sex life.

All virtual assistants have to deal with inappropriate comments and questions. From seasoned vets like Siri and Google Now to the rash of new specialists with names like Amy, Molly, Mia, and Robin.

When Microsoft launched Cortana in 2014, a good chunk of early queries was about her sex life, according to Microsoft’s Deborah Harrison.

You might also like: Sex Robot Has Traumatic Experience At Technology Festival And Is Forced To Retire Many Body Parts

It turns out people feel very comfortable talking freely with text and voice assistants. Humanizing the bots with names, faked emotions, personalities and genders (mostly female) helps build trust with users.

A side effect of creating friendly female personalities is that people also want to talk dirty, confess their love, role play or bombard them with insults.

Cortana is not about to put up with it.

You might also like: This financial dominatrix has made $1M in cryptocurrency

“If you say things that are particularly a**holeish to Cortana, she will get mad,” said Harrison during a talk at the Re•Work Virtual Assistant Summit in San Francisco last week. “That’s not the kind of interaction we want to encourage.”

Harrison is one of eight writers who create Cortana’s dialogue in the U.S. In addition to writing jokes and coming up with casual banter, her team has to figure out the best way to shut down vulgar conversations.

Cortana is clearly identified as a woman. She has a female avatar and is voiced by human woman Jen Taylor. But the writers are conscious about avoiding female-assistant stereotypes. Cortana isn’t self-deprecating and avoids saying sorry.

“We wanted to be very careful that she didn’t feel subservient in any way … or that we would set up a dynamic we didn’t want to perpetuate socially,” said Harrison.

You might also like: AI creates fake celebrity porn for Redditors to fap to

Not all assistants will take the same firm approach. Robin Labs, which makes a voice-assistant for drivers, thinks there might be a market customizing personalities. CEO Ilya Eckstein says there is a high demand for an assistant personality that’s “more intimate-slash-submissive with sexual undertones.”

Robin has held more than 100 million conversations. Looking through the data, Eckstein found people fall into a few basic categories. Some like lots of friendly banter, others want just the facts with no sass. Apparently, one group is quite fond of making Robin repeat profanities in her soft female voice.

Sometimes people flirt, not realizing the assistants are just programs. Sometimes, as with Facebook’s (FB) new M Messenger tool, there’s a real human doing some of the work. And though she is cheerful and friendly, x.ai’s meeting-scheduling robot Amy always keeps it professional. That hasn’t stopped people from sending her flowers, chocolates, and whiskey. One user even asked her out on a date (Amy declined).

One smart way to learn how to handle harassment is to talk to real assistants, which Microsoft did for Cortana. After all, they’ve been putting up with similar behavior for years.

You might also like: New Male Sex Robots With Bionic Penises May Just Replace Men For Good!