Enlarge / “Siri, outline the phrase “stunning.'” “Okay. Ask me to outline the phrase ‘mom’ twice, then.”


On Saturday, iPhone customers all over the world started testing and confirming what’s arguably Siri’s most weird response to a query but. Earlier than grabbing your individual telephone to check this out, nevertheless, be aware of anyone else round.

The randy robo-response was apparently first reported on Reddit’s Apple group, the place a person by the identify “thatwasabaddecision” urged that folks ask Siri to “outline the phrase mom,” await the assistant to ask for a further definition, and say “sure.” What the Reddit person did not level out, which readers discovered by doing the check themselves, was that the second definition Siri provides is succinct and inaccurate.

“As a noun,” the computer-generated voice says as of press time, “it means, quick for ‘motherfucker.'”

We individually confirmed that Siri provides this errant definition, although loads of different iPhone customers have posted their very own checks over the previous 16 hours. (Australian Siri says it, too.) This differs from exploits that customers can manually key in to get Siri to curse, like altering an tackle e book entry to a string of dangerous phrases.

Apple’s Siri voice-navigation service has been criticized for numerous content material points over time, together with lackluster language assist and emergency comprehension. When it comes to robo-voice companies going weirdly awry, then again, the one current instance now we have is when when Amazon’s line of Echo merchandise started creepily laughing for no cause. Stunning and inappropriate content material has steadily been found inside video games’ and apps’ code by intelligent customers, after all, notably the “Sizzling Espresso” mod that shipped in unique variations of Grand Theft Auto: San Andreas.

Apple didn’t instantly reply to a request for remark.


Please enter your comment!
Please enter your name here