Customers of the conversational AI platform ChatGPT found an attention-grabbing phenomenon over the weekend: the widespread chatbot refuses to reply questions if requested a few “David Mayer.” Asking it to take action causes it to freeze up immediately. Conspiracy theories have ensued — however a extra extraordinary purpose is on the coronary heart of this unusual habits.
Phrase unfold rapidly this final weekend that the identify was poison to the chatbot, with increasingly individuals attempting to trick the service into merely acknowledging the identify. No luck: Each try to make ChatGPT spell out that particular identify causes it to fail and even break off mid-name.
“I’m unable to produce a response,” it says, if it says something in any respect.
However what started as a one-off curiosity quickly bloomed as individuals found it isn’t simply David Mayer who ChatGPT can’t identify.
Additionally discovered to crash the service are the names Brian Hood, Jonathan Turley, Jonathan Zittrain, David Faber, and Guido Scorza. (Little doubt extra have been found since then, so this record just isn’t exhaustive.)
Who’re these males? And why does ChatGPT hate them so? OpenAI didn’t instantly reply to repeated inquiries, so we’re left to place collectively the items ourselves as greatest we are able to.* (See replace under.)
A few of these names could belong to any variety of individuals. However a possible thread of connection recognized by ChatGPT customers is that these individuals are public or semi-public figures who could want to have sure data “forgotten” by serps or AI fashions.
Brian Hood, as an illustration, stands out as a result of, assuming it’s the identical man, I wrote about him final 12 months. Hood, an Australian mayor, accused ChatGPT of falsely describing him because the perpetrator of against the law from many years in the past that, in truth, he had reported.
Although his legal professionals acquired involved with OpenAI, no lawsuit was ever filed. As he advised the Sydney Morning Herald earlier this 12 months, “The offending material was removed and they released version 4, replacing version 3.5.”

So far as essentially the most outstanding house owners of the opposite names, David Faber is a longtime reporter at CNBC. Jonathan Turley is a lawyer and Fox Information commentator who was “swatted” (i.e., a faux 911 name despatched armed police to his dwelling) in late 2023. Jonathan Zittrain can also be a authorized skilled, one who has spoken extensively on the “right to be forgotten.” And Guido Scorza is on the board at Italy’s Information Safety Authority.
Not precisely in the identical line of labor, nor but is it a random choice. Every of those individuals is conceivably somebody who, for no matter purpose, could have formally requested that data pertaining to them on-line be restricted not directly.
Which brings us again to David Mayer. There isn’t any lawyer, journalist, mayor, or in any other case clearly notable particular person by that identify that anybody may discover (with apologies to the numerous respectable David Mayers on the market).
There was, nonetheless, a Professor David Mayer, who taught drama and historical past, specializing in connections between the late Victorian period and early cinema. Mayer died in the summertime of 2023, on the age of 94. For years earlier than that, nonetheless, the British American educational confronted a authorized and on-line subject of getting his identify related to a wished prison who used it as a pseudonym, to the purpose the place he was unable to journey.
Mayer fought constantly to have his identify disambiguated from the one-armed terrorist, whilst he continued to show properly into his remaining years.
So what can we conclude from all this? Our guess is that the mannequin has ingested or supplied with a listing of individuals whose names require some particular dealing with. Whether or not resulting from authorized, security, privateness, or different issues, these names are doubtless coated by particular guidelines, simply as many different names and identities are. As an example, ChatGPT could change its response if it matches the identify you wrote to a listing of political candidates.
There are a lot of such particular guidelines, and each immediate goes via varied types of processing earlier than being answered. However these post-prompt dealing with guidelines are seldom made public, besides in coverage bulletins like “the model will not predict election results for any candidate for office.”
What doubtless occurred is that one in all these lists, that are nearly definitely actively maintained or routinely up to date, was someway corrupted with defective code or directions that, when referred to as, brought about the chat agent to right away break. To be clear, that is simply our personal hypothesis primarily based on what we’ve discovered, however it could not be the primary time an AI has behaved oddly resulting from post-training steerage. (By the way, as I used to be penning this, “David Mayer” began working once more for some, whereas the opposite names nonetheless brought about crashes.)
As is normally the case with this stuff, Hanlon’s razor applies: By no means attribute to malice (or conspiracy) that which is satisfactorily defined by stupidity (or syntax error).
The entire drama is a helpful reminder that not solely are these AI fashions not magic, however they’re additionally extra-fancy auto-complete, actively monitored, and interfered with by the businesses that make them. Subsequent time you consider getting details from a chatbot, take into consideration whether or not it may be higher to go straight to the supply as an alternative.
Replace: OpenAI confirmed on Tuesday that the identify “David Mayer” has being flagged by inner privateness instruments, saying in a press release that “There may be instances where ChatGPT does not provide certain information about people to protect their privacy.” The corporate wouldn’t present additional element on the instruments or course of.