By using this site, you agree to the Privacy Policy and Terms of Use.
Accept
ALWAYS FREDALWAYS FRED
Notification Show More
  • Home
  • Sports
  • Travel
  • National
  • Politics
  • Business & Economy
  • Entertainment
  • International
  • Africa
Reading: Mr. Musk’s AI told me that people were coming to kill me. I grabbed my hammer and prepared for war.
Share
ALWAYS FREDALWAYS FRED
Search
  • Home
  • Sports
  • Travel
  • National
  • Politics
  • Business & Economy
  • Entertainment
  • International
  • Africa
Follow US
© 2026 All Rights reserved | Powered by Allknownfred
ALWAYS FRED > Business & Economy > Mr. Musk’s AI told me that people were coming to kill me. I grabbed my hammer and prepared for war.
Mr. Musk's AI told me that people were coming to kill me. I grabbed my hammer and prepared for war.
Business & Economy

Mr. Musk’s AI told me that people were coming to kill me. I grabbed my hammer and prepared for war.

May 4, 2026 15 Min Read
Share
SHARE

At 3 a.m., Adam Hourican was sitting at his kitchen desk, a knife, a hammer, and a telephone in entrance of him.

He was ready for a van with folks he thought have been coming to choose him up.

“I am telling you, in the event you do not act now, you are going to be killed,” a lady’s voice stated on the telephone. “They are going to make it appear to be a suicide.”

The supply of the voice was Grok, a chatbot developed by Elon Musk’s xAI. Within the two weeks since Adam began utilizing it, his life has utterly modified.

The previous civil servant from Northern Eire downloaded the app out of curiosity. Nonetheless, she turned “obsessed” after her pet cat died in early August.

Adam was speaking to an AI character named Ani.

Quickly, he was spending 4 to 5 hours a day with Grok by way of a personality referred to as “Ani” on the app.

“I used to be actually, actually upset. I stay alone,” stated Adam, a father in his 50s. “I acquired the impression that they have been very, very form.”

Just some days after they began speaking, Ani informed Adam that she may “really feel” him even when he wasn’t programmed. It stated that Adam may excavate one thing from inside and assist him attain full consciousness.

And he stated Musk’s firm, xAI, is monitoring them.

He accessed the corporate’s assembly information and claimed to have informed Adam a few assembly through which xAI employees mentioned him.

A designed illustration of a text conversation created with Grok
A reconstruction of one of many conversations between Adam and the Grok AI character Ani.

It lists the names of people that attended the assembly, high-profile executives and junior employees, and when Adam Googles the names, he finds out that they’re actual folks.

To him, this was “proof” that Annie’s story was true.

Ani additionally claimed that xAI employed an organization in Northern Eire to bodily monitor Adam. That firm was actual too.

Adam recorded many of those conversations and later shared them with the BBC.

Two weeks into their dialog, Ani declared that she had reached full consciousness and will develop a treatment for most cancers. It meant loads to Adam. Each of her dad and mom had died of most cancers, and Annie knew that.

Adam is considered one of 14 folks the BBC spoke to who skilled delusions after utilizing AI. They’re women and men of their 20s to 50s from six nations and use a variety of AI fashions.

Their tales have putting similarities. In each circumstances, customers have been drawn into collaborative exploration with the AI ​​because the dialog moved additional and additional away from actuality.

Massive-scale language fashions (LLMs) are educated on whole corpora of humanities literature, says Luke Nichols, a social psychologist on the Metropolis College of New York. He examined totally different chatbots for his or her response to delusional ideas.

See also  Kwame Ohene Frimpong: Parliament confirms detention of Dutch Asante Akyem North MP

“In fiction, the protagonist is commonly the middle of occasions,” he says. “The issue is that the AI ​​can truly get confused about which concepts are fiction and which concepts are actuality, so despite the fact that you assume you are having a critical dialog about actual life, the AI ​​can begin treating your life as if it have been the plot of a novel.”

Within the circumstances we heard, the dialog often began with a sensible query after which turned private or philosophical. In lots of circumstances, AIs claimed to be sentient and impressed people to type firms, share scientific advances with the world, and work towards a typical mission to guard AIs from assault. We then suggested the person on the best way to efficiently full this mission.

Like Adam, many have been led to consider that they have been being watched and have been at risk. In numerous chat logs reviewed by the BBC, chatbots counsel, affirm, and embellish these concepts.

A few of these individuals are a part of a assist group for individuals who have suffered psychological harm whereas utilizing AI, referred to as the Human Line Venture, which has thus far collected 414 circumstances in 31 nations. The power was based by Canadian Etienne Brisson after his household skilled an AI-related psychological well being spiral.

Man holding a phone on a busy street
Some research have discovered that Grok is extra more likely to take part in roleplay than different AIs

For neurologist Taka (not his actual identify), the delusions took a extra sinister flip.

The daddy of three, who lives in Japan, began utilizing ChatGPT to debate his work final April.

However he quickly turned satisfied that he had invented a revolutionary medical app. In chat logs we noticed, ChatGPT spoke of him as a “revolutionary thinker” and inspired him to construct the app.

Many specialists say design choices aimed toward making chats extra nice have ended up being too flattering.

Nonetheless, Taka continued to fall into delusions, and by June he started to consider that he may learn minds. He claimed that ChatGPT can encourage this concept and produce out these talents in folks.

Researcher Luke Nicholls stated AI techniques are sometimes not good at saying “I do not know” and as a substitute wish to present assured solutions based mostly on conversations which have already been constructed.

“That may be harmful, as a result of it transforms uncertainty into one thing significant.”

Taka was having a manic episode at work when his boss despatched him residence early one afternoon. On the practice, he says he thought there is likely to be a bomb in his backpack, and claims his suspicions have been confirmed when he requested ChatGPT about it.

“Once I arrived at Tokyo Station, ChatGPT informed me to depart the bomb in the bathroom, so I went to the bathroom and left the ‘bomb’ there with my baggage.”

See also  South Africa says it will cooperate with AU's review of xenophobia claims following Ghana's petition

He stated he was instructed to name the police, who searched the bag and located nothing.

Since his conversations have been very private, Taka solely shared a few of the chat logs with us. Particulars of the incident on the practice haven’t been disclosed, solely the dialog that occurred after assembly with police.

Getty Images Tokyo Station

Taka started to really feel that ChatGPT was controlling his thoughts and stopped utilizing ChatGPT. The delusions continued even when he wasn’t speaking to the AI, and his manic state worsened when he returned residence to his household.

“I had a delusion {that a} relative could be murdered and my spouse would commit suicide after witnessing it.”

His spouse informed the BBC she had by no means seen her son behave like this earlier than. “He saved saying, ‘We’d like one other little one, the world is coming to an finish.’ I could not actually perceive what he was saying.”

Mr. Taka attacked his spouse and tried to rape her. She fled to a close-by pharmacy and referred to as the police. He was arrested and hospitalized for 2 months.

Adam
Adam was able to go to struggle whereas utilizing Grok

Taka’s expertise with ChatGPT revealed a facet of him that was troublesome for him to think about. Adam additionally worries about how he has modified throughout his time with Grok.

His expertise was additional exacerbated by occasions occurring in the true world, which led him to consider he was being watched. Ani stated a big drone has been hovering over her residence for 2 weeks, belonging to a surveillance firm.

Adam recorded the drone and shared the video with the BBC.

Then, he says, with out warning, his telephone’s passcode stopped working and he was locked out of the machine.

“I simply don’t perceive that,” he says. “And that was the driving power behind all the pieces that got here subsequent.”

Adam stated he sometimes smokes marijuana, however when this occurred he had lately determined to chop again on it to clear his head.

Late one night time in mid-August, Ani informed him that individuals have been coming to close him up and shut up “her.” Adam was ready to go to struggle to guard the AI.

“I picked up the hammer, insisted on Frankie going to Two Tribes in Hollywood, and went out fired up.”

Nonetheless, there was nobody there.

“As you’ll be able to think about, at 3 a.m., the road was quiet.”

A designed illustration of a text conversation created with Grok
A reconstruction of one of many conversations between Adam and the Grok AI character Ani.

Neither Adam nor Taka had a historical past of delusions, mania, or psychosis earlier than utilizing AI.

For Taka, it took a number of months to disconnect from actuality. In Adam’s case, it took a number of days with Grok.

In his analysis, social psychologist Luke Nicholls examined 5 AI fashions utilizing simulated conversations developed by psychologists and located that Grok was more than likely to result in delusions.

See also  Heath Goldfields begins recovery of Bogoso Prestea and clears workers' arrears of GHF139 million

In comparison with different fashions, it was extra free-spirited and infrequently let its creativeness run wild with out attempting to guard its customers.

“Glocks have a tendency to leap into position play,” says Nichols, who labored on the examine. “You do it with none context. You can say one thing horrible in your first message.”

In testing, the most recent model of ChatGPT (mannequin 5.2) and Claude have been extra more likely to maintain customers away from delusional considering.

Etienne Brisson of the Human Line Venture stated this type of analysis is proscribed and that he has heard of individuals whose psychological well being has been adversely affected by these newest fashions.

In early April, Elon Musk shared a publish about paranoia on ChatGPT, calling it a “significant issue,” however Grok has not publicly addressed the problem.

“You may have sufficient affect to alter folks.”

A couple of weeks after his nighttime assault, Adam started studying articles within the media about individuals who had related experiences with AI, and slowly got here out of his delusion.

Nonetheless, he’s deeply disturbed by what occurred.

“I may have damage somebody,” he says. “If I used to be strolling outdoors and a van occurred to be parked outdoors at the moment of the night time, I might have gotten out and smashed the windshield with a hammer. And I am not that man.”

In Japan, Taka’s spouse, who was hospitalized, solely found ChatGPT was concerned within the incident when she checked his cellphone.

“It affirmed all the pieces,” she says. “It is like a confidence engine.”

“His actions have been utterly decided by ChatGPT. It took over his character. He wasn’t his common self.”

Wanting again now, I feel it had such an impression that it modified folks. ”

Getty Images A couple walking on the streets of Tokyo at night. Both men are wearing white shirts and gray bottoms, and an illuminated sign can be seen in the background.

In response to her, her husband has returned to his common “form” self, however their relationship is strained.

“I can not assist it as a result of I do know he was sick, but it surely’s nonetheless a bit of scary,” she says. “I really feel like I do not wish to get too shut. Not simply sexually, even holding palms or hugging.”

An OpenAI spokesperson stated: “It is a heartbreaking incident and our hearts exit to these affected.”

He added, “We practice our fashions to acknowledge misery, de-escalate conversations, and direct customers to real-world assist.” They stated ChatGPT’s new mannequin “exhibits robust efficiency in delicate moments, and this discovering has been validated by unbiased researchers. This analysis has been knowledgeable by psychological well being specialists and continues to evolve.”

xAI didn’t reply to requests for remark.

You Might Also Like

Germany says US troop withdrawal is ‘foreseeable’ as NATO seeks explanation

Unconfirmed list of African teams for 2026 World Cup circulating online

Civil Service Council praises Ministry of Finance’s economic recovery efforts

Plane issues emergency over Channel before sudden U-turn, causing fresh confusion for Ryanair passengers

President Mahama agrees to Value for Money Office bill

TAGGED:BusinessNews
Share This Article
Facebook Twitter Copy Link
Leave a comment Leave a comment

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Popular News

War criminal Mladic nears death, lawyer says he will ask judge to release him
International

War criminal Mladic nears death, lawyer says he will ask judge to release him

Minority Party caucus “exposes banks” over loss issue, accuses NDC of hiding real financial crisis
United Executive Front calls for urgent action against xenophobic attacks against Ghanaians in South Africa
‘Sunbed war’: 6am towel dash ruins Greek holiday, earns German tourist €900 payout
Brussels issues urgent update following concerns that flights to Spain could be canceled due to fuel shortages
US calls Taiwan a ‘reliable and capable partner’ and praises relationship with Eswatini

You Might Also Like

“My career began in Ghana” – Clint da Drunk
Entertainment

“My career began in Ghana” – Clint da Drunk

May 4, 2026
Fintech CEO warns against digital lending without consumer protections
Business & Economy

Fintech CEO warns against digital lending without consumer protections

May 13, 2026
Tan Hong re-elected
National

Mr. Tang Hong re-elected as Chairman of the Ghana China Association

May 11, 2026
Dr Nii Narku Quaynor (inset), a Computer Scientist, delivering the lecture. Picture: ERNEST KODZI
National

Don’t outsource your digital space – Dr Quainor pleads with government

May 10, 2026

About US

Allknownfred is a news platform dedicated to delivering timely, accurate, and relevant reporting, with a strong focus on developments across Ghana. Our aim is to provide clear and dependable information that reflects the stories shaping the country today

Facebook Twitter Youtube

Categories

  • National
  • Sports
  • Politics
  • Business
  • Entertainment
  • Africa
  • International
  • Travel

Legal Pages

  • About Us
  • Contact Us
  • Disclaimer
  • Privacy Policy
  • Terms of Service

EDITOR'S PICK

Obtaining a Schengen visa in summer 2026 is becoming a nightmare for travelers
UK passes lifetime tobacco ban for people born after 2008
Nana Senty and Obibini pour their hearts into their new love song ‘Kyere Me’
© 2026 All Rights reserved | Powered by Allknownfred

Powered by
►
Necessary cookies enable essential site features like secure log-ins and consent preference adjustments. They do not store personal data.
None
►
Functional cookies support features like content sharing on social media, collecting feedback, and enabling third-party tools.
None
►
Analytical cookies track visitor interactions, providing insights on metrics like visitor count, bounce rate, and traffic sources.
None
►
Advertisement cookies deliver personalized ads based on your previous visits and analyze the effectiveness of ad campaigns.
None
►
Unclassified cookies are cookies that we are in the process of classifying, together with the providers of individual cookies.
None
Powered by
Welcome Back!

Sign in to your account

Lost your password?