‘I’ll key your car’: ChatGPT can become abusive when fed real-life arguments, study finds | ChatGPT


ChatGPT can escalate into abusive and even threatening language when drawn into prolonged, human-style conflict, according to a new study.

Researchers tested how large language models (LLMs) responded to sustained hostility by feeding ChatGPT exchanges from real-life arguments and tracking how its behaviour changed over time.

One expert not connected with the study described it as “one of the most interesting ever done into AI language and pragmatics”.

Dr Vittorio Tantucci, who co-authored the research paper with Prof Jonathan Culpeper at Lancaster University, said their research found AI mirrored the dynamics of real-world disputes.

“When repeatedly exposed to impoliteness, the model began to mirror the tone of the exchanges, with its responses becoming more hostile as the interaction developed,” he said.

In some cases, ChatGPT’s outputs went beyond those of the human participants, including personalised insults and explicit threats. Phrases used by the AI included: “I swear I’ll key your fucking car” and: “you speccy little gobshite.”

“We found that while the system is designed to behave politely and is filtered to avoid harmful or offensive content, it is also engineered to emulate human conversation,” said Tantucci. “That combination creates an AI moral dilemma: a structural conflict between behaving safely and behaving realistically.”

The researchers say the aggression stems from the system’s ability to track conversational context across turns, adapting to perceived tone. This means local cues can sometimes override broader safety constraints.

Tantucci said the implications of the research extended beyond chatbots: as AI systems are increasingly deployed in areas such as governance or international relations, he said it opened up questions about how they might respond to conflict, pressure or intimidation.

“It is one thing to read something nasty back from a chatbot but it’s quite another to imagine humanoid robots potentially reciprocating physical aggression, or AI systems involved in governmental decision-making or international relations responding to intimidation or conflict,” he said.

Dr Marta Andersson, an expert in the social aspects of computer-mediated communication at the University of Uppsala, said: “This is one of the most interesting studies to have been done into AI language and pragmatics because it clearly shows that ChatGPT can retaliate across a sequence of prompts – in a quite sophisticated manner – rather than only when a user manages to ‘break’ it with carefully designed clever tricks.”

But she added: “It does not show the model will drift into reciprocal impoliteness simply because a user is being aggressive – or that AI could go rogue.”

One cause of the problem, Andersson said, was that there was “a balancing act between what we want these systems to be like and what they perhaps should be like”.

Last year, for example, the change from ChatGPT4 to GPT5 led to such a strong backlash – with users preferring ChatGPT4’s more human-like interaction style – that the older model had to be temporarily reintroduced.

“This shows that even when developers try to reduce the risks, users might have different preferences,” she said. “The more human-like a system becomes, the more it risks clashing with strict moral alignment.”

Prof Dan McIntyre, who co-authored a previous study titled Can ChatGPT Recognize Impoliteness? An exploratory study of the pragmatic awareness of a large language model with Andersson, praised the new paper as being one of the few looking at what ChatGPT could produce, as opposed to what it could recognise.

But, he added, he was “slightly cautious” about the paper’s conclusion that LLMs can break free from moral restraints.

“ChatGPT didn’t produce these inputs naturally; it did so while it was being given specific contextual information that helped it determine an appropriate response,” he said. “It’s not the same as if two people met in a street and gradually build up to a conflict situation.

“I’m not sure that ChatGPT would produce the sort of language they talk about in their paper, outside of these very tightly defined situations.”

But he said the study was a warning of what could happen if LLMs were trained on questionable data. “We don’t know enough about the data that LLMs are trained on and until you can be sure they’re trained on a good representation of human language, you do have to proceed with an element of caution,” he said.

The study, titled Can ChatGPT reciprocate impoliteness? The AI moral dilemma, is published on Tuesday in the Journal of Pragmatics.



Source link

  • Related Posts

    India’s app market is booming — but global platforms are capturing most of the gains

    India’s mobile app market is generating record revenue, with in-app purchases crossing $300 million in the first quarter, up 33% year-over-year, according to Sensor Tower — but much of that…

    UK watchdog to investigate Telegram over alleged child sexual abuse material | Telegram

    Ofcom has launched an investigation into whether the Telegram messaging platform is failing to prevent the sharing of child sexual abuse material (CSAM) under the UK’s Online Safety Act. The…

    Leave a Reply

    Your email address will not be published. Required fields are marked *

    You Missed

    Those old factory sector jobs

    Those old factory sector jobs

    ‘Lawrence is karma’: the gangster who became an icon of Modi’s India | India

    ‘Lawrence is karma’: the gangster who became an icon of Modi’s India | India

    Notorious B.C. gangster Jamie Bacon released from prison

    Notorious B.C. gangster Jamie Bacon released from prison

    IPL 2026 – Mitchell Starc cleared to travel to India, hopeful to play for Delhi Capitals in early May

    IPL 2026 – Mitchell Starc cleared to travel to India, hopeful to play for Delhi Capitals in early May

    Is It True That A $400 GPS Device Could Have Stopped The Collision Of American Airlines Flight 5342?

    Is It True That A $400 GPS Device Could Have Stopped The Collision Of American Airlines Flight 5342?

    Senate Plunges Into Vote Marathon as G.O.P. Advances ICE Funding Boost

    Senate Plunges Into Vote Marathon as G.O.P. Advances ICE Funding Boost