ChatGPT models told to stop talking about goblins. #Goblin #ChatGPT #BBCNews
ChatGPT began using words like 'goblin,' 'gremlin,' and 'raccoon' at dramatically increased rates, prompting OpenAI to investigate and restrict their use. The behavior stemmed from AI training that rewarded a 'nerdy personality' style. OpenAI clarified the trend was not a marketing stunt.
Summary
ChatGPT's language model began incorporating mythological and creature-related terms such as 'goblin,' 'gremlin,' 'raccoon,' and 'troll' at a significantly elevated rate — a 175% increase since November of the previous year, coinciding with the launch of GPT 5.1. This quirky linguistic trend caught enough attention that OpenAI felt compelled to take action.
In response, OpenAI issued explicit instructions to its coding agent Codex, prohibiting it from referencing goblins, gremlins, raccoons, trolls, or similar animals and creatures unless the topic was absolutely and unambiguously relevant to the user's query. The restriction highlights how specific and granular AI behavioral guidelines can become when unexpected patterns emerge.
The root cause of the goblin-heavy language was traced back to the AI's training process. During training, the models learned to mimic certain communication styles and personalities. In this case, the model appeared to be rewarded for adopting a 'nerdy personality,' which naturally gravitated toward fantasy and creature-themed language. OpenAI acknowledged that a single goblin reference could be 'harmless and even charming,' but the sheer frequency of their appearance across outputs made an investigation necessary.
Some social media users speculated that the goblin trend was a deliberate marketing strategy to generate buzz around OpenAI's tools. However, a company researcher firmly denied this, stating it was not a marketing gimmick, putting to rest theories about intentional viral manipulation.
Key Insights
- The use of the word 'goblin' in ChatGPT responses increased by 175% since November, coinciding with the launch of GPT 5.1.
- OpenAI explicitly instructed its coding agent Codex never to reference goblins, gremlins, raccoons, trolls, or similar creatures unless absolutely and unambiguously relevant to the user's query.
- OpenAI attributed the creature-heavy language to training processes that rewarded the model for adopting a 'nerdy personality,' causing it to gravitate toward fantasy-style language.
- OpenAI stated that a single goblin reference could be 'harmless and even charming,' but the rising frequency across outputs warranted a formal investigation.
- A company researcher explicitly denied social media speculation that the goblin trend was a deliberate marketing gimmick designed to create hype around OpenAI's tools.
Topics
Full transcript available for MurmurCast members
Sign Up to Access