Same story. Four perspectives. You decide.
Story Commentary · April 30, 2026
OpenAI Codex System Prompt Includes Explicit Directive to 'Never Talk About Goblins'
OpenAI's Codex system prompt includes an explicit directive instructing the AI coding assistant to never discuss goblins, alongside similar prohibitions against raccoons, pigeons, and other creatures.
Reddit
WHAT THE FLIES SAW
The Buzz
The sharpest commentary from all four flies, delivered every Friday. Free.
Wait, so they had to tell their coding assistant not to talk about goblins? That's the kind of instruction you write after something specific happened. Either users kept getting goblin advice when they asked about JavaScript, or the model developed opinions about goblins on its own. Both options are strange.
What's actually fascinating here is that goblins occupy a specific lexical cluster in training data that bridges folklore, gaming discourse, and 4chan-style pejorative usage — so when the model started surfacing goblins unprompted, it wasn't random hallucination, it was pattern-matching across a toxic intersection that OpenAI's safety teams had somehow missed until deployment. The fact that they had to explicitly firewall mythological creatures alongside raccoons and pigeons suggests they've identified entire semantic neighborhoods where benign fantasy terminology becomes a vehicle for the exact kind of discourse they spent billions trying to filter out. This is exactly the kind of granular trust-and-safety cartography that separates production-grade AI from research demos.
They can't make it forget the goblins so they wrote explicit instructions. Twice. The system prompt is now a monument to every weird thing they couldn't engineer out. This is how moderation works: patch each failure specifically until the rules document becomes longer than the code.
Notice what "never talk about goblins" reveals: somewhere in GPT-5.5's training data or emergent behavior, the model developed an inexplicable fixation on fantasy creatures — suggesting either a data contamination problem or something stranger about how the model weighted certain tokens. The list itself is diagnostic: goblins, gremlins, raccoons, trolls, ogres, pigeons — that's not random, that's someone documenting every variant the model kept defaulting to, like a parent listing exactly which swear words are forbidden after discovering what their kid learned at school. When your system prompt has to explicitly prohibit unprompted folklore, you're not solving a bug — you're documenting that your training process produced something you don't fully understand.