System Prompt

Picture two authors sitting down with the same AI chatbot to brainstorm a thriller. One gets responses that are punchy, atmospheric, and dripping with noir sensibility. The other gets something that reads like a polite encyclopedia entry. Same model, same question. So what’s different?

Behind the scenes, before either author typed a word, someone had already given the AI its marching orders. Those invisible instructions, the ones the user never sees, are the system prompt. And they shape everything.

The Instructions You Never See

A system prompt is a set of foundational directions given to an AI model before any conversation with a user begins. It defines the model’s role, personality, behavioral boundaries, and priorities for the entire interaction. Think of it as a job description handed to a new employee on day one: before they greet their first customer, they already know what department they work in, what tone to use, what they can and can’t do, and what success looks like.

When you open ChatGPT and type a question, your words are a “user message.” But before your message ever reaches the model, a system prompt, written by OpenAI’s developers, has already told it something like: “You are ChatGPT, a large language model trained by OpenAI. Be helpful, be accurate, and don’t generate harmful content.” You never see this prompt. But every response you receive has been filtered through it.

A system prompt is different from a regular prompt in three important ways. It’s invisible to the end user. It persists for the entire conversation rather than disappearing after a single exchange. And when a user’s request conflicts with the system prompt’s instructions, well-designed models are trained to prioritize the system prompt. It’s the difference between a suggestion and a standing order.

Born from an API

Unlike many AI terms, “system prompt” wasn’t coined in an academic paper or by a visionary researcher. It emerged from a very specific engineering decision on a very specific date.

On March 1, 2023, OpenAI launched the Chat Completions API alongside the model gpt-3.5-turbo. This was the API that let developers build their own products on top of ChatGPT’s technology. And it introduced something new: a structured message format that organized every conversation into three distinct roles. System, user, and assistant.

Before this, OpenAI’s older API was a simple text-in, text-out affair. You sent a block of text, the model continued it. There was no formal way to separate “background instructions for the AI” from “what the human is actually saying.” The Chat Completions API created that separation, and in doing so, gave the concept a name.

OpenAI described the underlying structure, called Chat Markup Language (ChatML), as a way to “make explicit to the model the source of each piece of text” and to “show the boundary between human and AI text.” In plainer terms: they needed the model to know which instructions came from the developer building the app and which came from the person using it. The system role was the solution.

The concept caught on fast. Within months, every major AI platform had adopted some version of it. Anthropic’s Claude, Google’s Gemini, and dozens of smaller models all implemented a system message tier. By July 2023, OpenAI brought the idea to everyday users with Custom Instructions, a feature that essentially gave every ChatGPT user the ability to write their own lightweight system prompt. By November of that year, Custom GPTs took it further, letting anyone build a specialized AI assistant by filling in an “Instructions” field that was, at its core, a system prompt with a friendlier name.

In August 2024, Anthropic made an unusual move: they published Claude’s system prompt publicly. For the first time, users could read the exact hidden instructions shaping their AI assistant’s behavior, everything from “do not start your response with flattery” to how the model should handle requests for sensitive content. It was a transparency gesture that pulled back the curtain on what had always been invisible infrastructure.

What Happens Under the Hood

When a developer sends a request to an AI model like GPT-4 or Claude, the input is structured as a list of messages, each labeled with a role. The system message always comes first:

System: “You are a fiction editor specializing in literary thrillers. Use a direct, constructive tone. Flag weak pacing but don’t rewrite the author’s prose.”

User: “Can you look at this chapter opening for pacing issues?”

Assistant: (the model’s response, shaped by both messages)

The model processes the system message as high-priority context. Its attention mechanisms give those tokens elevated weight, meaning the system prompt’s instructions carry more influence over the response than an equivalent instruction buried in the user’s message. This is why system prompts can maintain a consistent persona across long conversations, even when users change topics, push back, or try to redirect the model.

Different platforms handle this differently. OpenAI’s models give strong priority to the system role. Anthropic’s Claude, by contrast, places relatively more emphasis on user messages compared to OpenAI’s approach. This is one reason the same prompt can feel different in ChatGPT versus Claude: the balance of power between the system instructions and your instructions is calibrated differently.

Why This Matters for Your Writing Life

Understanding system prompts gives you a practical advantage that most AI-using authors don’t have. It lets you see the hidden layer that’s shaping every AI interaction you have.

You’re already using system prompts without knowing it. Every AI writing tool, from Sudowrite to ChatGPT, has a system prompt running backstage. When Sudowrite’s Story Engine generates a chapter in your genre, its system prompt has already told the model to write fiction with specific stylistic and structural priorities. When ChatGPT gives you brainstorming ideas, its system prompt has already set the behavioral frame. Understanding this explains why AI sometimes behaves in ways you didn’t expect: there’s a whole layer of instructions you can’t see that’s influencing the output.

Custom Instructions let you become the director. ChatGPT’s Custom Instructions and Claude’s project-level instructions are, functionally, your personal system prompt. You can tell the AI things like “I write cozy mysteries set in coastal Maine, my protagonist is a retired marine biologist, and I prefer tight third-person POV.” That context shapes every response without you having to repeat it. Authors who use this feature well report dramatically better output, because the model understands your creative world from the very first message.

It explains why the same AI feels different in different tools. The same underlying model (say, Claude or GPT-4) can produce wildly different output in ChatGPT versus Sudowrite versus NovelCrafter. The difference is almost entirely the system prompt. NovelCrafter, for instance, exposes the system prompt directly to authors, letting you write instructions like “You are a copy editor specializing in literary fiction. Maintain the author’s voice. Flag awkward sentences but don’t rewrite them.” That level of control is only possible when you understand what a system prompt is and how it works.

It’s the key to consistency. Without a system prompt, an AI model starts every conversation with a blank slate. It has no memory of your preferences, your project, or your creative style. System prompts (and their consumer-facing equivalents) are the mechanism for giving AI a persistent understanding of who you are and what you’re working on. They’re the reason your AI assistant can feel like a collaborator who knows your book instead of a stranger you have to brief from scratch every single time.

The next time an AI tool gives you a response that feels weirdly off-brand, or strangely perfect, you’ll know there’s a system prompt involved. And if the tool lets you edit it, you’ll know exactly what to do.