← All posts
AI / LLM Strategy

Prompt Engineering Is Dead. Long Live Context Engineering.

The shift from clever prompts to context assembly pipelines — why the 'prompt engineer' role is dissolving, and what replaced it.

Craig Hoffmeyer7 min read

In 2023, "prompt engineer" was a job title. In 2024, it was a chapter in every AI book. In 2025, it was starting to feel quaint. In 2026, it is mostly a punchline. Not because prompts stopped mattering — they still matter — but because the center of gravity has moved. The hard part of getting an LLM to do the right thing is no longer the wording of the prompt. It is the pipeline that assembles the context the model gets to see.

I have watched this shift happen across dozens of engagements. Teams that were obsessing over the last 5% of prompt wording in 2024 are now obsessing over their retrieval quality, their tool design, and their context compression in 2026. The move is from "magic words" to "engineered context." This article is about what that shift looks like in practice and why your team probably needs to make it if they have not already.

The old model: the prompt is the feature

The mental model in early 2024 went like this: you had a prompt, the prompt was carefully crafted with examples and instructions, and the prompt was the thing that made the feature work. You iterated on the prompt. You versioned the prompt. You tested the prompt. The prompt was the product.

This worked because models were less capable, context windows were small, and tasks were narrow. A clever prompt could make a meaningful difference between a feature that worked and one that did not. "Let's think step by step" was a genuinely powerful trick. Few-shot examples could carry a task that zero-shot could not.

What changed

Two things changed between 2024 and 2026 that made prompt engineering less important:

Models got better at following simple instructions. The gap between a cleverly-worded prompt and a plainly-worded one shrunk dramatically. Frontier models in 2026 do the right thing when you tell them the right thing in plain English. The room for cleverness is smaller.

Tasks got more complex. The features people build now are not "summarize this paragraph." They are "look at this customer's entire history, figure out what they are asking, decide whether to answer from documentation or escalate to a human, and draft the response." That task has a dozen context inputs, a sequence of tool calls, and multiple branching decisions. The prompt is a small part of the work. The rest is pipeline.

The combination means that the craft of "writing a good prompt" — the skill set that used to be the whole game — is now a subset of a larger craft: engineering the context the model sees.

What context engineering actually is

Context engineering is the work of deciding what the model gets to see, in what order, in what format, at what moment. It includes prompt wording but is much larger than that. The concrete activities:

Retrieval. Finding the right data to include. This is the RAG pipeline work (I wrote a whole post on this) but also covers database lookups, API calls, and whatever else needs to happen to get the right facts in front of the model.

Compression. The context window is finite and expensive. If the raw material is bigger than what fits, you need to compress it without losing the parts the model needs. Summarization, extraction, and filtering are all context engineering tools.

Structuring. The same facts in different orders produce different outputs. A good context engineer knows that the model pays more attention to the beginning and the end of long contexts, that structured headings help the model find what it needs, and that delimiters matter.

Tool design. For agent-like features, the tools the model has access to and the descriptions of those tools are part of the context. A well-designed tool API with clear descriptions is often the difference between an agent that works and one that does not.

State management. In multi-turn interactions, what you carry forward from previous turns is context engineering. What you drop matters as much as what you keep.

Prompt wording. Yes, still. But it is one item on this list, not the whole list.

The new workflow

A team doing context engineering well has a workflow that does not look like the old prompt engineering workflow. The changes:

The prompt is small. Maybe 50–300 tokens of instructions. Mostly role, task, format, and what to do when things go wrong. Not the 2,000-token system prompts with 15 few-shot examples that used to be the norm.

Most of the context is dynamic. Retrieved facts, tool results, summaries, user state. These get assembled per request by the pipeline before the model sees them.

The pipeline is the thing you iterate on. When quality is bad, you do not start by editing the prompt. You look at what the pipeline pulled in and ask whether the right information made it through. Usually it did not.

The eval is per-stage. You can test the retrieval stage independently of the generation stage. You can test the compression stage independently of the tool use stage. The eval is a set of stage-specific tests plus an end-to-end test, not a single black-box test of the whole prompt.

A concrete example

Take a customer-support drafting feature. The team wants the model to read an incoming ticket and draft a reply using the company's voice and knowledge base.

The old approach was: "Write a massive prompt that includes the company voice guide, the top FAQ entries, and a few example tickets. Shove the incoming ticket at the end. Iterate on the prompt until the outputs look good."

The new approach is:

  1. Retrieve the top 3 most relevant knowledge base articles for the ticket (retrieval).
  2. Pull the last 3 interactions the customer has had (state).
  3. Load the company voice guide — but only the 200 most relevant tokens for this type of response (compression).
  4. Assemble these into a structured context with clear section headers (structuring).
  5. Add a small prompt with role, task, and format instructions (prompt wording).
  6. Run the model.

The prompt in step 5 is boring and small. The interesting work is in steps 1–4. When the outputs are bad, the team looks at which step let the model down. Almost always it is a retrieval problem or a compression problem, not a wording problem.

The skills that replaced "prompt engineering"

The valuable skills in 2026 are not the same as they were in 2024. What I look for in people building AI features now:

Information retrieval intuition. Understanding how to find relevant material across different corpora and formats. This is closer to search engineering than to creative writing.

Data pipeline instincts. Comfort with ETL, cleaning, chunking, and embedding. These are boring traditional engineering skills that LLM features happen to need.

Evaluation mindset. The instinct to measure before changing. See why your eval suite matters more than your prompt.

Tool and API design sensibility. For agent features, the ability to design tools that a model can use effectively — clear names, good descriptions, minimal ambiguity, obvious failure modes.

Cost and latency awareness. Knowing the budget implications of adding another retrieval step or another model call. Understanding where the expensive steps are and where the cheap ones are.

What I look for less than I used to: mastery of specific prompt tricks. Cleverness with few-shot examples. Knowledge of jailbreaks and workarounds. These are less important now than they were two years ago, and the trend is continuing.

Where prompts still matter

Prompts are not dead dead. Three places they still matter a lot:

Short, high-volume tasks. Classification, routing, extraction — when you are calling the model thousands of times a day with small inputs, the wording of the system prompt has real impact on both quality and cost.

Agent tool selection. When an agent is choosing between tools, the prompt (and the tool descriptions) is most of the context. Careful wording moves the needle.

Output format constraints. Getting the model to reliably produce valid JSON, specific structured outputs, or particular formats is still a prompt craft problem, though vendor-specific structured output APIs have taken most of this pain away.

So "prompt engineering is dead" is an oversimplification. "Prompt engineering is one slice of a larger discipline" is more accurate. The larger discipline is context engineering.

Counterpoint: not every team needs to go full context engineering

A warning. If you are building a narrow AI feature on small inputs with stable requirements, the old prompt-centric workflow still works. Do not over-engineer a simple feature with a full pipeline if a good prompt is all it needs. The shift I am describing matters for complex features — anything with retrieval, state, or tools — not for every LLM call.

Your next step

Pick the AI feature in your product that has been hardest to improve. List out everything the model sees when it runs that feature. Be honest about what percentage of that context is static prompt wording and what percentage is dynamic retrieved/assembled material. If the balance is 90% prompt and 10% dynamic, you are probably still doing prompt engineering and there is leverage in moving toward context engineering.

Where I come in

Helping teams make the jump from prompt-centric to context-centric workflows is a common part of my AI engagements. Usually 2–4 weeks of pairing with the team on their hardest feature. Book a call if your team is stuck iterating on prompts and not seeing improvement.


Related reading: Your AI Feature Is a RAG Pipeline · Why Your Eval Suite Matters More Than Your Prompt · Agents Are Eating SaaS

Need help on your AI feature's context pipeline? Book a call.

Get in touch →