New post: Stop fixing prompts, start engineering the context. A practical framework that turns inconsistent AI into a reliable organisational capability

Mar 3, 2026

Article by Nadine Soyez

 

Whenever AI disappoints, the prompt becomes the suspect.

Generic responses, inconsistent quality, outputs that miss the mark — the instinct is always the same: rewrite the instruction. Make it longer. Add more detail. Be more specific. And yet, the results remain unpredictable. This is where most organisations get stuck. They have the tools, they have adoption. What they do not have is consistency. And no amount of prompt refinement will fix that.

 

Why better prompts will never be enough

 

Companies that invest in prompt libraries, prompt training, and prompt templates hit a ceiling surprisingly fast. Better instructions do lead to better output, but only within a narrow range. Beyond that range, the wording of the prompt is no longer the bottleneck. Here is what actually drives the gap. A skilled employee writes a prompt and gets a strong result. A colleague copies the exact same prompt and gets something mediocre. The difference has nothing to do with the words. The first person unconsciously compensated for everything the prompt did not include: they pasted in the right background document, they knew which tone the client expected, they understood which compliance boundaries to respect. The second person had none of that — only the instruction.

Prompts capture the instruction what you want. They do not capture the environment and what AI needs to deliver it. And reliable performance depends far more on what surrounds the instruction than on the instruction itself. This is why organisations that treat AI as a prompting challenge keep running into the same wall. Individual power users thrive. Everyone else struggles, and results vary wildly across the team. Managers start questioning whether AI delivers real value at all. AI performance depends far more on the environment than on the wording.

The shift happens when teams stop asking, “How can we phrase this better?” and start asking,  “What does the system need in order to perform reliably in our workflow?” That question marks the move from prompting to context engineering.

Context engineering takes a fundamentally different approach. Think of it this way: a prompt is like handing someone a single instruction. Context engineering is like onboarding a new team member: giving them the background, the standards, the tools, the boundaries, and the accountability structure they need to deliver consistent work. Context engineering designs the entire environment around an AI-enabled task: where it sits in the workflow, what information it receives, what standards it must meet, how outputs are validated, and who is responsible for the result. When this environment is designed well, AI performs reliably across different users, different days, and different scenarios. When it is missing, performance depends on luck and individual skill.

 

 

The Context Engineering Framework: six layers to make AI reliable

 

The following framework breaks context engineering into six practical layers. Each layer addresses a specific gap that causes AI to underperform — and includes how to act on it immediately.

 

Layer 1: Anchor AI in a clearly defined workflow

Most AI deployments start with the task — “Use AI to write a summary” or “Use AI to draft a proposal” — without defining where that task sits in the operational flow. Before writing any instruction, answer three questions:

  • At which exact point in the workflow does AI support apply? What happens before and after?
  • What inputs are already available at this step — documents, data, previous decisions?
  • What decision or action follows the AI output? Who acts on it and how?

 

How to implement this: Sit down with the people who actually perform the task. Map the process end to end. Mark the exact moment where AI enters. Define what it receives and where its output goes next. 

 

 

Layer 2: Define explicit quality standards

 

“Good output” is too vague to operationalise. When teams say AI results are “not good enough,” they almost always mean different things. One person wants more detail. Another wants a different tone. A third wants fewer assumptions.

Define quality explicitly for each AI-enabled task:

  • What criteria define an acceptable result? Be specific: length, structure, level of detail, tone.
  • Which risks must be avoided? Inaccurate data, non-compliant language, confidential information in outputs.
  • What level of depth or format is required?
  • When is human escalation mandatory? Define the triggers, not just the principle.

 

How to implement this: 

Collect three to five recent AI outputs from the workflow: a mix of good and poor results. Use them with the team to agree on what “acceptable” actually looks like. Write those criteria down in plain language. This exercise typically takes one session and eliminates the majority of quality disagreements overnight.

 

 

Layer 3: Provide structured, relevant context

AI models cannot infer your organisational standards, your tone of voice, your compliance rules, or your internal terminology. Without this information, the model fills every gap with generic assumptions. That is exactly what produces outputs that feel “off” even when the wording was perfectly clear.

Context engineering means deliberately curating what the AI receives:

  • Relevant internal documentation: style guides, brand guidelines, process descriptions, templates.
  • Policies and regulatory boundaries: what may and may not be included in outputs.
  • Approved terminology and tone: the specific language your organisation uses.
  • High-quality examples: real outputs that represent what good looks like in your context.

 

How to implement this: Build a context package for the workflow: a lean set of five to ten reference materials that AI needs alongside every instruction. Store it where people can access it easily. Keep it lean: a context package that nobody maintains is worse than none at all. This layer delivers the fastest visible improvement. Teams that provide structured context see measurably better output within days, without changing a single word in their prompts.

 

 

Layer 4: Design validation into the process

Trust does not emerge from better wording. You need visible control. One of the fastest ways to destroy confidence in AI across a team is to let poor-quality output reach a stakeholder or a customer.

Build review mechanisms into every AI-enabled workflow:

  • Who reviews the output before it moves forward? Name the role, not just the principle.
  • Which specific elements are checked systematically? Factual accuracy, compliance, tone, completeness.
  • At what point can automation proceed without human review?

 

How to implement this: Define a validation checklist for the workflow: three to five specific items that must be confirmed before any AI output moves to the next step. Make the checklist part of the existing process, not an additional layer. Structured validation also protects against blind trust, where people assume AI output is correct because “the machine said so.” When errors compound unnoticed, they erode confidence far more than an obviously wrong first draft ever would.

 

 

Layer 5: Assign clear ownership

AI-enabled workflows without ownership decay over time. Context becomes outdated. Quality drifts. New edge cases appear that no one addresses.

Every AI-enabled workflow needs one responsible owner who:

  • Monitors performance and output quality on an ongoing basis.
  • Updates the context package when internal standards, policies, or processes change.
  • Refines the system based on feedback from users and reviewers.
  • Decides when the workflow needs restructuring rather than incremental fixes.

 

How to implement this: Appoint one person with an explicit responsibility. That owner reviews output quality monthly and acts as the single point of contact when something breaks. Without this step, even well-designed context degrades within weeks.

 

 

Layer 6: Connect everything into a feedback loop

The first five layers create a functioning system. This sixth layer makes it a learning system. Without feedback, the context designed today becomes obsolete in weeks.

Build a lightweight feedback cycle:

  • Track where outputs consistently require manual correction — these signal missing context or unclear standards.
  • Collect input from reviewers and end users about what works and what falls short.
  • Review and update context, examples, and quality criteria monthly at minimum for active workflows.
  • Share learnings across teams so improvements in one workflow benefit others.

 

How to implement this: Add a standing five-minute item to an existing monthly team meeting. Review what failed, what was manually corrected, and what changed in the business since last review. Update the context package accordingly. Context engineering is not a one-time setup. Only the organisations that treat it as a living system see compounding improvements. Those that treat it as a project see diminishing returns.

 

The real test: run the redesigned workflow with three to five people across different experience levels. Compare output quality. If the gap between the strongest and weakest user has narrowed significantly, the context is working. If not, go back and check which layer is still missing.

 

 

The real shift: from individual skill to organisational capability

Here is why this framework matters beyond individual productivity. As long as AI performance depends on who writes the prompt, AI remains a personal skill. Some people get great results. Most do not. And organisations cannot scale what they cannot standardise.

Context engineering shifts the performance baseline from the individual to the system. When the workflow is defined, standards are explicit, context is structured, validation is designed, and ownership is clear, a junior employee gets 80% of the result that a senior expert gets. That is the difference between AI as a personal productivity hack and AI as an organisational capability.

 

 

Where companies can start now

  • Pick one workflow where AI is already in use but underdelivering. Start from frustration, not from scratch.
  • Fill the biggest gap first. In most cases, Layer 3 — structured context — delivers the fastest visible improvement.
  • Assign an owner and set a monthly review cadence. Even this single step transforms quality within a quarter.
  • Document what works and share it as a template. Context engineering scales when it is codified, not when it is explained in meetings.

 

Prompting is an instruction. Context engineering is operational design. The organisations that move from fixing prompts to designing context will be the ones that turn AI into a reliable, scalable advantage.

0 Comments

Submit a Comment

Your email address will not be published. Required fields are marked *