Skip to main content

The Context Problem

"Every AI assistant I have tried knows everything about the world and nothing about my world. It can write a perfect OKR framework but doesn't know that we call our quarterly targets 'Boulders' not OKRs. It can produce a flawless meeting summary but doesn't know that when Zara says 'let's take this offline' she means it's politically sensitive and should not be minuted. It knows how to draft a project update but doesn't know that Project Nighthawk is what we call the Karachi expansion internally. I spend five minutes of every conversation re-explaining things that a good colleague would already know."

— Chief of Staff, 300-person technology company


This gap — the distance between how a sophisticated AI answers a question in the abstract and how a knowledgeable colleague answers the same question in context — is the Context Problem.

Every domain chapter in Part 3 of this book solved a specific professional problem. Chapter 28 deployed AI for financial analysis. Chapter 34 automated revenue operations. Chapter 35 optimised supply chains. Chapter 37 (planned) transforms HR. Chapter 38 (planned) builds an operations intelligence layer.

Each of those domain agents is valuable in isolation. Together, they represent something significant: an AI-native organisation where every major work function has a layer of intelligent automation alongside the humans who run it.

But every one of those agents starts each session the same way: from zero.

No memory of the conversation yesterday. No knowledge that "Boulders" means quarterly strategic priorities. No awareness that Project Nighthawk is the Karachi expansion and that it has been stalled for ten days. No understanding that Omar Farooq needs three business days' lead time and dislikes last-minute requests. No knowledge that AgentFactory is the P1 initiative and that anything connected to it carries elevated urgency.

This is not a failure of intelligence. Claude's reasoning capability is not in question — it can analyse as well as expert humans in almost every domain. The constraint is context: the accumulated organisational knowledge that a good colleague carries and that Claude, without memory, does not.

What a Colleague Knows

The difference between a brilliant new hire and an experienced colleague is not intelligence. It is context. The brilliant new hire knows the theory. The experienced colleague knows the organisation.

A good colleague — someone who has worked alongside you for a year — knows:

  • Your name, role, and working style
  • What your current priorities are and what is at risk
  • Who the key people are and how to work with each of them
  • What your organisation calls things, including the terms that would confuse an outsider
  • What meetings exist and what each one is actually for
  • What decisions have been made, why, and what cannot be revisited
  • What is in flight, what is stalled, and what is genuinely urgent

A chatbot knows none of this unless you tell it every single time.

The productivity gap between an AI assistant that starts from zero every session and one that carries meaningful context is enormous. Industry surveys suggest that professionals spend between five and fifteen minutes per AI session re-establishing context that a colleague would already have. Across a working week, that amounts to a significant fraction of AI interaction time spent not on work — but on briefing.

Chapter 39 closes this gap. It introduces the Workplace Memory Architecture — a four-layer context system that gives Claude persistent, structured knowledge of your people, projects, terminology, and priorities. It introduces the plugin system that makes this memory actionable. And it shows how the Digital Chief of Staff emerges: an AI that can act like a knowledgeable colleague, not a generic tool.

But before the solution, it is worth understanding the problem precisely. There are four distinct failure modes, and naming them makes them recognisable — and fixable.

Failure Mode 1: Terminology Blindness

An organisation is, among other things, a vocabulary. It has words for things that outsiders do not know, and it uses those words constantly. Some are official. Many are not.

Panaversity runs on a specific vocabulary. "Boulders" are the quarterly strategic priorities — not "OKRs" or "goals" or "priorities." "AgentFactory" is the internal codename for the AI Agent Factory book project. "Project Nighthawk" is the Karachi expansion — a term used internally; the word "Karachi" is never used in internal communications about it. "Digital FTE" means a fully configured AI agent performing a specific professional role — not a chatbot, not an "AI tool." "The Compass" is the annual strategic planning document reviewed quarterly.

When Zia asks Claude to "write an update on AgentFactory for the quarterly Boulders review," what does Claude produce?

If Claude has no organisational context, it produces something like this:

"I'd be happy to write a project update. Could you clarify what 'AgentFactory' refers to, and what format the quarterly Boulders review uses?"

Two clarifying questions. Neither needed. A colleague would have written the update.

Or, worse, Claude attempts the update and fills in the blanks with generic AI industry language — producing an output that sounds professional but requires complete rewriting because it uses "OKRs" instead of Boulders, calls the project "the AI curriculum" instead of AgentFactory, and refers to "digital agents" instead of Digital FTEs.

Terminology blindness does not prevent work from getting done. It adds a translation layer to every output — a continuous tax paid in manual editing and re-briefing.

Failure Mode 2: People Anonymity

Every organisation is built around relationships. The quality of a communication depends not just on what is said but on who is saying it to whom — and what the recipient's preferences, sensitivities, and current state are.

Omar Farooq is the Head of Analytics at Panaversity. If you know Omar, you know several things: he is data-driven, he expects specific and well-scoped requests, he dislikes last-minute asks, he needs at least three business days' lead time for data work, and the best channel for routine requests is Slack DM while formal requests should go via email. He will push back on scope creep — being vague with Omar is a mistake.

None of this is unusual knowledge. It is the kind of thing any colleague who has worked with Omar for three months would know without thinking about it.

When Zia asks Claude to "draft a message to Omar asking for the analytics brief for the investor deck," what does Claude produce without that context?

A generic, politely worded message that treats Omar as an unknown recipient. Probably slightly too informal given the formality of the request, or slightly too formal given the channel. Almost certainly does not acknowledge the lead time he needs. May not specify the format or deadline clearly enough to avoid a follow-up. Might open with preamble that Omar's direct communication style will find irritating.

The draft needs editing. Not much, perhaps — but every time. Every message to every stakeholder. Because Claude does not know who they are.

People anonymity is the most operationally expensive failure mode for anyone who communicates frequently on behalf of their organisation. A good Chief of Staff knows every key stakeholder well enough to draft a message that sounds like the principal and lands correctly with the recipient. Without people memory, no AI assistant can do this.

Dr. Sana Mirza is joining Panaversity next Monday as Head of Curriculum. She has a PhD from the Aga Khan University in Learning Sciences, brings academic precision and evidence-based rigour, and owns the PHM framework — Panaversity's seven-approach adaptive teaching methodology. She and Omar have not worked together before, and there is a relationship there that will need careful mediation as they define their respective domains.

All of this matters immediately — and none of it is in Claude's context unless it is put there.

Failure Mode 3: Project Amnesia

At any given moment, Panaversity has three primary initiatives in flight.

AgentFactory is P1 — the AI Agent Factory book, currently in Part 3, with a Q2 2026 curriculum launch target. Chapter 39 (this chapter) is in progress. The sequencing of chapters matters — skills files must accompany each chapter, and each chapter depends on the work done in the previous one.

Project Nighthawk is P2 — the Karachi expansion. The facility agreement negotiations have been stalled for over ten days. Every day of delay compresses the Q3 launch timeline downstream. Escalation is overdue.

BankersAI is P2-recurring — the monthly AI upskilling workshop for banking sector professionals. Workshop #7 is the next one, with the topic focused on Digital FTEs for compliance teams. Content must be reviewed seven days before delivery.

Zia knows all of this. Every relevant decision, every current risk, every open action item, every key contact. It is the operating context in which he works every day.

When he asks Claude to "draft a project status update for the board," without that context, Claude asks: "Which project? What is its current status? Who are the key stakeholders? What format does the board prefer?"

Or it drafts something that sounds authoritative but is generic — no mention of the Q3 risk on Nighthawk, no reference to the Q2 launch target for AgentFactory, no indication of what is actually at risk versus what is on track.

Project amnesia means every project briefing starts from scratch. No history. No decisions. No risks. No context about what was agreed in the last review and what changed since then. The AI is always new to the project — no matter how many times you have discussed it.

Failure Mode 4: Priority Blindness

The most subtle failure mode, and often the most costly.

Claude treats all requests as equally important. If you ask it to help draft a message and separately ask it to analyse a strategic document, it will approach both with the same level of care and urgency — because it has no way of knowing that the strategic document is connected to the P1 initiative and has a board deadline tomorrow, while the message is a courtesy communication with a flexible timeline.

In Zia's world, AgentFactory carries elevated urgency by default. Any task connected to it should be treated with more thoroughness, more care, and more urgency than a task connected to a P2 or P3 initiative. A chapter that is "behind schedule" is a different kind of problem from a workshop that is "behind schedule" — the chapter delay has cascading effects on the curriculum launch; the workshop delay is recoverable.

Claude does not know this without being told. And "being told" — every session, for every request — is exactly the problem.

Priority blindness also manifests in how Claude approaches ambiguous requests. When Zia says "can you take a look at this?" with no further context, a colleague who knows the organisation would ask: "Is this urgent? What is it for? How much time should I spend?" Not because they are slow, but because the answer changes significantly based on what it is connected to.

Claude, without priority context, either asks the same question (fine, but it happens every session) or makes an assumption (usually wrong, because the assumption cannot be informed by organisational knowledge it does not have).

The Common Thread

All four failure modes are the same problem viewed from four angles. They are all information problems.

The data exists. Zia knows what "Boulders" means. He knows who Omar is and how to work with him. He knows the status of Project Nighthawk and why it matters. He knows that AgentFactory is P1 and that connected tasks carry elevated urgency.

The failure is not that the information does not exist. The failure is the gap between data and connected, actionable intelligence — and that gap exists because there is no structured mechanism to make the information available to Claude at the start of every session.

This is an architectural problem. And it has an architectural solution.

What Chapter 39 Builds

The next thirteen lessons build the system that closes all four failure modes.

Lessons 2-5 establish the Workplace Memory Architecture — the four-layer context system that gives Claude persistent knowledge of your personal context, your people, your projects, and your organisation. By the end of Lesson 5, every session starts not from zero but from a fully briefed context.

Lessons 6-11 introduce the nine skills that make this memory actionable — task intelligence, delegation quality, the daily digest, meeting support, the executive dashboard, and cross-domain context injection.

Lessons 12-13 introduce the four persistent agents — the Digital Chief of Staff, the Memory Keeper, the Meeting Intelligence agent, and the Work Tracker — that automate the maintenance and activation of this system.

Lesson 14 brings it all together: a complete, configured Agentic Office running as a coordinated system.

The goal is not a better chatbot. It is a Digital Chief of Staff: an AI that knows your world well enough to act like a knowledgeable colleague — one who does not need five minutes of re-briefing at the start of every conversation.

Try With AI

Use these prompts in Cowork or your preferred AI assistant.

Reproduce: Ask Claude to draft a project update without any organisational context — then compare with a fully contextual version.

Write a project update for "Project Nighthawk" for our
Executive Weekly meeting. The update should cover current
status, what is at risk, and what decisions are needed.

Read the output. Note what it does not know — what questions it asks, what generic language it uses, what context it fills in with assumptions.

Now try again with context:

Write a project update for Project Nighthawk (our internal
codename for the Karachi expansion) for the Executive Weekly
(Monday 09:00 PKT, 30-minute standing meeting, no slides).
Current status: PLANNING — AT RISK. The facility agreement
negotiations have been stalled for 10+ days. This is blocking
the Q3 2026 launch target. Escalation to the government
liaison is overdue. Attendees are CEO, COO, and Heads of
Department. Format: brief narrative followed by three
bullet points — status, risk, decision needed.

What you are learning: The second prompt is better — but you wrote forty-five words of context before you could start the actual work. Multiply that across every project, every person, every meeting, every week. The goal of work.local.md is to write that context once and have it available automatically, every session.

Adapt: Audit your own terminology gap.

I work in [describe your organisation and your role in
one sentence]. Here are three terms my organisation uses
that an outsider would not understand:
1. [Term 1] — what it actually means: [definition]
2. [Term 2] — what it actually means: [definition]
3. [Term 3] — what it actually means: [definition]

For each term: explain what generic language an AI would
use instead, and describe the impact on output quality when
the AI uses the generic term rather than our specific one.

What you are learning: Mapping your own terminology gap makes the problem concrete. Most professionals find they have more organisation-specific terms than they initially realise — often twenty to fifty that appear regularly in their work. Each one that Claude does not know is a translation step or a re-briefing cost.

Apply: Map all four failure modes to your own organisation.

I am going to describe four AI context failure modes.
For each one, I want you to help me identify a concrete
example from my own work context.

Failure Mode 1 — Terminology Blindness: AI uses generic
vocabulary when my organisation has specific terms.

Failure Mode 2 — People Anonymity: AI has no knowledge
of who key stakeholders are or how to work with them.

Failure Mode 3 — Project Amnesia: AI has no awareness
of what is currently in flight or what has been decided.

Failure Mode 4 — Priority Blindness: AI treats all
requests as equally urgent regardless of actual stakes.

My role: [describe your role]
My organisation: [describe your organisation in 2-3 sentences]

For each failure mode, help me identify one specific
example from my work where this gap has caused me to
re-brief an AI assistant or manually correct an output.
Then estimate: how many minutes per week does each
failure mode cost me?

What you are learning: Quantifying the time cost of the Context Problem transforms it from an abstract frustration to a business case. The investment in building work.local.md (typically two to four hours, across Lessons 3-5) should be weighed against the ongoing weekly cost of not having it. For most senior professionals, the return on that investment arrives within the first week.

Flashcards Study Aid


Continue to Lesson 2: Two Plugins, One System →