Part 0: Thinking is the Curriculum
10 Chapters. 10 Skills. Every Exercise Requires You to Think Before AI Thinks for You.
This part comes before agent foundations, before programming, before architecture, before deployment. It comes first because everything that follows depends on it.
AI tools can retrieve facts, generate analyses, write code, and produce polished documents. What they cannot do is think. They cannot ask the right question. They cannot detect their own errors. They cannot reason from first principles when no pattern exists. They cannot decide under uncertainty, navigate ethical dilemmas, or create something genuinely new.
These ten chapters teach the skills that AI amplifies but cannot replace. But teaching thinking in the age of AI creates an immediate paradox: if students have access to AI tools, what stops them from letting AI do all the thinking? The answer is not to ban AI. The answer is to design exercises where AI cannot do the work — because the work is the student's own cognitive process, not the output.
The deliverable is never the answer. The deliverable is the documented evidence of thinking.
Part 0 is about how to teach humans to thrive in the future. The rest of the book is how to build the future.
What You Need
- A web browser
- Access to claude.ai and chatgpt.com (free tiers work)
- No code. No setup. No IDE. Just your brain and a browser.
The Six Layers of Thinking-Proof Assessment
To make the foundational principle operational, every chapter in Part 0 is built on six layers. These layers are embedded into the exercises so that it becomes structurally impossible to succeed by outsourcing thinking to AI.
Before the student touches any AI tool, they must commit to a position in writing. This is timestamped and cannot be changed. Only after this prediction is sealed do they open claude.ai or chatgpt.com. AI cannot do this part because it requires the student to have had an independent thought first that they are now comparing against.
Applied in: Every chapter. The prediction lock is the universal starting point.
Students do not submit answers. They submit a reasoning trail — every prompt they wrote, every AI response they received, every decision they made to accept, reject, or modify. The grade is on the decisions. AI can generate answers but it cannot generate a genuine record of someone else's decision-making process.
Applied in: Chapters 1, 2, 5, 6, 8, 9. The reasoning receipt is the primary grading artifact.
The student submits their work, then defends it without AI access. If the student outsourced all thinking to AI, they collapse under the first question. This is the oldest assessment method in academia — the oral examination — and it is suddenly the most AI-proof one.
Applied in: Chapters 1, 3, 5, 7, 10. Every chapter with live performance uses this layer.
After a student submits their work, it is fed into AI with the prompt "argue against this." The student must respond in real-time. The student who copied from AI does not understand what they submitted well enough to defend it against AI's own counter-attack.
Applied in: Chapters 1, 2, 4, 7, 9. Submitting work is never the end — defending it is the real assessment.
When 30 students receive the same problem and AI tools, those who outsourced thinking produce nearly identical outputs. Those who actually thought will diverge. Originality becomes a measurable signal of cognitive engagement.
Applied in: Chapters 1, 4, 6, 8. Divergence is measured whenever the class receives the same prompt.
Students submit three drafts: before AI, after AI collaboration, and after reflection. The grade lives in the gaps between drafts. AI can produce a polished final draft but it cannot produce a genuine record of thinking evolving across three stages.
Applied in: Chapters 2, 3, 4, 6, 7, 8, 9, 10. The three-draft structure makes cognitive growth visible.
How the Layers Work Together
No single layer is sufficient on its own. But surviving all six simultaneously requires genuine thinking. The prediction lock creates a baseline the reasoning receipt must be consistent with. The reasoning receipt documents decisions the live defence will interrogate. The contradiction challenge attacks the position committed to in the prediction lock. The divergence test catches students who bypassed earlier layers. And iterative drafts make the entire trajectory visible.
How AI Checks Your Thinking in Every Exercise
Every exercise in this part follows a four-step cycle:
- You think first and produce a deliverable without AI or with documented AI collaboration.
- You submit your deliverable to AI using an exact prompt provided in the exercise.
- AI grades your work, identifies your blind spots, and gives you specific feedback.
- You reflect on the gap between your self-assessment and AI's assessment — that gap is where the deepest learning happens.
This is not AI doing your thinking for you. This is AI acting as a rigorous, tireless, infinitely patient evaluator of your thinking. A human instructor cannot read 30 students' reasoning receipts in real-time and give each one detailed feedback. AI can. The instructor's role shifts from grading to designing exercises and conducting live defences — the parts that require human judgment.
When AI Feedback Seems Wrong
Part 0 uses AI to evaluate your thinking, while Chapter 2 teaches you that AI is frequently wrong. Both are true. AI is a powerful evaluator but not a perfect one. Sometimes its feedback will be inaccurate, unfair, or shallow. When that happens, do not ignore it and do not blindly accept it. Use the following protocol:
- Identify the specific feedback point you disagree with.
- Write your counter-argument explaining why the AI's evaluation is incorrect (100-150 words).
- Submit your counter-argument back to AI with this prompt: "I disagree with your evaluation on [specific point]. Here is my counter-argument: [paste]. Either defend your original evaluation with specific reasoning or acknowledge the error."
- Include the full exchange (AI feedback, your challenge, AI response) in your portfolio.
- This is graded as bonus evidence of critical thinking.
Challenging AI feedback is not a sign of failure. It is the highest application of the skills this part teaches. A student who accepts bad feedback uncritically has missed the entire point.
The Thinking Score Card
Every AI check prompt in this part ends with the same standardized scoring request: the Thinking Score Card. This gives you five consistent scores (each 1-10) across all 40 exercises, all 10 chapters, allowing you to track your growth on a single chart from Chapter 1, Exercise 1 to Chapter 10, Exercise 4.
- Independent Thinking (1-10): Did the student produce genuine thought before or beyond AI? Evidence of predictions, original analysis, or ideas that AI would not generate independently.
- Critical Evaluation (1-10): Did the student evaluate AI output critically? Evidence of accepting, rejecting, or modifying AI responses with justified reasoning rather than passive acceptance.
- Reasoning Depth (1-10): How deep is the student's reasoning? Evidence of second-order thinking, tracing causes to consequences, identifying hidden assumptions, and building logical chains rather than surface-level responses.
- Originality (1-10): How much of the student's work diverges from what AI would produce if given the same prompt? Evidence of unique perspectives, novel connections, or creative approaches that go beyond standard AI output.
- Self-Awareness (1-10): Does the student accurately understand their own strengths, weaknesses, and confidence levels? Evidence of calibrated confidence, honest gap identification, and reflections that match the quality of the work.
Score Tracking Table
Maintain this table throughout Part 0. After each exercise, record your five scores:
| Exercise | Independent Thinking | Critical Evaluation | Reasoning Depth | Originality | Self-Awareness | Average |
|---|---|---|---|---|---|---|
| Ch1.Ex1 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 |
| Ch1.Ex2 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 |
| Ch1.Ex3 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 |
| Ch1.Ex4 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 |
| Ch2.Ex1 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 |
| Ch2.Ex2 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 |
| Ch2.Ex3 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 |
| Ch2.Ex4 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 |
| Ch3.Ex1 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 |
| Ch3.Ex2 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 |
| Ch3.Ex3 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 |
| Ch3.Ex4 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 |
| Ch4.Ex1 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 |
| Ch4.Ex2 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 |
| Ch4.Ex3 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 |
| Ch4.Ex4 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 |
| Ch5.Ex1 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 |
| Ch5.Ex2 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 |
| Ch5.Ex3 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 |
| Ch5.Ex4 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 |
| Ch6.Ex1 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 |
| Ch6.Ex2 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 |
| Ch6.Ex3 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 |
| Ch6.Ex4 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 |
| Ch7.Ex1 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 |
| Ch7.Ex2 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 |
| Ch7.Ex3 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 |
| Ch7.Ex4 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 |
| Ch8.Ex1 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 |
| Ch8.Ex2 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 |
| Ch8.Ex3 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 |
| Ch8.Ex4 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 |
| Ch9.Ex1 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 |
| Ch9.Ex2 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 |
| Ch9.Ex3 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 |
| Ch9.Ex4 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 |
| Ch10.Ex1 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 |
| Ch10.Ex2 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 |
| Ch10.Ex3 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 |
| Ch10.Ex4 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 |
| Average | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 | ___/10 |
At the end, calculate your average per dimension across all 40 exercises and compare with your Thinking Baseline scores.
The Ten Chapters
| Chapter | Core Skill | What You Build |
|---|---|---|
| 1. Asking Better Questions | Question Formulation | A Question Quality Portfolio with prediction locks, reasoning receipts, and AI-graded evaluations |
| 2. Detecting Broken Reasoning | Verification and Discernment | An Error Detection Portfolio with annotated AI outputs, confidence calibration, and error taxonomy |
| 3. Thinking in Systems | Systems Thinking | A Systems Thinking Portfolio with cascade maps, human-AI comparisons, and variable shift analyses |
| 4. Reasoning From First Principles | First Principles Reasoning | A First Principles Portfolio with derivations, assumption autopsies, and constraint rebuilds |
| 5. Communicating What Matters | Communication Under Pressure | A Communication Portfolio with audience predictions, live adaptations, and hard conversations |
| 6. Working With AI, Not For AI | AI Collaboration | An AI Collaboration Portfolio with three-path comparisons, collaboration logs, and override tests |
| 7. Reasoning Through Dilemmas | Ethical Reasoning | An Ethical Reasoning Portfolio with position locks, adversarial defences, and stakeholder swaps |
| 8. Building Something From Nothing | Creation and Originality | A Creation Portfolio with blank page sprints, creation logs, and three-draft evolutions |
| 9. Deciding Under Uncertainty | Decision-Making | A Decision-Making Portfolio with sealed decisions, reversal triggers, and decision audits |
| 10. Learning How to Learn | Meta-Learning | A Meta-Learning Portfolio with learning plans, 72-hour sprints, and a Personal Learning Framework |
For Solo and Online Learners
Several exercises involve peer interaction — question tournaments, cross-examination, live defence, hard conversations, and teach-back sessions. If you are learning alone or asynchronously online, every peer exercise includes a Solo Learner Alternative. These alternatives use AI to simulate the peer role:
Instead of a peer panel questioning your work, you submit your deliverable to AI with a specific adversarial prompt that generates tough, unpredictable questions — then you respond in writing. Instead of a live role-play partner, AI plays the stakeholder and you practice adapting in real-time through a multi-turn conversation. Instead of peer feedback forms, you use a structured AI evaluation prompt designed to replicate what an engaged peer would notice.
The solo path is not inferior to the peer path — it is different. Peers provide unpredictability, social pressure, and perspectives you cannot anticipate. AI provides consistency, tirelessness, and the ability to generate adversarial challenges on demand. If possible, combine both.
How This Scales: Assessment for 16,000 Students (Instructor Reference)
Tier 1 — AI First-Pass (every student, every exercise): Every exercise includes an exact AI check prompt. The student submits their work, receives AI-generated scores and feedback, and includes this in their portfolio. This provides immediate, personalized feedback at unlimited scale. AI scores serve as the baseline assessment.
Tier 2 — Peer Review Circles (every student, per chapter): Students are organized into review circles of 4-5 people. At the end of each chapter, circles exchange portfolios and evaluate one peer's work using a provided rubric. Peer reviewers submit their evaluation along with their own portfolio. Reviewing others' thinking is itself a thinking exercise — it reinforces the skills being taught.
Tier 3 — Instructor Spot-Check (flagged portfolios): Instructors do not review every portfolio. They review flagged cases: portfolios where AI scores and peer scores diverge significantly, portfolios with suspiciously high divergence-test similarity, students who challenged AI feedback (to verify the challenge was legitimate), and a random 10% sample for calibration. This keeps instructor load manageable at any scale while ensuring quality control.
This three-tier system means every student gets personalized AI feedback within minutes, peer feedback within days, and instructor attention where it matters most.
Before You Begin
Complete the Thinking Baseline — a 30-minute ungraded assessment that snapshots your current thinking skills. You will repeat it after Chapter 10 to measure your growth.
Learning Path
Thinking Baseline → Chapters 1-10 → Portfolio Assembly → Post-Assessment → Growth Map
Each chapter builds on the previous. Skills introduced early (Prediction Lock, Reasoning Receipt, Error Taxonomy) recur throughout. By Chapter 10, you are using every skill from every preceding chapter simultaneously.
The litmus test for every skill in this part: Does this make AI a more powerful tool in your hands, or does it make you a slower version of the tool?