Course 03 Implementation Spec: The "No Playbook" Case Studies
1. Title and Source Files Used
Course title: The "No Playbook" Case Studies
Owned output file: 03-no-playbook-case-studies/03_no_playbook_case_studies_course_content.md
Source files used as base truth:
03-no-playbook-case-studies/curriculum.md03-no-playbook-case-studies/website-prompt.md
Purpose of this file: Translate the course curriculum into a delivery-ready course-content specification that a facilitator, LMS builder, evaluator, or AI-assisted teaching workflow can execute without inventing missing instructional structure.
Source-of-truth commitments for this spec:
- Preserve the curriculum's 3-week, 6-session, 90-minute live format.
- Preserve the central thesis: the advantage is not following a playbook but operating well when none exists.
- Preserve the named core cases from the curriculum:
- XPRIZE origin
- SpaceX early years and Falcon 1 period
- OpenAI formation and restructuring
- Cross-case synthesis including Airbnb
- Preserve the curriculum's four assessed outputs:
- case study analysis
- cross-case matrix plus personal map
- ambiguity simulation participation
- final reflection
- Preserve the website prompt's tone and visual direction as implementation guidance for slides, handouts, and LMS framing, without changing the curriculum itself.
2. Design Decisions at the Top
Instructional design decisions
- This course teaches decision quality under ambiguity, not founder trivia.
Students should leave with a reusable way of thinking about unclear, high-stakes situations, not a scrapbook of startup stories.
- Case studies are structured around decision points.
Every major case is taught through moments when the actors had to choose with incomplete information, contradictory signals, and visible downside.
- Reasoning matters more than outcome worship.
Students should learn to analyze whether a decision was rigorous given the information available then, not whether the eventual outcome became famous.
- The emotional load of ambiguity is part of the curriculum.
Students need language for fear, doubt, social pressure, and team morale, because no-playbook situations are not purely analytical.
- The course escalates from recognition to reenactment.
The sequence intentionally moves from spotting ambiguity, to dissecting cases, to extracting a framework, to applying it to the student's own life, to live simulation.
- Students must repeatedly separate knowns, unknowns, and unknowables.
This distinction is a core operating habit for ambiguous environments and should appear in discussion, worksheets, and grading.
- Outputs are designed to be LLM-evaluable.
Every major deliverable uses explicit sections and evidence expectations so an AI evaluator can assess depth of reasoning instead of rewarding style alone.
- The website prompt informs presentation style.
Slides, worksheets, and LMS modules should echo the documentary framing:
- episode-like case structure
- high-contrast decision moments
- recurring language such as "The Four Constants," "decision reenactment," and "when the map didn't exist"
- Supplemental mini-cases support comparison but do not replace the core cases.
Airbnb is the required cross-case comparator from the curriculum. Stripe and Square may appear as short examples only if used to sharpen pattern recognition.
Facilitation design decisions
- Facilitators should act like disciplined case moderators.
The job is to keep students close to evidence, timing, incentives, constraints, and decision logic.
- Hero worship and cynicism both need correction.
If students romanticize founders, ask about costs and tradeoffs. If students dismiss them, ask what alternative they would have chosen with the same information.
- Discussion should privilege specific claims.
The facilitator should frequently ask:
- What did they actually know?
- What did they think they knew?
- What was still unknowable?
- What options were live at that moment?
- Personal transfer is required, not optional.
The course only lands if students can identify a real no-playbook situation in school, work, family, community, or creative life.
3. Delivery Model Assumptions
Cohort and format assumptions
- Duration: 3 weeks
- Live sessions: 6 sessions
- Session length: 90 minutes each
- Recommended cohort size: 12 to 28 students
- Recommended staffing: 1 lead facilitator; add 1 teaching assistant for cohorts above 20
- Delivery mode: in-person, live online, or hybrid
- Default assumption for this spec: live synchronous discussion/workshop plus short asynchronous assignments
- Student age range: 15 to 25
- Prior background: mixed; no prerequisite technical or startup knowledge required
Instructional environment assumptions
- Students have access to a slide deck, shared docs or LMS, and either breakout rooms or small-group tables.
- Facilitator can project or distribute short case packets, quotes, timelines, and decision-point handouts.
- Students can submit written work digitally.
- Students may use AI tools for brainstorming or first-pass outlining, but all final submissions must show the student's own judgment, synthesis, and evidence selection.
Time and workload assumptions
- Live time: 9 total hours
- Async workload: approximately 5 to 7 hours over 3 weeks
- Reading load should stay light and targeted:
- brief excerpts
- short interview clips or transcripts
- decision-point summaries
- The course should feel demanding in reasoning, not burdensome in reading volume.
Materials assumptions
Required materials:
- facilitator slide deck
- case packets for XPRIZE, SpaceX, OpenAI, and Airbnb mini-case comparison
- decision memo worksheet
- Four Constants worksheet
- cross-case matrix template
- personal no-playbook map template
- scenario room packet for Session 6
- grading rubrics and LLM evaluator prompts
Recommended instructional routines:
- quickwrite
- pair share
- small-group decision lab
- cold-call evidence defense
- structured debrief
- exit ticket
Accessibility and inclusion assumptions
- Examples should be translated into terms students from non-startup backgrounds can engage with.
- "No-playbook" situations may come from school leadership, family responsibilities, community organizing, freelancing, sports, religious groups, or creative projects.
- Students may anonymize personal situations in written work.
Assessment assumptions
- The LLM acts as a first-pass grader and feedback assistant.
- A human reviews:
- borderline submissions
- emotionally sensitive reflections
- suspected fabricated case evidence
- unusually high or low scores
4. Detailed Course Content
Course-Level Throughline
Students repeatedly practice one operating sequence:
- Detect that a situation is genuinely no-playbook.
- Name the underlying problem instead of hiding inside inherited language.
- Separate what is known, unknown, and unknowable.
- Clarify stakes, constraints, and downside tolerance.
- Generate options without pretending certainty exists.
- Choose a move and defend why it is appropriate now.
- Treat failure, surprise, and feedback as information for the next move.
Course-Level Recurring Framework: The Four Constants
The curriculum names four constants in no-playbook situations. They should recur across sessions, worksheets, and grading:
- The Physics Test
What is actually true about constraints, timing, cost, capabilities, and reality?
- The Stakes Clarity
What are we actually willing to lose, spend, delay, or risk?
- The Information Asymmetry Map
What do we know, need to know, and accept that we cannot know in time?
- The First-Mover Calculus
Is the advantage in being correct, or in moving before certainty exists?
Module 1 / Session 1: Recognizing No-Playbook Moments
Session goal: Students can distinguish routine, complicated, complex, and chaotic situations and identify what makes a situation "no-playbook."
Session outcomes:
- Students can define a no-playbook situation in their own words.
- Students can classify ambiguous scenarios with justification.
- Students can analyze the XPRIZE origin as a case of creating a structure where no template existed.
Session artifacts:
- scenario classification sheet
- XPRIZE decision note
- exit ticket definition
Session 1 run of show
| Time | Segment | Lesson / Activity | Facilitator moves | Student outputs |
|---|---|---|---|---|
| 0-10 min | Opening provocation | Quickwrite: "Tell me about a time no one could tell you what to do." | Normalize school, family, team, and work examples. Push students away from generic "hard things." | 4-6 sentence quickwrite |
| 10-25 min | Lesson 1.1 | What counts as a no-playbook moment? | Define routine, complicated, complex, chaotic. Stress that no-playbook usually lives in complex or chaotic conditions but can appear inside any domain when precedent breaks. | Notes with draft definition |
| 25-40 min | Activity | Spectrum Test scenario sort | Give 10 scenarios. Require students to classify and defend each one with one sentence of reasoning. Ask what decision style fits each category. | Completed classification sheet |
| 40-50 min | Debrief | Why classification matters | Ask: "What goes wrong if you treat complex as routine?" and "What goes wrong if you dramatize a routine problem into chaos?" | Annotated corrections |
| 50-65 min | Lesson 1.2 | Case study: XPRIZE origin | Teach the 1995 context, lack of private-space precedent, and why prize incentives were structurally unusual. Keep the case on decision architecture, not biography. | XPRIZE case notes |
| 65-80 min | Activity | Decision dissection: Build the structure | Prompt: "If you wanted private space innovation but had no industry, no proven capital model, and no dominant customer, what structure would you invent?" | 1-page structure sketch in pairs |
| 80-87 min | Whole-group discussion | Reveal and compare | Ask which assumptions each pair kept from existing institutions and which they discarded. | Marked-up pair draft |
| 87-90 min | Exit ticket | Session close | Ask students to define a no-playbook moment in one sentence and name one signal that indicates they are in one. | Exit ticket |
Session 1 facilitator guidance
- Do not let students call everything "chaotic." Precision matters.
- When students confuse difficulty with ambiguity, ask whether a competent expert already knows the path.
- In the XPRIZE discussion, redirect from "genius idea" language toward incentive design, legitimacy, and timing.
Session 1 assignment launch
Homework: Ambiguity observation log
- Students collect 3 examples from real life or media where a person or organization faced unclear precedent.
- For each example, students record:
- What made it no-playbook?
- What was known?
- What was unknown?
- What was the main perceived risk?
Module 2 / Session 2: SpaceX Deep Dive I - Early Formation and Failure
Session goal: Students analyze how SpaceX operated when the private-space playbook did not yet exist and evaluate early decisions under uncertainty.
Session outcomes:
- Students can identify the specific assumptions SpaceX broke.
- Students can analyze early SpaceX decisions using knowns, unknowns, and alternatives.
- Students can produce a defensible judgment about whether persistence after failure was rational.
Session artifacts:
- SpaceX timeline sheet
- failure analysis worksheet
- decision memo draft
Session 2 run of show
| Time | Segment | Lesson / Activity | Facilitator moves | Student outputs |
|---|---|---|---|---|
| 0-10 min | Reactivation | Review ambiguity logs | Ask for one example that clearly was and one that clearly was not a no-playbook moment. | Shared examples |
| 10-22 min | Context briefing | SpaceX 2001-2002 formation | Teach the initial goal, the cost structure of launch, the Russia trip context, and why buying vs building mattered. | Timeline notes |
| 22-35 min | Lesson 2.1 | What assumptions did SpaceX reject? | Surface assumptions around aerospace-grade procurement, outsourcing, talent pedigree, and customer legitimacy. | Assumption list |
| 35-50 min | Activity | Constraint sort: physics vs convention in aerospace | Give claims such as "rockets must be built by legacy contractors" and "launch reliability demands extreme process discipline." Make students sort them. | Sorted claims sheet |
| 50-65 min | Lesson 2.2 | Failure as a decision environment | Introduce Falcon 1 failure context. Ask what failure changes: information, morale, capital, credibility, and risk tolerance. | Failure environment notes |
| 65-82 min | Activity | Decision in the room after Failure #1 | Students receive a short packet: remaining runway, technical uncertainty, team state, external skepticism. In groups, they choose one next move and justify it. | Group decision memo |
| 82-90 min | Debrief | Compare group choices | Ask which groups optimized for survival, learning, reputation, or mission. Name tradeoffs explicitly. | Debrief notes |
Session 2 facilitator guidance
- Keep the case anchored to timing. Students must answer as if they are in 2002-2006, not with hindsight.
- If students say "they should just raise more money," ask whether credibility, timing, and investor appetite made that realistic then.
- Push the distinction between mission commitment and irrational stubbornness.
Session 2 homework
Homework: SpaceX decision memo
- 500-700 words
- Prompt: "After one failed launch, what should SpaceX do next, and why?"
- Required structure:
- situation summary
- what is known
- what is unknown
- options considered
- recommended move
- biggest downside of your recommendation
Module 2 / Session 3: SpaceX Deep Dive II - Persistence, Reuse, and Framework Extraction
Session goal: Students analyze later SpaceX decision points and extract general principles for operating under uncertainty.
Session outcomes:
- Students can evaluate persistence after repeated failure.
- Students can analyze vertical integration and reuse as no-playbook strategic choices.
- Students can extract transferable decision principles from the SpaceX case.
Session artifacts:
- post-failure option matrix
- principle extraction sheet
- case analysis outline
Session 3 run of show
| Time | Segment | Lesson / Activity | Facilitator moves | Student outputs |
|---|---|---|---|---|
| 0-12 min | Opening | Memo share and cold-call defense | Ask several students to defend their Session 2 recommendation against critique. | Revised memo notes |
| 12-28 min | Lesson 3.1 | Failure #2 and Failure #3 as escalating ambiguity | Teach how money, morale, and credibility compress over time. Ask what changes between first, second, and third failure. | Escalation notes |
| 28-45 min | Activity | Pivot, persist, or quit matrix | Students map three options across cost, learning, morale, mission, and probability of survival. | Option matrix |
| 45-58 min | Lesson 3.2 | Vertical integration and off-pattern execution | Frame in-house manufacturing, unconventional talent, and first-principles cost reasoning as deliberate departures from industry norms. | Strategy notes |
| 58-70 min | Lesson 3.3 | Reuse before reuse was proven | Ask why building for reuse before proof looked reckless and why it might still have been rational. | Short written response |
| 70-82 min | Activity | Framework extraction lab | Small groups answer: "What decision rules seem to govern this case?" Force them to express rules, not admiration. | 5-rule group list |
| 82-90 min | Debrief | From case to transferable method | Consolidate class list into a visible set of candidate principles. | Class principle sheet |
Required principle categories for Session 3
Each group must state one principle in each category:
- how to treat failure
- how to treat expertise
- how to think about cost
- how to decide when persistence becomes irrational
- how to protect morale without lying
Session 3 facilitator guidance
- Reject vague principles such as "never give up" or "believe in yourself."
- Ask, "What evidence from the case supports that rule?"
- If students ignore downside, ask what would make the same behavior look irresponsible.
Session 3 assignment launch
Major assignment launched: Case Study Analysis
- Students choose SpaceX or OpenAI.
- Due after Session 4.
- See Section 5 for full requirements.
Module 3 / Session 4: OpenAI's Formation and Structural Ambiguity
Session goal: Students examine a different type of no-playbook problem: not rockets, but organization design under emerging technology, mission tension, and capital constraints.
Session outcomes:
- Students can explain why OpenAI's original structure was unusual relative to academic labs and corporate AI labs.
- Students can analyze the tension between safety, openness, talent competition, and capital needs.
- Students can compare structural ambiguity to technical ambiguity.
Session artifacts:
- OpenAI structure map
- tradeoff analysis chart
- case analysis planning notes
Session 4 run of show
| Time | Segment | Lesson / Activity | Facilitator moves | Student outputs |
|---|---|---|---|---|
| 0-10 min | Re-entry | Bridge from SpaceX to OpenAI | Ask: "What remains the same when the frontier is organizational, not hardware?" | Bridge notes |
| 10-25 min | Lesson 4.1 | Context: 2015 AI landscape | Frame academic AI, big-tech labs, AGI concern, talent concentration, and the problem of mission control. | Context timeline |
| 25-40 min | Lesson 4.2 | The original structure problem | Walk through nonprofit origins, safety mission, open publication impulse, and why a standard lab model did not fit. | Structure map |
| 40-55 min | Activity | Build the structure exercise | In teams, students design an organization that must pursue frontier AI, retain safety commitments, attract talent, and raise enough capital. | Team structure proposal |
| 55-68 min | Debrief | Compare student structures to what happened | Ask what tradeoffs each structure optimizes and what pathologies it invites. | Tradeoff notes |
| 68-80 min | Lesson 4.3 | 2019 capped-profit restructuring | Present this as a live response to competing constraints, not as a moral verdict. Ask what changed between the original ideal and practical requirements. | Restructuring notes |
| 80-88 min | Activity | Tension map | Students list 4 tensions: openness vs competitive necessity, safety vs speed, mission vs capital, governance vs agility. | Tension map |
| 88-90 min | Close | Assignment reminder | Re-state case analysis due date and required evidence standard. | Submission plan |
Session 4 facilitator guidance
- Keep the class away from tribal debates about whether a given organization is "good" or "bad."
- The teaching question is: what structure would you design when your aims conflict and no stable institutional template exists?
- If students produce simplistic answers like "just stay nonprofit," ask how talent, compute, and capital would work in practice.
Optional supplemental mini-case notes
If time allows, the facilitator may introduce 5-minute comparison snapshots of Airbnb, Stripe, or Square to illustrate different forms of no-playbook uncertainty:
- Airbnb: market legitimacy and trust creation
- Stripe: infrastructure simplification in a painful domain
- Square: productizing behavior change for small merchants
These are support examples only and should not displace the core OpenAI session content.
Module 4 / Session 5: Cross-Case Synthesis and Personal Mapping
Session goal: Students synthesize patterns across cases and apply the Four Constants to a personal no-playbook situation.
Session outcomes:
- Students can compare technical, organizational, and market ambiguity across cases.
- Students can use the Four Constants as a practical framework.
- Students can locate one current or upcoming no-playbook situation in their own life.
Session artifacts:
- cross-case matrix
- personal no-playbook map
- draft final reflection notes
Session 5 run of show
| Time | Segment | Lesson / Activity | Facilitator moves | Student outputs |
|---|---|---|---|---|
| 0-10 min | Opening synthesis | What do these cases actually share? | Ask for common patterns and force evidence-backed answers. | Class synthesis notes |
| 10-25 min | Lesson 5.1 | The Four Constants framework | Teach the Physics Test, Stakes Clarity, Information Asymmetry Map, and First-Mover Calculus explicitly as a reusable decision tool. | Framework notes |
| 25-45 min | Activity | Cross-case matrix build | Teams place SpaceX, OpenAI, Airbnb, and one self-selected case on axes of technical uncertainty and market/social uncertainty. | Cross-case matrix draft |
| 45-58 min | Debrief | Compare placements | Ask why cases land where they do, and what changes in strategy at each quadrant. | Revised matrix |
| 58-72 min | Lesson 5.2 | Personal application workshop | Students identify a real ambiguity they are in or approaching. Examples: choosing a project path, launching something, stepping into responsibility, navigating new tech, or building in a space with weak precedent. | Personal case selection |
| 72-84 min | Activity | Personal no-playbook map | Students complete a structured map: situation, stakes, knowns, unknowns, unknowables, first move, anti-delusion check. | Personal map |
| 84-90 min | Close | Preview Scenario Room | Explain that Session 6 will test reasoning quality under timed uncertainty. | Preparedness note |
Required structure for the personal no-playbook map
Each student's map must include:
- one-sentence situation statement
- why this is genuinely no-playbook
- 3 knowns
- 3 unknowns
- 2 unknowables
- main downside if they move too early
- main downside if they wait too long
- one next action
- one evidence signal that would cause them to update the plan
Session 5 facilitator guidance
- Some students will choose situations that are merely stressful, not actually ambiguous. Ask whether a clear playbook exists and they are simply avoiding it.
- Some students will choose cases too large and abstract. Push them toward a decision they may realistically face in the next 3 to 12 months.
- Reward specificity over drama.
Session 5 assignment launch
Assignment: Cross-Case Matrix + Personal Map
- Due before Session 6
- See Section 5 for details
Module 5 / Session 6: Operating in Ambiguity - The Scenario Room
Session goal: Students demonstrate live reasoning in a simulated no-playbook environment and reflect on how they behave under ambiguity.
Session outcomes:
- Students can apply the Four Constants under time pressure.
- Students can explain and defend a recommendation without pretending certainty.
- Students can reflect on their instincts under ambiguity and identify what to improve.
Session artifacts:
- scenario analysis worksheet
- group reasoning presentation
- final reflection launch notes
Session 6 run of show
| Time | Segment | Lesson / Activity | Facilitator moves | Student outputs |
|---|---|---|---|---|
| 0-8 min | Briefing | Scenario Room rules | Emphasize: grade reasoning, not "right answer." Teams must show assumptions, tradeoffs, and update logic. | Scenario packet received |
| 8-18 min | Scenario assignment | Groups receive A, B, or C | Assign mixed-strength groups. Clarify packet facts only; do not answer strategic questions. | Scenario read-through |
| 18-45 min | Team work | Group analysis using Four Constants | Circulate and ask: "What are you smuggling in as an assumption?" "What would you need to believe for this to work?" | Completed worksheet |
| 45-65 min | Presentations | Group reasoning presentations | Require teams to present recommendation, rejected alternative, biggest uncertainty, and first update trigger. | 4-minute team presentation |
| 65-78 min | Cross-group discussion | Compare logic across scenarios | Ask what patterns repeated even across different industries and settings. | Peer notes |
| 78-88 min | Debrief | Psychological dimension of ambiguity | Ask what felt hardest: lack of certainty, conflict, fear of being wrong, time pressure, social disagreement, or unclear ownership. | Debrief response |
| 88-90 min | Close | Launch final reflection | Re-state final reflection prompt and remind students to use a real situation, not a generic ambition story. | Final assignment brief |
Scenario Room required deliverable structure
Each group must submit:
- Problem framing
- Four Constants analysis
- Three options considered
- Recommended move now
- Biggest unresolved uncertainty
- Trigger that would cause a course correction
- One sentence on what failure would teach them next
Session 6 facilitator guidance
- Teams often jump to solutions too fast. Slow them down by asking what reality they are assuming.
- Do not reward polished confidence that skips evidence.
- If teams fixate on only one scenario detail, ask what second-order consequences they are missing.
- The debrief should include feeling-state analysis. Students need to notice whether they freeze, bluff, over-control, or over-index on consensus.
5. Assignments and Artifacts
A. Case Study Analysis
Weight: 30%
Student choice: SpaceX or OpenAI
Deliverable length: 1,400 to 1,800 words
Purpose: Analyze a real no-playbook decision environment, reconstruct the options available at the time, and judge the quality of the decision logic.
Required sections:
- case context and timeline
- the no-playbook moment
- knowns, unknowns, and unknowables
- options available at the time
- decision taken and why it may have seemed rational
- strongest alternative path
- what this case teaches about operating without a playbook
Minimum evidence requirements:
- at least 4 concrete case details
- at least 2 clearly stated tradeoffs
- at least 1 paragraph that resists hindsight bias
- explicit use of at least 2 of the Four Constants
Common failure modes:
- summary without analysis
- hindsight certainty
- founder worship or moral grandstanding
- vague claims like "they believed in the mission"
B. Cross-Case Matrix + Personal No-Playbook Map
Weight: 20%
Deliverable format: one visual matrix plus a 350-500 word personal explanation
Purpose: Compare ambiguity types across cases and transfer the framework to the student's own situation.
Required components:
- matrix with SpaceX, OpenAI, Airbnb, and one student-selected comparator
- labels for both axes
- 2-3 sentence justification for each placement
- personal no-playbook map using the required structure from Session 5
- short explanation of what type of ambiguity the student's own situation most resembles
Minimum evidence requirements:
- at least one comparison sentence across cases
- at least one sentence explaining why the student's situation is not merely "difficult" but actually no-playbook
C. Scenario Room Participation and Group Reasoning Artifact
Weight: 20%
Deliverable format: group worksheet plus live participation score
Purpose: Evaluate reasoning quality under time pressure and ambiguous conditions.
Scored elements:
- evidence of structured analysis
- contribution to team reasoning
- quality of recommendation
- quality of uncertainty acknowledgment
- quality of verbal defense during presentation or discussion
Required group artifact:
- the 7-part Scenario Room structure listed in Session 6
Participation note: This component scores the quality of the student's reasoning moves, not whether the team picked the same option the facilitator would have chosen.
D. Final Reflection: "My No-Playbook Moment"
Weight: 30%
Deliverable length: 1,000 to 1,200 words
Purpose: Apply the course to a real ambiguity in the student's life and articulate a practical framework for action.
Required sections:
- the situation
- why no ready-made playbook exists
- stakes and downside
- what is known, unknown, and unknowable
- how the Four Constants apply
- what the student is currently tempted to do
- what the student now believes is the best next move
- what would cause them to update their view
Minimum evidence requirements:
- one clearly stated tension or fear
- one rejected option and why it was rejected
- one update trigger
- one specific next action within the next 14 days
Formative Artifacts Across the Course
These may be ungraded or low-stakes completion items, but they should be collected because they support final performance and can be used by LLM tutors:
- ambiguity observation log
- Session 2 SpaceX decision memo
- Session 3 principle extraction sheet
- Session 4 structure design proposal
- Session 5 personal map draft
- Session 6 scenario worksheet
6. AI/LLM Grading and Assessment Framework
Assessment model
The LLM should function as a disciplined first-pass evaluator. Its job is to:
- check completion
- score visible evidence against rubrics
- identify weak reasoning patterns
- generate revision guidance
The LLM is not the sole authority on final grades. Human review is required for emotional nuance, suspected fabrication, and edge cases.
Core grading principles
- Grade reasoning under uncertainty, not rhetorical confidence.
Confident tone should not raise scores unless the submission shows disciplined analysis.
- Reward accurate reconstruction of the decision environment.
High scores require evidence that the student understands what the actors knew then, not just what we know now.
- Penalize hindsight bias.
If the student judges a decision mainly because the eventual outcome was good or bad, scores should drop.
- Reward explicit uncertainty handling.
Strong work names unknowns, unknowables, tradeoffs, and update triggers.
- Reward transfer, not just case recall.
The best submissions move from case facts to reusable principles or personal application.
- Do not over-penalize writing polish.
Minor grammar issues matter only if meaning becomes unclear.
Artifact-by-artifact LLM evidence checks
Case Study Analysis
The LLM should check for:
- clear case timeline
- a real decision point rather than broad narrative summary
- differentiation between knowns, unknowns, and unknowables
- at least one alternative path evaluated seriously
- visible resistance to hindsight bias
- use of the Four Constants
Automatic downgrades:
- if the essay mostly retells history with little judgment
- if the student treats outcome as proof of decision quality
- if no downside or tradeoff is named
Cross-Case Matrix + Personal Map
The LLM should check for:
- all required cases present
- justification for each placement
- consistent axis logic
- evidence that the student's own situation is genuinely ambiguous
- one plausible next move and one update trigger
Automatic downgrades:
- if the matrix is visually present but unjustified
- if the student's personal example is just a routine task with discomfort attached
- if "unknowns" are really just missing effort rather than ambiguity
Scenario Room Participation and Group Artifact
The LLM should check for:
- structured use of the Four Constants
- clear option generation
- recommendation with uncertainty acknowledged
- course-correction trigger
- evidence of reasoning rather than bluffing
Automatic downgrades:
- if the group jumps from problem to answer without mapping constraints
- if the presentation claims certainty where the scenario does not allow it
- if no rejected option is discussed
Final Reflection
The LLM should check for:
- a real situation with consequences
- clear explanation of why no ready-made playbook exists
- honest articulation of tension, risk, or fear
- use of all Four Constants
- a next action and an update rule
Automatic downgrades:
- if the reflection reads like generic ambition advice
- if the student avoids real stakes
- if no rejected option appears
- if the "next move" is too vague to execute
Concrete assessment heuristics for LLM evaluation
High-signal indicators
The submission is likely strong if it includes several of these:
- the student names what would have had to be true for a decision to work
- the student distinguishes missing information from unknowable information
- the student identifies a non-obvious tradeoff
- the student names an update trigger or falsification condition
- the student explains why a plausible alternative was rejected
- the student uses precise situational language rather than motivational slogans
Low-signal indicators
The submission is likely weak if it includes several of these:
- "They just believed" or "they took a risk" with no mechanism
- "No one knew" when some things were in fact knowable
- "The lesson is never give up" without boundary logic
- generic praise or condemnation of a founder
- personal reflection with no specific next action
- repeated buzzwords such as innovation, disruption, mindset, vision, or courage without attached evidence
Heuristic caps for scoring
- If the student never separates knowns from unknowns, the work cannot score above competent on reasoning quality.
- If the student never names a real tradeoff, the work cannot score above competent on analysis quality.
- If the student evaluates a decision entirely through eventual outcome, the work cannot score above partial on decision judgment.
- If the personal reflection contains no concrete next step within 14 days, it cannot score above partial on actionability.
- If the Scenario Room artifact contains no update trigger, it cannot score above partial on uncertainty handling.
Automatic human-review flags
- suspected fabricated case details or invented quotations
- emotional distress, coercion, or high-risk personal situations in the final reflection
- clear evidence of plagiarism or AI-generated filler
- major contradictions within the submission
- too little content to verify completion
Recommended LLM output structure
For each assessed artifact, the LLM should return:
- completion check
- rubric scores by criterion
- 2 to 4 evidence-based observations
- 2 concrete revision moves
- overall level: Strong / Competent / Weak / Incomplete
- human-review flag: Yes / No, with reason if yes
Suggested scoring mechanics
- Score each rubric criterion on a 1 to 4 scale.
- Use artifact-specific weights from Section 7.
- Convert rubric totals into the assignment's weighted course percentage.
Score meaning
- 4 - Strong: Specific, evidence-backed, and structurally sound reasoning
- 3 - Competent: Mostly complete and credible, with some thin areas
- 2 - Partial: Some correct elements, but important reasoning gaps remain
- 1 - Weak: Missing, generic, confused, or unsupported
7. Rubrics, Scoring Criteria, and Evaluator Prompt Guidance
A. Case Study Analysis Rubric
| Criterion | Weight | 4 - Strong | 3 - Competent | 2 - Partial | 1 - Weak |
|---|---|---|---|---|---|
| Decision-point framing | 20% | Names a precise no-playbook moment and stakes clearly | Decision point mostly clear | Broad moment with blurry stakes | No clear decision point |
| Knowns / unknowns / unknowables | 20% | Clean distinction with strong examples | Distinction mostly present | Some categories blur | No meaningful distinction |
| Tradeoff analysis | 20% | Evaluates multiple real tradeoffs and alternatives | Some tradeoffs named | Tradeoffs thin or generic | Little tradeoff analysis |
| Hindsight resistance | 15% | Explicitly analyzes what was visible then vs later | Some attempt to avoid hindsight | Outcome bias still visible | Judges mostly by results |
| Framework use | 15% | Uses Four Constants accurately and meaningfully | Uses framework adequately | Framework mentioned but thin | Framework absent or misused |
| Communication clarity | 10% | Clear, organized, and easy to follow | Mostly clear | Uneven clarity | Hard to follow |
B. Cross-Case Matrix + Personal Map Rubric
| Criterion | Weight | 4 - Strong | 3 - Competent | 2 - Partial | 1 - Weak |
|---|---|---|---|---|---|
| Cross-case comparison quality | 25% | Placements are well-justified and comparative logic is sharp | Placements mostly justified | Placements present but weakly defended | Matrix is incomplete or arbitrary |
| Ambiguity-type reasoning | 20% | Distinguishes technical, market, and organizational ambiguity clearly | Mostly accurate distinctions | Some confusion across types | Major confusion or omission |
| Personal map quality | 25% | Personal situation is real, specific, and genuinely no-playbook | Situation mostly clear | Situation is thin or only partly ambiguous | Situation does not fit course concept |
| Update logic | 15% | Strong next move and clear update trigger | Action and update mostly present | One exists but is weak | No usable action/update rule |
| Communication clarity | 15% | Easy to interpret visually and in prose | Mostly clear | Some ambiguity in presentation | Hard to follow |
C. Scenario Room Participation and Group Artifact Rubric
| Criterion | Weight | 4 - Strong | 3 - Competent | 2 - Partial | 1 - Weak |
|---|---|---|---|---|---|
| Problem framing | 20% | Frames the problem cleanly without premature solution jump | Problem mostly clear | Framing partial or noisy | Poorly framed |
| Use of Four Constants | 25% | All four used concretely and appropriately | Most used meaningfully | Mentioned but not deeply used | Little or no framework use |
| Option quality | 20% | Multiple plausible options considered and compared | Options present with some comparison | Limited or repetitive options | One shallow option only |
| Uncertainty handling | 20% | Recommendation includes explicit unresolved uncertainty and update trigger | Some uncertainty acknowledged | Thin acknowledgment | Pretends certainty |
| Presentation / participation | 15% | Clear defense, evidence use, and productive engagement | Solid participation | Uneven contribution or clarity | Minimal or unhelpful contribution |
D. Final Reflection Rubric
| Criterion | Weight | 4 - Strong | 3 - Competent | 2 - Partial | 1 - Weak |
|---|---|---|---|---|---|
| Situation authenticity | 20% | Real, specific, and consequential situation | Situation mostly clear and relevant | Situation somewhat vague | Situation feels generic or low-stakes |
| No-playbook diagnosis | 20% | Clearly explains why precedent is insufficient | Explanation mostly convincing | Some ambiguity but not well-argued | Not convincingly no-playbook |
| Framework application | 25% | All Four Constants applied accurately and concretely | Most constants applied well | Partial or generic application | Framework largely absent |
| Judgment and self-awareness | 20% | Names fear, temptation, and rejected options honestly | Reasonable reflection present | Reflection shallow or defensive | Little self-awareness |
| Actionability | 15% | Clear next action and update trigger | Action mostly clear | Action vague | No usable next step |
Course score conversion guidance
- 90-100: Strong command of ambiguous decision-making
- 75-89: Credible and mostly complete reasoning
- 60-74: Partial mastery; needs stronger handling of uncertainty and tradeoffs
- Below 60: Insufficient evidence of the course method
Suggested evaluator prompt for the LLM
Use this prompt with the relevant submission pasted after it:
You are evaluating student work for Course 03, "The 'No Playbook' Case Studies."
Your job is to assess the student's reasoning under ambiguity based only on visible evidence in the submission.
Do not reward confidence, polish, or admiration for famous founders unless the student shows disciplined analysis.
Reward:
- clear decision-point framing
- distinction between knowns, unknowns, and unknowables
- explicit tradeoffs
- resistance to hindsight bias
- use of the Four Constants
- concrete next actions or update triggers where relevant
Penalize:
- generic storytelling
- founder worship or reflexive moral condemnation
- judging decisions only by outcomes
- vague claims like "they took a big risk" without mechanism
- motivational language that replaces analysis
Return your evaluation in this order:
1. Completion check against the required components
2. Criterion-by-criterion scores from 1 to 4
3. Short evidence-based justification for each score
4. Two strengths
5. Two revision moves
6. Overall level: Strong, Competent, Weak, or Incomplete
7. Human-review flag: Yes or No, with reason if Yes
Important constraints:
- Do not invent missing evidence.
- If the student does not separate knowns from unknowns, cap reasoning quality at 3.
- If the student evaluates a decision mainly by outcome, cap decision judgment at 2.
- If the personal artifact lacks a concrete next action, cap actionability at 2.
- Minor writing issues should not heavily affect the score unless they block meaning.
```
## Optional structured output schema for LLM evaluators
```json
{
"assignment_type": "",
"completion_check": {
"complete": true,
"missing_components": []
},
"scores": {
"criterion_1": 0,
"criterion_2": 0,
"criterion_3": 0,
"criterion_4": 0,
"criterion_5": 0
},
"weighted_total": 0,
"strengths": [
"",
""
],
"revision_moves": [
"",
""
],
"overall_level": "",
"human_review_flag": {
"flagged": false,
"reason": ""
}
}
```
## Additional evaluator heuristics
- If the student never names what would cause them to change course, treat the work as weak on adaptive reasoning.
- If the student treats uncertainty as an excuse for hand-waving, score down on rigor.
- If the student includes a rejected option with a clear reason, score up on judgment.
- If the student recognizes that some information cannot be obtained in time, score up on realism.
- If the student confuses disagreement with ambiguity, score down on conceptual precision.
---
## 8. Feedback Strategy: What Strong, Average, and Weak Responses Look Like and How an LLM Should Respond
## Global feedback posture
The LLM should be direct, sober, and developmental. It should not flatter students with vague praise. It should identify:
- what evidence is working
- what reasoning move is missing
- what revision would materially improve the submission
## Strong responses usually look like this
- They reconstruct the decision environment rather than retell the whole story.
- They distinguish knowns, unknowns, and unknowables with precision.
- They analyze at least one real alternative rather than pretending only one path existed.
- They acknowledge fear, downside, or institutional tension.
- They make a recommendation or personal next move without pretending certainty.
**How the LLM should respond to strong work:**
- name the exact reasoning move that made the work strong
- reinforce discipline, not just insight
- push for one layer deeper, usually second-order effects or update logic
**Example LLM response pattern:**
"Your strongest move was separating what the actors knew from what later observers know. That kept your analysis from collapsing into hindsight. Your tradeoff section is also strong because you treated capital, morale, and credibility as competing constraints rather than a single 'risk' bucket. To improve further, add one explicit signal that would have told the team their current path was no longer rational."
## Average responses usually look like this
- They include the right sections but some are thin.
- They summarize the case better than they analyze it.
- They identify uncertainty but do not classify it carefully.
- They have a personal application, but the next move is under-specified.
**How the LLM should respond to average work:**
- identify the strongest existing element
- name the specific missing layer of reasoning
- ask for one concrete revision that upgrades the work
**Example LLM response pattern:**
"You clearly identified that this was a no-playbook situation and you included the main case facts. The weakest part is the treatment of uncertainty: several items listed as 'unknown' were actually unknowable in the moment, and that distinction matters. Revise by separating those categories and adding one rejected option with a clear reason it was not chosen."
## Weak responses usually look like this
- They rely on slogans such as "take bold risks" or "trust the vision."
- They judge famous decisions mainly by whether they worked.
- They never identify a serious alternative.
- They confuse stress with ambiguity.
- Their personal reflection is abstract, sanitized, or non-actionable.
**How the LLM should respond to weak work:**
- state plainly that the work is too generic to score highly
- identify the exact missing evidence
- give a narrow rescue plan with 2 or 3 concrete fixes
**Example LLM response pattern:**
"This submission does not yet show strong reasoning under ambiguity. It names a famous case and a broad lesson, but it does not identify a specific decision point, real tradeoffs, or what was unknown at the time. Start by choosing one exact moment, listing three knowns and three unknowns, and explaining one alternative the actors could have taken."
## Artifact-specific feedback guidance
### Case Study Analysis
**Strong:** Specific decision point, disciplined uncertainty map, clear tradeoffs, low hindsight bias.
**Average:** Good summary and some analysis, but tradeoffs or alternatives are thin.
**Weak:** Mostly biography, opinion, or outcome-based judgment.
**LLM response move:** Ask, "What made this decision hard in real time?" and "What plausible alternative did you take seriously?"
### Cross-Case Matrix + Personal Map
**Strong:** Placements are justified, comparisons are meaningful, and the personal case is genuinely ambiguous.
**Average:** Matrix exists but some placements feel intuitive rather than argued.
**Weak:** Cases are dropped on the chart with little logic; personal case is not truly no-playbook.
**LLM response move:** Ask the student to defend one matrix placement and one personal unknown more precisely.
### Scenario Room Artifact
**Strong:** The team shows structure, options, uncertainty, and update logic.
**Average:** Reasonable recommendation, but the analysis skipped one or two framework steps.
**Weak:** Fast opinion masquerading as strategy.
**LLM response move:** Ask what they ruled out, what they still do not know, and what would force a pivot.
### Final Reflection
**Strong:** Honest, concrete, and actionable; the student shows tension and judgment.
**Average:** Relevant but somewhat abstract; the framework is present but not fully operational.
**Weak:** Generic self-help essay with little real decision logic.
**LLM response move:** Ask for one actual next move inside 14 days and one condition that would change that move.
## Final note for evaluators
The course succeeds when students become more disciplined in situations where precedent is weak. The highest-scoring work should show calm reasoning without false certainty, ambition without mythmaking, and action without pretending the unknown has disappeared.