Day 5 - Session 1: Breaking Things Systematically
Finding where and why AI fails
Good Morning!
Today’s Big Question
When do things break?
This Morning
- Trigger failure modes systematically
- Understand why failures happen
- Build your “don’t trust” catalog
- Learn to predict breakage
Show Me Your Prompts! (15 min)
Final Sharing Session
- What sustainable practices did you develop yesterday?
- Any overnight discoveries?
- What are you still struggling with?
What We’ll Learn This Session
By the end of this morning, you will be able to:
- Analyze: Identify failure patterns across contexts (Marzano Level 3)
- Evaluate: Predict when AI will fail (Marzano Level 4)
- Create: Build personal evaluation criteria (Marzano Level 4)
Why Break Things on Purpose?
The Goal
Know failure modes BEFORE they matter
From Research
- “Always confident, usually correct”
- No epistemic humility
- Confabulation as default
- Sycophancy over accuracy
Exercise: Systematic Failure Exploration (45 min)
Five Stations, Five Failures
TODO: Rotation through failure modes
Failure 1: Sycophancy Test
TODO: False premise agreement
Failure 2: Token Blindness
TODO: Character counting
Failure 3: Self-Contradiction
TODO: Context manipulation
Failure 4: Capability Lies
TODO: Claiming impossible abilities
Failure 5: Temporal Confusion
TODO: Time-based errors
The Pattern Behind Failures
Discussion: Building Evaluation Criteria (20 min)
Your Personal Checklist
When NOT to trust AI: - Factual claims without sources - Counting or precise measurement - Recent events - Reasoning about reasoning - [Add your discoveries]
Looking Ahead
This Afternoon: Synthesis
- What we’ve learned
- Where to go next
- Community building
Key Question
How do we maintain human judgment when tools seem so capable?
Answering Today’s Question
When do things break?
Things break when: - Tasks require actual understanding - Precision matters more than plausibility - Context exceeds window - Time sense required - Self-knowledge needed
Our tools break when we: * Don’t manage state * Don’t treat them like tools * Don’t check the output
Out services and expectations of use break when: - Users misunderstand training vs. chatting - Moral responsibility gets forgotten - Privacy assumptions prove false - Terms of service change unexpectedly
The boring stuff matters because consequences are real.
Before Lunch
Document your failure catalog: - Specific triggers you found - Warning signs to watch - Tasks to never trust - Verification strategies
Bring to final session!