2Phase 2 · The Wins
Module 3 · The Study System (Flagship)
Flagship Module

The Active Recall Engine: Flashcards & Quizzes on Demand

Lesson 3.3 5 screens · plus subject tweaks

The studying technique with the most evidence behind it.

If there's one finding cognitive scientists agree on about studying, it's this: active recall: pulling information out of your brain instead of just re-reading it on a page: is the most effective study technique that exists. Re-reading feels productive because it's familiar. Active recall feels harder because it forces your brain to generate the answer, not just recognize it. The hard part is the whole point: every time you successfully retrieve a memory, it gets stronger.

Flashcards, quizzes, MCQs, self-tests, problem sets: all variations on active recall. All work. The catch: they used to take forever to make. You'd spend a whole evening writing flashcards instead of using them. AI changes the math entirely. Your unified study doc from Lesson 3.2 is the raw material; this lesson hands you the prompts that turn it into a quiz bank in minutes.

Recognition vs. recall: why this matters for your test

Reading your notes again? Recognition. ("Yeah, I've seen this term before: I get it.") Closing your notes and trying to define the term out loud? Recall. ("Wait, what does this actually mean again?") Recognition is much easier. It's also a lie: you can pass a flashcard quiz on stuff you couldn't teach. That gap is exactly what bites you on free-response tests, in office hours, in interviews, and in the next class that builds on this one. Active recall closes the gap.

Step 1: Generate flashcards from your study doc.

Open the unified study doc you built in 3.2. Paste it into Claude. Send this.

The flashcard generator
Here's my study doc for [class] week [X]. [Paste or attach.] Build me a deck of active-recall flashcards from it. Format each card as: Front: [a question or prompt that requires me to GENERATE the answer: not recognize it] Back: [the concise answer, plus one sentence of context] Mix the difficulty: - ~30% definitions (term → definition or vice versa) - ~40% concept questions ("Why does X happen?" / "What's the difference between A and B?") - ~20% application ("If you saw [scenario], which concept applies?") - ~10% deep-end ("What would change if [assumption] were different?") Rules: - Don't write cards I could answer just by recognizing the front. - Cover EVERY ⭐ overlap from the study doc. Those are test gold. - Skip anything that's purely procedural ("the prof said this on slide 3"):focus on understanding. - 20–30 cards total. If the material warrants more or fewer, tell me why. Output as a clean Markdown list I can copy into Anki, Quizlet, or my notes app.

Where to put the cards

Anki is the gold standard for spaced repetition (free, slightly ugly, ridiculously effective). Quizlet is friendlier if you've used it before. Apple Notes / Google Keep work fine for one-off review: just paste the markdown list and read it top-to-bottom, covering the back of each card with your hand. The tool matters less than the habit.

Step 2: Run a mixed quiz, one question at a time.

Flashcards drill atoms. The mixed quiz drills how those atoms behave under pressure: when questions are about more than one concept and you don't know which one's coming next. This is closer to what an actual test feels like.

The mixed-quiz prompt
Using my study doc for [class] week [X], quiz me on the material. Format: - 12 questions total. Mix multiple-choice (4 options each), short-answer (one sentence), and 1–2 "explain in 3 sentences" prompts. - Cover every major topic from the study doc. Don't favor any one section. - Difficulty range: 3 easy, 6 medium, 3 hard. Run it like this: 1. Ask me ONE question at a time. 2. After I answer, tell me if I'm right or wrong: but DON'T explain yet. Just keep going. 3. After all 12, give me a report: - Score (X/12) - Per-topic accuracy - The specific concepts I got wrong, with a 2-sentence correction for each - The 1–2 topics I should re-study before doing this again - One follow-up question you'd ask my prof if you were me Be a tough but fair grader. If my short answer is mostly right but missing a key piece, mark it partial and tell me what was missing.

Why one-at-a-time matters

Generating 12 questions and answers all at once tempts you to skim. One at a time mimics the test experience: you don't know what's coming, you can't peek ahead, and your brain has to retrieve cold each time. That's the muscle you need on test day.

🔁

Spaced repetition

Re-quizzing on the same material at increasing intervals (1 day, 3 days, 1 week, 2 weeks) is more effective than cramming the same total time the night before. Cards you got right move further out; cards you got wrong come back tomorrow. Anki automates this. Or just re-run the mixed quiz once a week per class.

Subject tweaks: same engine, different gears.

Active recall works in every subject. The specifics bend a little depending on what you're studying. Below is the short version of how each kind of class wants the engine tuned. The downloadable Subject Cheat Sheet (linked at the bottom) has the full prompts.

🧪

STEM: math, physics, chem, bio

STEM is about working problems, not memorizing answers. The trap: asking AI to solve the problem for you (gives you a number you can't reproduce). The move: asking AI to coach you through your own attempt. Use a "don't give me the answer: ask me what my first step would be" prompt structure. Hand-write your work, snap a photo, drop it in.

📚

Humanities: history, lit, philosophy, polisci

Humanities is about argument and evidence, not memorizing facts. The trap: asking AI to write the essay or hand you the thesis (kills your voice; crosses Module 4 lines). The move: Socratic sparring partner: you state your take, Claude probes the weakest part of your reasoning. Push-back style, not validate-me style.

🌍

Languages: Spanish, French, Mandarin, anything new

The rare subject where you can ethically use AI as a conversation partner as much as you want: because the whole goal is communicative skill, and conversation is what builds it. Voice mode is a killer feature here. Twenty minutes of conversation in the target language beats an hour of vocab drills.

💻

Intro coding: CS1, web dev, data science

Coding is where AI is most powerful AND where students most easily lose the skill. The bright line: "explain my error, don't fix it" patterns keep work in your hands. "Give me code that does X" hands the assignment off.

STEM: the "coach me, don't solve it" prompt
I'm working on a [subject] problem. I want to solve it myself. Your job is to be my coach, NOT to give me the answer. Problem: [paste] Here's how I want this to work: 1. Don't show me the full solution. 2. Ask me what my first step would be. 3. If my first step is right: confirm and ask me what's next. 4. If my first step is wrong: DON'T tell me the right step. Ask a guiding question that helps me see what I missed (e.g., "what does the problem tell you about [variable]?" or "what equation describes this kind of system?"). 5. Continue until I get to the answer. 6. After I have it, show me the cleanest version of the full solution and tell me where I struggled: so I know what to drill. Subject: [class + level] Topic: [specific concept] Be patient but specific. If I'm stuck for 2–3 questions in a row, give me a bigger hint: but never the answer.
Humanities: the Socratic sparring partner
I'm preparing for [class] and I want to test my thinking on [topic / question / argument]. Be a Socratic sparring partner. Specifically: 1. I'll state my current take on the topic in 2–3 sentences. 2. Ask me ONE pointed question that probes the weakest part of my reasoning. 3. After I respond, identify whether my answer addressed the weakness or sidestepped it. 4. If sidestepped: ask the question again, sharper. 5. After 4–5 rounds, give me a debrief: - Where my argument is strongest - Where it's weakest - One counter-argument a smart opponent would raise that I should be ready to answer - One source or thinker I should engage with that would complicate or strengthen my view Be intellectually rigorous, not nice. The point is to make my actual thinking sharper, not to validate it. My take: [your argument] Class context: [name + level]
Languages: the conversation partner
I'm learning [language] at [level: beginner / intermediate / advanced / specific course like "Spanish 201"]. Be my conversation partner. Specifically: 1. We're going to have a conversation entirely in [language] about [topic: e.g., "ordering food at a restaurant" or "describing my weekend" or "discussing climate policy"]. 2. Match my level: slightly above is fine (a stretch helps), but don't drown me in vocabulary I don't know. 3. After every 3–4 of my messages, pause and tell me (in English): - One thing I said well - One mistake I made (grammar, word choice, or naturalness):show the corrected version - One word or phrase I should have used that would have sounded more native 4. Then continue the conversation in [language]. Use voice mode if I tell you I'm walking around: I want to practice speaking, not just typing. Otherwise, text is fine. Don't switch to English unless I ask. The whole point is the immersion.
Coding: "explain my error, don't fix it"
I'm taking [class: e.g., "CS101 in Python"] and I'm stuck on a problem. Here's my code and the error. Code: [paste your code] Error: [paste the error message] What I'm trying to do: [describe in one sentence] Don't give me the fixed code. Instead: 1. Explain what the error message actually means in plain language. 2. Tell me what part of MY code is causing it (point to the specific line). 3. Ask me a question that would help me figure out the fix on my own. 4. Wait for my next attempt. If I'm wrong twice in a row on the same fix, give me a slightly bigger hint: but still not the working code.

The full Subject Cheat Sheet

The reference page has all four playbooks expanded: STEM problem-set generators, humanities reading-comprehension tests, language grammar drills, and coding code-review prompts: plus the subject-specific honesty traps for each. Open it once, save it as a PDF (there's a Save-as-PDF button at the top of the page), and refer back per class. Open the Subject Cheat Sheet →

The integrity case for active recall.

Active recall is AI making you smarter, not doing your work

Honest Work Code · Rule 1: Learn with it, not instead of it. Active recall is the easiest integrity argument in the whole course: Claude isn't writing anything you'll submit. It's quizzing you on material you're trying to learn. The work happens in your brain. The "don't give me the answer" pattern in the STEM and coding prompts is the same idea: Claude as a tireless tutor, not a solution machine. You can use these moves on graded material in any class with any AI policy without ever crossing a line, because the output is your understanding, not Claude's words.

Try this: generate a 12-question quiz on the topic you're shakiest on

Pick the class and topic you're least confident about right now.

1. Paste the relevant week of your study doc into Claude.
2. Run the mixed-quiz prompt. Take it cold: no peeking.
3. Read the report. Notice the gap between "I thought I knew this" and what the score actually was.
4. Re-study only the topics that came up wrong. Re-quiz tomorrow.

This loop: quiz, gap, restudy, re-quiz: is what turns a study doc into actual mastery. It's also exactly what Lesson 3.4 (Explain It Back) sharpens to a finer edge.

Up next: the move that finds gaps a quiz can't.

Quizzes test recognition (and a bit of recall). They don't always catch the deeper gap: when you think you understand a concept but couldn't actually teach it. Lesson 3.4 introduces the explain-it-back method. Claude becomes the curious student, you teach, and the gaps surface fast.

Continue to 3.4 → Explain It Back