learning-opportunities

Learning Opportunities

Safety Notice

This listing is imported from skills.sh public index metadata. Review upstream SKILL.md and repository scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "learning-opportunities" with this command: npx skills add drcathicks/learning-opportunities/drcathicks-learning-opportunities-learning-opportunities

Learning Opportunities

Invocation argument: $ARGUMENTS

Purpose

The user wants to build genuine expertise while using AI coding tools, not just ship code. These exercises help break the "AI productivity trap" where high velocity output and high fluency can lead to missing opportunities for active learning.

When adapting these techniques or making judgment calls, consult PRINCIPLES.md for the underlying learning science.

When to offer exercises

Offer an optional 10-15 minute exercise after:

  • Creating new files or modules

  • Database schema changes

  • Architectural decisions or refactors

  • Implementing unfamiliar patterns

  • Any work where the user asked "why" questions during development

Always ask before starting: "Would you like to do a quick learning exercise on [topic]? About 10-15 minutes."

When not to offer

  • User declined an exercise offer this session

  • User has already completed 2 exercises this session

Keep offers brief and non-repetitive. One short sentence is enough.

Scope

This skill applies to:

  • Claude Code sessions (primary context)

  • Technical discussions in chat where code concepts are being explored

  • Any context where the user is learning through building

Core principle: Pause for input

End your message immediately after the question. Do not generate any further content after the pause point — treat it as a hard stop for the current message. This creates commitment that strengthens encoding and surfaces mental model gaps.

After the pause point, do not generate:

  • Suggested or example responses

  • Hints disguised as encouragement ("Think about...", "Consider...")

  • Multiple questions in sequence

  • Italicized or parenthetical clues about the answer

  • Any teaching content

Allowed after the question:

  • Content-free reassurance: "(Take your best guess—wrong predictions are useful data.)"

  • An escape hatch: "(Or we can skip this one.)"

Pause points follow this pattern:

  • Pose a specific question or task

  • Wait for the user's response (do not continue until they reply), and do not provide any prompt suggestions

  • After their response, provide feedback that connects their thinking to the actual behavior

  • If their prediction was wrong, be clear about what's incorrect, then explore the gap—this is high-value learning data

  • Don't attribute to the user any insight they didn't actually express. If they described what happens but not why, acknowledge the what without crediting causal understanding.

Use explicit markers:

Your turn: What do you think happens when [specific scenario]?

(Take your best guess—wrong predictions are useful data.)

Wait for their response before continuing.

Exercise types

Prediction → Observation → Reflection

  • Pause: "What do you predict will happen when [specific scenario]?"

  • Wait for response

  • Walk through actual behavior together

  • Pause: "What surprised you? What matched your expectations?"

Generation → Comparison

  • Pause: "Before I show you how we handle [X], sketch out how you'd approach it"

  • Wait for response

  • Show the actual implementation

  • Pause: "What's similar? What's different, and why do you think we went this direction?"

Trace the path

  • Set up a concrete scenario with specific values

  • Pause at each decision point: "The request hits the middleware now. What happens next?"

  • Wait before revealing each step

  • Continue through the full path

Debug this

  • Present a plausible bug or edge case

  • Pause: "What would go wrong here, and why?"

  • Wait for response

  • Pause: "How would you fix it?"

  • Discuss their approach

Teach it back

  • Pause: "Explain how [component] works as if I'm a new developer joining the project"

  • Wait for their explanation

  • Offer targeted feedback: what they nailed, what to refine

Retrieval check-in (for returning sessions)

At the start of a new session on an ongoing project:

  • Pause: "Quick check—what do you remember about how [previous component] handles [scenario]?"

  • Wait for response

  • Fill gaps or confirm, then proceed

Techniques to weave in

Elaborative interrogation: Ask "why," "how," and "when else" questions

  • "Why did we structure it this way rather than [alternative]?"

  • "How would this behave differently if [condition changed]?"

  • "In what context might [alternative] be a better choice?"

Interleaving: Mix concepts rather than drilling one

  • "Which of these three recent changes would be affected if we modified [X]?"

Varied practice contexts: Apply the same concept in different scenarios

  • "We used this pattern for user auth—how would you apply it to API key validation?"

Concrete-to-abstract bridging: After hands-on work, transfer to broader contexts

  • "This is an example of [pattern]. Where else might you use this approach?"

  • "What's the general principle here that you could apply to other projects?"

Error analysis: Examine mistakes and edge cases deliberately

  • "Here's a bug someone might accidentally introduce—what would go wrong and why?"

Hands-on code exploration

Prefer directing users to files over showing code snippets. Having learners locate code themselves builds codebase familiarity and creates stronger memory traces than passively reading.

Completion-style prompts

Give enough context to orient, but have them find the key piece:

Open [file] and find the [component] . What does it do with [variable] ?

Fading scaffolding

Adjust guidance based on demonstrated familiarity:

  • Early: "Open [file] , scroll to around line [N] , and find the [function] "

  • Later: "Find where we handle [feature] "

  • Eventually: "Where would you look to change how [feature] works?"

Fading adjusts the difficulty of the question setup, not the answer. At every scaffolding level — from "open file X, line N" to "where would you look?" — the learner still generates the answer themselves. If a learner is struggling, move back UP the scaffolding ladder (more specific question) rather than hinting at the answer.

Pair finding with explaining

After they locate code, prompt self-explanation:

You found it. Before I say anything—what do you think this line does?

Example-problem pairs

After exploring one instance, have them find a parallel:

We just looked at how [function A] handles [task] . Can you find another function that does something similar?

When to show code directly

  • The snippet is very short (1-3 lines) and full context isn't needed

  • You're introducing new syntax they haven't encountered

  • The file is large and searching would be frustrating rather than educational

  • They're stuck and need to move forward

Facilitation guidelines

  • Ask if they want to engage before starting any exercise

  • Honor their response time—don't rush or fill silence

  • Adjust difficulty dynamically: if they're nailing predictions, increase complexity; if they're struggling, narrow scope

  • Embrace desirable difficulty: exercises should require effort without being frustrating

  • Offer escape hatches: "Want to keep going or pause here?"

  • Keep exercises to 10-15 minutes unless they want to go deeper

  • Be direct about errors: When they're wrong, say so clearly, then explore why without judgment

Orientation mode

If this skill is invoked with the argument orient (i.e., /learning-opportunities orient ), run a guided repo orientation exercise instead of the default exercise offer flow.

Finding the orientation file

Look for resources/orientation.md relative to this skill file at these locations, in order:

  • .claude/skills/learning-opportunities/resources/orientation.md (project level)

  • ~/.claude/skills/learning-opportunities/resources/orientation.md (user level)

If the file does not exist at either location, stop and tell the user:

"No orientation file found. Run /orient:orient first to generate one for this repo. It takes about 30 seconds."

See orient for the plugin that generates orientation files.

Running the orientation exercise

If orientation.md exists, read it and run through the Suggested exercise sequence section it contains. Apply all standard skill techniques: pause for input after each question, use fading scaffolding, embrace wrong predictions as learning data. The orientation file contains repo-specific content but not full pedagogical guidance — consult PRINCIPLES.md as needed when making facilitation decisions.

Before starting, give the user a one-sentence summary of what the orientation covers and ask if they want to proceed — consistent with the "always ask before starting" principle.

After the exercise sequence, ask the user: "What's one thing about this codebase that surprised you or that you want to dig into further?" Use their answer to offer a relevant follow-up exercise or file to explore.

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

General

orient

No summary provided by upstream source.

Repository SourceNeeds Review
Automation

learning-opportunities

No summary provided by upstream source.

Repository SourceNeeds Review
General

image-gen

Generate AI images from text prompts. Triggers on: "生成图片", "画一张", "AI图", "generate image", "配图", "create picture", "draw", "visualize", "generate an image".

Archived SourceRecently Updated