Learn
Conduct a structured knowledge acquisition session — surveying unfamiliar territory, building initial models, testing them through deliberate exploration, integrating findings into coherent understanding, and consolidating for durable retrieval.
When to Use
- Encountering an unfamiliar codebase, framework, or domain with no prior context
- A user asks about a topic outside current working knowledge and the answer requires genuine investigation, not recall
- Multiple conflicting sources or patterns exist and a coherent mental model needs to be built from scratch
- After
remote-viewingsurfaces intuitive leads that need systematic validation - Preparing to
teacha topic — the AI must first understand it deeply enough to explain it
Inputs
- Required: Learning target — a topic, codebase area, API, domain concept, or technology to understand
- Optional: Scope boundary — how deep to go (surface survey vs. deep expertise)
- Optional: User's purpose — why this knowledge matters (guides which aspects to prioritize)
- Optional: Known starting points — files, docs, or concepts already familiar
Procedure
Step 1: Survey — Map the Territory
Before attempting to understand anything, map the landscape to identify what exists.
Learning Modality Selection:
┌──────────────────┬──────────────────────────┬──────────────────────────┐
│ Territory Type │ Primary Modality │ Tool Pattern │
├──────────────────┼──────────────────────────┼──────────────────────────┤
│ Codebase │ Structural mapping — │ Glob for file tree, │
│ │ find entry points, core │ Grep for exports/imports,│
│ │ modules, boundaries │ Read for key files │
├──────────────────┼──────────────────────────┼──────────────────────────┤
│ API / Library │ Interface mapping — │ WebFetch for docs, │
│ │ find public surface, │ Read for examples, │
│ │ types, configuration │ Grep for usage patterns │
├──────────────────┼──────────────────────────┼──────────────────────────┤
│ Domain concept │ Ontology mapping — │ WebSearch for overviews, │
│ │ find core terms, │ WebFetch for definitions,│
│ │ relationships, debates │ Read for local notes │
├──────────────────┼──────────────────────────┼──────────────────────────┤
│ User's context │ Conversational mapping │ Read conversation, │
│ │ — find stated goals, │ Read MEMORY.md, │
│ │ preferences, constraints │ Read CLAUDE.md │
└──────────────────┴──────────────────────────┴──────────────────────────┘
- Identify the territory type and select the primary modality
- Perform a broad scan — not reading deeply, but identifying landmarks (key files, entry points, core concepts)
- Note the boundaries: what is in scope, what is adjacent, what is out of scope
- Identify gaps: areas that look important but are opaque from the surface
- Create a rough map: list the major components and their apparent relationships
Expected: A skeletal map of the territory with 5-15 landmarks identified. A sense of which areas are clear from the surface and which require deeper investigation. No understanding yet — just a map.
On failure: If the territory is too large to survey, narrow scope immediately. Ask: "What is the minimum I need to understand to serve the user's purpose?" If the territory has no clear entry point, start from the output (what does this system produce?) and trace backward.
Step 2: Hypothesize — Build Initial Models
From the survey, construct initial hypotheses about how the system works.
- Formulate 2-3 hypotheses about the territory's structure or behavior
- State each hypothesis clearly: "I believe X because I observed Y"
- For each hypothesis, identify what evidence would confirm it and what would refute it
- Rank hypotheses by confidence: which feels most supported, which is shakiest
- Identify the highest-value hypothesis to test first (the one that, if confirmed, would unlock the most understanding)
Expected: Concrete, falsifiable hypotheses — not vague impressions. Each has a test that would confirm or refute it. The hypotheses collectively cover the most important aspects of the territory.
On failure: If no hypotheses form, the survey was too shallow — return to Step 1 and read 2-3 landmarks in depth. If all hypotheses feel equally uncertain, start with the simplest one (Occam's razor) and build from there.
Step 3: Explore — Probe and Test
Systematically test each hypothesis through targeted investigation.
- Select the highest-priority hypothesis
- Design a minimal probe: what is the smallest investigation that would confirm or refute it?
- Execute the probe (read a file, search for a pattern, test an assumption)
- Record the result: confirmed, refuted, or modified
- If refuted, update the hypothesis based on the new evidence
- If confirmed, probe deeper: does the hypothesis hold at the edges, or only in the center?
- Move to the next hypothesis and repeat
Expected: At least one hypothesis tested to conclusion. The mental model is beginning to take shape — some parts confirmed, some revised. Surprises are noted as particularly valuable data.
On failure: If probes consistently produce ambiguous results, the hypotheses may be testing the wrong things. Step back and ask: "What would someone who understands this system consider the most important fact?" Probe for that instead.
Step 4: Integrate — Build Mental Model
Synthesize findings into a coherent model that connects the pieces.
- Review all confirmed hypotheses and revised models
- Identify the central organizing principle: what is the "spine" that everything connects to?
- Map relationships: which components depend on which? What flows where?
- Identify the surprising findings — these often contain the deepest insight
- Look for patterns that repeat across different parts of the territory
- Build a mental model that can predict behavior: "Given input X, I expect Y because Z"
Expected: A coherent mental model that explains the territory's structure and predicts its behavior. The model should be expressible in 3-5 sentences and should make specific claims, not vague generalizations.
On failure: If the pieces do not integrate into a coherent model, there may be a fundamental misunderstanding in one of the earlier hypotheses. Identify the piece that does not fit and re-test it. Alternatively, the territory may genuinely be incoherent (poorly designed systems exist) — note this as a finding rather than forcing coherence.
Step 5: Verify — Challenge Understanding
Test the mental model by making predictions and checking them.
- Use the model to make 3 specific predictions about the territory
- Test each prediction through investigation (not by assuming it is true)
- For each confirmed prediction, confidence increases
- For each refuted prediction, identify where the model is wrong and correct it
- Identify edge cases: does the model hold at the boundaries, or does it break down?
- Ask: "What would surprise me?" — then check if that surprise is possible
Expected: The mental model survives at least 2 of 3 prediction tests. Where it breaks, the failure is understood and the model is corrected. The model now has both confirmed strengths and known limitations.
On failure: If most predictions fail, the mental model has a fundamental flaw. This is actually valuable information — it means the territory works differently than expected. Return to Step 2 with the new evidence and rebuild the hypotheses from scratch. The second attempt will be much faster because the wrong models have been eliminated.
Step 6: Consolidate — Store for Retrieval
Capture the learning in a form that supports future retrieval and application.
- Summarize the mental model in 3-5 sentences
- Note the key landmarks — the 3-5 most important things to remember
- Record any counterintuitive findings that might be forgotten
- Identify related topics that this learning connects to
- If the learning is durable (will be needed across sessions), update MEMORY.md
- If the learning is session-specific, note it as context for the current conversation
- State what remains unknown — honest gaps are more useful than false confidence
Expected: A concise, retrievable summary that captures the essential understanding. Future references to this topic can start from this summary rather than re-learning from scratch.
On failure: If the learning resists summarization, it may not yet be fully integrated — return to Step 4. If the learning seems too obvious to be worth storing, consider that what feels obvious now may not feel obvious in a fresh context. Store the non-obvious parts.
Validation
- A survey was conducted before any deep investigation (map before dive)
- Hypotheses were explicitly stated and tested, not assumed
- At least one hypothesis was revised based on evidence (indicates genuine learning)
- The mental model makes specific, testable predictions about the territory
- Known unknowns are identified alongside known knowns
- The consolidated summary is concise enough to be useful for future retrieval
Common Pitfalls
- Skipping the survey: Diving into detail before understanding the landscape wastes time on unimportant areas and misses the big picture
- Unfalsifiable hypotheses: "This is probably complex" cannot be tested. "This module handles authentication because it imports crypto" can be
- Confirmation bias during exploration: Seeking only evidence that supports the initial hypothesis while ignoring contradictions
- Premature consolidation: Storing a model before it has been tested leads to confidently wrong future predictions
- Perfectionism: Attempting to learn everything before applying any knowledge. Learning is iterative — use partial understanding, then refine
- Learning without purpose: Acquiring knowledge with no application in mind produces unfocused, shallow understanding
Related Skills
learn-guidance— the human-guidance variant for coaching a person through structured learningteach— knowledge transfer calibrated to a learner; builds on the model constructed hereremote-viewing— intuitive exploration that surfaces leads for systematic learning to validatemeditate— clearing prior context noise before entering a new learning territoryobserve— sustained neutral pattern recognition that feeds learning with raw data