Codex compatibility note:
-
Invoke repository skills with $skill-name in Codex; this mirrored copy rewrites legacy Claude /skill-name references.
-
Prefer the plan-hard skill for planning guidance in this Codex mirror.
-
Task tracker mandate: BEFORE executing any workflow or skill step, create/update task tracking for all steps and keep it synchronized as progress changes.
-
User-question prompts mean to ask the user directly in Codex.
-
Ignore Claude-specific mode-switch instructions when they appear.
-
Strict execution contract: when a user explicitly invokes a skill, execute that skill protocol as written.
-
Do not skip, reorder, or merge protocol steps unless the user explicitly approves the deviation first.
-
For workflow skills, execute each listed child-skill step explicitly and report step-by-step evidence.
-
If a required step/tool cannot run in this environment, stop and ask the user before adapting.
Codex Project-Reference Loading (No Hooks)
Codex does not receive Claude hook-based doc injection. When coding, planning, debugging, testing, or reviewing, open project docs explicitly using this routing.
Always read:
-
docs/project-reference/docs-index-reference.md (routes to the full docs/project-reference/* catalog)
-
docs/project-reference/lessons.md (always-on guardrails and anti-patterns)
Situation-based docs:
-
Backend/CQRS/API/domain/entity changes: backend-patterns-reference.md , domain-entities-reference.md , project-structure-reference.md
-
Frontend/UI/styling/design-system: frontend-patterns-reference.md , scss-styling-guide.md , design-system/README.md
-
Spec/test-case planning or TC mapping: feature-docs-reference.md
-
Integration test implementation/review: integration-test-reference.md
-
E2E test implementation/review: e2e-test-reference.md
-
Code review/audit work: code-review-rules.md plus domain docs above based on changed files
Do not read all docs blindly. Start from docs-index-reference.md , then open only relevant files for the task.
Quick Summary
Goal: Convert Markdown files to Microsoft Word (.docx) format with GFM support and proper formatting.
Workflow:
-
Install -- Ensure pandoc is available (required dependency)
-
Convert -- Run pandoc with docx output, apply reference template if provided
-
Verify -- Check output file for formatting fidelity
Key Rules:
-
Requires pandoc installed on the system
-
Supports GFM tables, code blocks, and images
-
Optional reference.docx template for custom styling
Be skeptical. Apply critical thinking, sequential thinking. Every claim needs traced proof, confidence percentages (Idea should be more than 80%).
markdown-to-docx
Convert markdown files to editable Microsoft Word documents with support for tables, code blocks, images, and LaTeX math equations.
Installation Required
This skill requires npm dependencies. Run one of:
Option 1: ClaudeKit CLI (recommended)
ck init
Option 2: Manual
cd .claude/skills/markdown-to-docx npm install
Dependencies: markdown-docx , gray-matter
Quick Start
Basic conversion
node .claude/skills/markdown-to-docx/scripts/convert.cjs --input ./README.md
Specify output path
node .claude/skills/markdown-to-docx/scripts/convert.cjs -i ./doc.md -o ./output.docx
With custom theme
node .claude/skills/markdown-to-docx/scripts/convert.cjs -i ./doc.md --theme ./theme.json
CLI Options
Option Short Description Default
--input
-i
Input markdown file (required)
--output
-o
Output DOCX path {input}.docx
--theme
-t
Custom theme JSON built-in
--title
Document title filename
--help
-h
Show help
Features
-
GFM Support: Tables, strikethrough, task lists
-
Code Blocks: Syntax preserved with monospace font
-
Images: Local and URL images embedded
-
Math: LaTeX equations rendered by default ($...$ , $$...$$ )
-
Frontmatter: YAML metadata extracted for title
-
No System Dependencies: Pure JavaScript, no Chrome needed
Output
Returns JSON on success:
{ "success": true, "input": "/path/to/input.md", "output": "/path/to/output.docx" }
Compatibility
Generated DOCX files work with:
-
Microsoft Word (2007+)
-
Google Docs (upload to Drive)
-
LibreOffice Writer
-
Apple Pages
[IMPORTANT] Use task tracking to break ALL work into small tasks BEFORE starting — including tasks for each file read. This prevents context loss from long files. For simple tasks, AI MUST ATTENTION ask user whether to skip.
AI Mistake Prevention — Failure modes to avoid on every task: Check downstream references before deleting. Deleting components causes documentation and code staleness cascades. Map all referencing files before removal. Verify AI-generated content against actual code. AI hallucinates APIs, class names, and method signatures. Always grep to confirm existence before documenting or referencing. Trace full dependency chain after edits. Changing a definition misses downstream variables and consumers derived from it. Always trace the full chain. Trace ALL code paths when verifying correctness. Confirming code exists is not confirming it executes. Always trace early exits, error branches, and conditional skips — not just happy path. When debugging, ask "whose responsibility?" before fixing. Trace whether bug is in caller (wrong data) or callee (wrong handling). Fix at responsible layer — never patch symptom site. Assume existing values are intentional — ask WHY before changing. Before changing any constant, limit, flag, or pattern: read comments, check git blame, examine surrounding code. Verify ALL affected outputs, not just the first. Changes touching multiple stacks require verifying EVERY output. One green check is not all green checks. Holistic-first debugging — resist nearest-attention trap. When investigating any failure, list EVERY precondition first (config, env vars, DB names, endpoints, DI registrations, data preconditions), then verify each against evidence before forming any code-layer hypothesis. Surgical changes — apply the diff test. Bug fix: every changed line must trace directly to the bug. Don't restyle or improve adjacent code. Enhancement task: implement improvements AND announce them explicitly. Surface ambiguity before coding — don't pick silently. If request has multiple interpretations, present each with effort estimate and ask. Never assume all-records, file-based, or more complex path.
Critical Thinking Mindset — Apply critical thinking, sequential thinking. Every claim needs traced proof, confidence >80% to act. Anti-hallucination: Never present guess as fact — cite sources for every claim, admit uncertainty freely, self-check output for errors, cross-reference independently, stay skeptical of own confidence — certainty without evidence root of all hallucination.
MUST ATTENTION apply critical thinking — every claim needs traced proof, confidence >80% to act. Anti-hallucination: never present guess as fact.
MUST ATTENTION apply AI mistake prevention — holistic-first debugging, fix at responsible layer, surface ambiguity before coding, re-read files after compaction.
Closing Reminders
IMPORTANT MUST ATTENTION break work into small todo tasks using task tracking BEFORE starting IMPORTANT MUST ATTENTION search codebase for 3+ similar patterns before creating new code IMPORTANT MUST ATTENTION cite file:line evidence for every claim (confidence >80% to act) IMPORTANT MUST ATTENTION add a final review todo task to verify work quality
[TASK-PLANNING] Before acting, analyze task scope and systematically break it into small todo tasks and sub-tasks using task tracking.
Hookless Prompt Protocol Mirror (Auto-Synced)
Source: .claude/hooks/lib/prompt-injections.cjs
- .claude/.ck.json
[WORKFLOW-EXECUTION-PROTOCOL] [BLOCKING] Workflow Execution Protocol — MANDATORY IMPORTANT MUST CRITICAL. Do not skip for any reason.
-
DETECT: Match prompt against workflow catalog
-
ANALYZE: Find best-match workflow AND evaluate if a custom step combination would fit better
-
ASK (REQUIRED FORMAT): Use a direct user question with this structure:
-
Question: "Which workflow do you want to activate?"
-
Option 1: "Activate [BestMatch Workflow] (Recommended)"
-
Option 2: "Activate custom workflow: [step1 → step2 → ...]" (include one-line rationale)
-
ACTIVATE (if confirmed): Call $workflow-start <workflowId> for standard; sequence custom steps manually
-
CREATE TASKS: task tracking for ALL workflow steps
-
EXECUTE: Follow each step in sequence [CRITICAL-THINKING-MINDSET] Apply critical thinking, sequential thinking. Every claim needs traced proof, confidence >80% to act. Anti-hallucination principle: Never present guess as fact — cite sources for every claim, admit uncertainty freely, self-check output for errors, cross-reference independently, stay skeptical of own confidence — certainty without evidence root of all hallucination. AI Attention principle (Primacy-Recency): Put the 3 most critical rules at both top and bottom of long prompts/protocols so instruction adherence survives long context windows.
Learned Lessons
Lessons Learned
[CRITICAL] Hard-won project debugging/architecture rules. MUST ATTENTION apply BEFORE forming hypothesis or writing code.
Quick Summary
Goal: Prevent recurrence of known failure patterns — debugging, architecture, naming, AI orchestration, environment.
Top Rules (apply always):
-
MUST ATTENTION verify ALL preconditions (config, env, DB names, DI regs) BEFORE code-layer hypothesis
-
MUST ATTENTION fix responsible layer — NEVER patch symptom sites with caller-specific defensive code
-
MUST ATTENTION use ExecuteInjectScopedAsync for parallel async + repo/UoW — NEVER ExecuteUowTask
-
MUST ATTENTION name by PURPOSE not CONTENT — adding member forces rename = abstraction broken
-
MUST ATTENTION persist sub-agent findings incrementally after each file — NEVER batch at end
-
MUST ATTENTION Windows bash: verify Python alias (where python /where py ) — NEVER assume python /python3 resolves
Debugging & Root Cause Reasoning
-
[2026-04-11] Holistic-first: verify environment before code. Failure → list ALL preconditions (config, env vars, DB names, endpoints, DI regs, credentials, permissions, data prerequisites) → verify each via evidence (grep/cat/query) BEFORE code-layer hypothesis. Worst rabbit holes: diving nearest layer while bug sits elsewhere — e.g., hours debugging "sync timeout", real cause: test appsettings pointing wrong DB. Cheapest check first.
-
[2026-04-01] Ask "whose responsibility?" before fixing. Trace: bug in caller (wrong data) or callee (wrong handling)? Fix responsible layer — NEVER patch symptom site masking real issue.
-
[2026-04-01] Trace data lifecycle, not error site. Follow data: creation → transformation → consumption. Bug usually where data created wrong, not consumed.
-
[2026-04-01] Code is caller-agnostic. Functions/handlers/consumers don't know who invokes them. Comments/guards/messages describe business intent — NEVER reference specific callers (tests, seeders, scripts).
Architecture Invariants
-
[2026-03-31] ParallelAsync + repo/UoW MUST use ExecuteInjectScopedAsync , NEVER ExecuteUowTask . ExecuteUowTask creates new UoW but reuses outer DI scope (same DbContext) — parallel iterations sharing non-thread-safe DbContext silently corrupt data. ExecuteInjectScopedAsync creates new UoW + new DI scope (fresh repo per iteration).
-
[2026-03-31] Bus message naming MUST include service name prefix — core services NEVER consume feature events. Prefix declares schema ownership (AccountUserEntityEventBusMessage = Accounts owns). Core services (Accounts, Communication) are leaders. Feature services (Growth, Talents) sending to core MUST use {CoreServiceName}...RequestBusMessage — never define own event for core to consume.
Naming & Abstraction
- [2026-04-12] Name PURPOSE not CONTENT — "OrXxx" anti-pattern. HrManagerOrHrOrPayrollHrOperationsPolicy names set members, not what it guards. Add role → rename = broken abstraction. Rule: names express DOES/GUARDS, not CONTAINS. Test: adding/removing member forces rename? YES = content-driven = bad → rename to purpose (e.g., HrOperationsAccessPolicy ). Nuance: "Or" fine in behavioral idioms (FirstOrDefault , SuccessOrThrow ) — expresses HAPPENS, not membership.
Environment & Tooling
- [2026-04-20] Windows bash: NEVER assume python /python3 resolves — verify alias first. Python may not be in bash PATH under those names. Check: where python / where py . Prefer py (Windows Python Launcher) for one-liners, node if JS alternative exists.
Test-specific lessons → docs/project-reference/integration-test-reference.md Lessons Learned section. Production-code anti-patterns → docs/project-reference/backend-patterns-reference.md Anti-Patterns section. Generic debugging/refactoring reminders → System Lessons in .claude/hooks/lib/prompt-injections.cjs .
Closing Reminders
-
IMPORTANT MUST ATTENTION holistic-first: verify ALL preconditions (config, env, DB names, endpoints, DI regs) BEFORE code-layer hypothesis — cheapest check first
-
IMPORTANT MUST ATTENTION fix responsible layer — NEVER patch symptom site; trace caller (wrong data) vs callee (wrong handling), fix root owner
-
IMPORTANT MUST ATTENTION parallel async + repo/UoW → ALWAYS ExecuteInjectScopedAsync , NEVER ExecuteUowTask (shared DbContext = silent data corruption)
-
IMPORTANT MUST ATTENTION bus message prefix = schema ownership; feature services NEVER define events for core services — use {CoreServiceName}...RequestBusMessage
-
IMPORTANT MUST ATTENTION name by PURPOSE — adding/removing member forces rename = broken abstraction
-
IMPORTANT MUST ATTENTION sub-agents MUST write findings after each file/section — NEVER batch all findings into one final write
-
IMPORTANT MUST ATTENTION Windows bash: NEVER assume python /python3 resolves — run where python /where py first, use py launcher or node
[LESSON-LEARNED-REMINDER] [BLOCKING] Task Planning & Continuous Improvement — MANDATORY. Do not skip.
Break work into small tasks (task tracking) before starting. Add final task: "Analyze AI mistakes & lessons learned".
Extract lessons — ROOT CAUSE ONLY, not symptom fixes:
-
Name the FAILURE MODE (reasoning/assumption failure), not symptom — "assumed API existed without reading source" not "used wrong enum value".
-
Generality test: does this failure mode apply to ≥3 contexts/codebases? If not, abstract one level up.
-
Write as a universal rule — strip project-specific names/paths/classes. Useful on any codebase.
-
Consolidate: multiple mistakes sharing one failure mode → ONE lesson.
-
Recurrence gate: "Would this recur in future session WITHOUT this reminder?" — No → skip $learn .
-
Auto-fix gate: "Could $code-review //simplify /$security /$lint catch this?" — Yes → improve review skill instead.
-
BOTH gates pass → ask user to run $learn . [TASK-PLANNING] [MANDATORY] BEFORE executing any workflow or skill step, create/update task tracking for all planned steps, then keep it synchronized as each step starts/completes.