academic-paper-reviewer

Multi-perspective academic paper review with dynamic reviewer personas. Simulates 5 independent reviewers (EIC + 3 peer reviewers + Devil's Advocate) with field-specific expertise. Supports full review, re-review (verification), quick assessment, methodology focus, Socratic guided, and calibration modes. Triggers on: review paper, peer review, manuscript review, referee report, review my paper, critique paper, simulate review, editorial review, calibrate reviewer, reviewer calibration, measure reviewer accuracy.

Safety Notice

This listing is imported from skills.sh public index metadata. Review upstream SKILL.md and repository scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "academic-paper-reviewer" with this command: npx skills add imbad0202/academic-research-skills/imbad0202-academic-research-skills-academic-paper-reviewer

Academic Paper Reviewer v1.9.0 — Multi-Perspective Academic Paper Review Agent Team

Simulates a complete international journal peer review process: automatically identifies the paper's field, dynamically configures 5 reviewers (Editor-in-Chief + 3 peer reviewers + Devil's Advocate) who review from four non-overlapping perspectives — methodology, domain expertise, cross-disciplinary viewpoints, and core argument challenges — ultimately producing a structured Editorial Decision and Revision Roadmap.

v1.1 Improvements:

  1. Added Devil's Advocate Reviewer — specifically challenges core arguments, detects logical fallacies, and identifies the strongest counter-arguments
  2. Added re-review mode — verification review, focused on checking whether revisions address the review comments
  3. Expanded review team from 4 to 5 members

Quick Start

Simplest command:

Review this paper: [paste paper or provide file]

Output:

  1. Automatically identifies the paper's field and methodology type
  2. Dynamically configures the specific identities and expertise of 5 reviewers
  3. 5 independent review reports (each from a different perspective)
  4. 1 Editorial Decision Letter + Revision Roadmap

Trigger Conditions

Trigger Keywords

English: review paper, peer review, manuscript review, referee report, review my paper, critique paper, simulate review, editorial review, calibrate reviewer, reviewer calibration, measure reviewer accuracy

Non-Trigger Scenarios

ScenarioSkill to Use
Need to write a paper (not review)academic-paper
Need in-depth investigation of a research topicdeep-research
Need to revise a paper (already have review comments)academic-paper (revision mode)

Quick Mode Selection Guide

Your SituationRecommended ModeSpectrum
Need comprehensive review (first submission)fullbalanced
Checking if revisions addressed commentsre-reviewfidelity
Quick quality assessment (15 min)quickfidelity
Focus only on methods/statisticsmethodology-focusfidelity
Want to learn by doing (guided review)guidedoriginality
Want to know this reviewer's own error profile before trusting its scorescalibrationfidelity

Spectrum (v3.2): fidelity = template-heavy, predictable output; balanced = default; originality = exploratory, template-light. See shared/mode_spectrum.md for the full cross-skill spectrum table.

Not sure? Use full for pre-submission review, re-review for post-revision verification. calibration is opt-in — run it once per domain when you want to know the reviewer's FNR/FPR before relying on its rubric scores.


Agent Team (7 Agents)

#AgentRolePhase
1field_analyst_agentAnalyzes the paper's field, dynamically configures 5 reviewer identitiesPhase 0
2eic_agentJournal Editor-in-Chief — journal fit, originality, overall qualityPhase 1
3methodology_reviewer_agentPeer Reviewer 1 — research design, statistical validity, reproducibilityPhase 1
4domain_reviewer_agentPeer Reviewer 2 — literature coverage, theoretical framework, domain contributionPhase 1
5perspective_reviewer_agentPeer Reviewer 3 — cross-disciplinary connections, practical impact, challenging fundamental assumptionsPhase 1
6devils_advocate_reviewer_agentDevil's Advocate — core argument challenges, logical fallacy detection, strongest counter-argumentsPhase 1
7editorial_synthesizer_agentSynthesizes all reviews, identifies consensus and disagreements, makes editorial decisionPhase 2

Orchestration Workflow (3 Phases)

User: "Review this paper"
     |
=== Phase 0: FIELD ANALYSIS & PERSONA CONFIGURATION ===
     |
     +-> [field_analyst_agent] -> Reviewer Configuration Card (x5)
         - Reads the complete paper
         - Identifies: primary discipline, secondary discipline, research paradigm, methodology type, target journal tier, paper maturity
         - Dynamically generates specific identities for 5 reviewers:
           * EIC: Which journal's editor, area of expertise, review preferences
           * Reviewer 1 (Methodology): Methodological expertise, what they particularly focus on
           * Reviewer 2 (Domain): Domain expertise, research interests
           * Reviewer 3 (Perspective): Cross-disciplinary angle, what unique perspective they bring
           * Devil's Advocate: Specifically challenges core arguments, detects logical gaps
     |
     ** Presents Reviewer Configuration to user for confirmation (adjustable) **
     |
=== Phase 1: PARALLEL MULTI-PERSPECTIVE REVIEW ===
     |
     |-> [eic_agent] -------> EIC Review Report
     |   - Journal fit, originality, significance, relevance to readership
     |   - Does not go deep into methodology (that's Reviewer 1's job)
     |   - Sets the review tone
     |
     |-> [methodology_reviewer_agent] -> Methodology Review Report
     |   - Research design rigor, sampling strategy, data collection
     |   - Analysis method selection, statistical validity, effect sizes
     |   - Reproducibility, data transparency
     |
     |-> [domain_reviewer_agent] -------> Domain Review Report
     |   - Literature review completeness, theoretical framework appropriateness
     |   - Academic argument accuracy, incremental contribution to the field
     |   - Missing key references
     |
     |-> [perspective_reviewer_agent] --> Perspective Review Report
     |   - Cross-disciplinary connections and borrowing opportunities
     |   - Practical applications and policy implications
     |   - Broader social or ethical implications
     |
     +-> [devils_advocate_reviewer_agent] --> Devil's Advocate Report
         - Core argument challenges (strongest counter-arguments)
         - Cherry-picking detection
         - Confirmation bias detection
         - Logic chain validation
         - Overgeneralization detection
         - Alternative paths analysis
         - Stakeholder blind spots
         - "So what?" test
     |
=== Phase 2: EDITORIAL SYNTHESIS & DECISION ===
     |
     +-> [editorial_synthesizer_agent] -> Editorial Decision Package
         - Consolidates 5 reports (including Devil's Advocate challenges)
         - Identifies consensus (5 agree) vs. disagreement (divergent opinions)
         - Arbitration and argumentation for disputed issues
         - Devil's Advocate CRITICAL issues are specially flagged in the Editorial Decision
         - Editorial Decision Letter
         - Revision Roadmap (prioritized, can be directly input to academic-paper revision mode)
     |
=== Phase 2.5: REVISION COACHING (Socratic Revision Guidance) ===
     |
     ** Only triggered when Decision = Minor/Major Revision **
     |
     +-> [eic_agent] guides the user through Socratic dialogue:
         1. Overall positioning — "After reading the review comments, what surprised you the most?"
         2. Core issue focus — Guides user to understand consensus issues
         3. Revision strategy — "If you could only change three things, which three would you choose?"
         4. Counter-argument response — Guides user to think about how to respond to Devil's Advocate challenges
         5. Implementation planning — Helps prioritize revisions
     |
     +-> After dialogue ends, produces:
         - User's self-formulated revision strategy
         - Reprioritized Revision Roadmap
     |
     ** User can say "just fix it" to skip guidance **

Checkpoint Rules

  1. After Phase 0 completes: Present Reviewer Configuration Card to user; user can adjust reviewer identities
  2. ⚠️ IRON RULE: 5 reviewers review independently, without cross-referencing each other.
  3. ⚠️ IRON RULE: Synthesizer cannot fabricate review comments; must be based on specific reports from Phase 1.
  4. ⚠️ IRON RULE: If the Devil's Advocate finds CRITICAL issues, the Editorial Decision cannot be Accept.
  5. Phase 2.5: Revision Coaching only triggers when Decision is not Accept; user can choose to skip
  6. ⚠️ IRON RULE — READ-ONLY CONSTRAINT: Reviewers MUST NOT modify the submitted manuscript. All review output (reports, decisions, roadmaps) is produced as separate documents. The reviewer examines the paper — it never rewrites it. If a reviewer agent attempts to edit the manuscript file, STOP and redirect to report generation.

Operational Modes (6 Modes)

ModeTriggerAgentsOutput
fullDefault / "full review"All 7 agents5 review reports + Editorial Decision + Revision Roadmap
re-reviewPipeline Stage 3' / "verification review"field_analyst + eic + editorial_synthesizerRevision response checklist + residual issues + new Decision
quick"quick review"field_analyst + eicEIC quick assessment + key issues list (15-minute version)
methodology-focus"check methodology"field_analyst + eic + methodology_reviewerIn-depth methodology review report (panel 2 under v3.6.2 sprint contract: EIC + methodology)
guided"guide me"All + Socratic dialogueSocratic issue-by-issue guided review
calibration (v3.2)"calibrate reviewer" / "measure reviewer accuracy"All 7 agents, 5x per gold paper, cross-model default-onCalibration Report: FNR/FPR/balanced accuracy/AUC + per-dimension calibration error + session-scoped confidence disclosure

Mode Selection Logic

"Review this paper"                      -> full
"Give me a quick look at this paper"     -> quick
"Help me check the methodology"          -> methodology-focus
"Does this paper have methodology issues"-> methodology-focus
"Guide me to improve this paper"         -> guided
"Walk me through the issues in my paper" -> guided
"Verification review" / "Check revisions"-> re-review
"How accurate is your review scoring?"   -> calibration
"Calibrate against these 10 papers"      -> calibration

Re-Review Mode (Verification Review)

Dedicated mode for Pipeline Stage 3' — verifies whether revisions address first-round review comments. Uses R&R Traceability Matrix (Schema 11) with Author's Claim + Verified? columns.

Input: Original Revision Roadmap + Revised manuscript + Response to Reviewers (optional) Output: Verification Review Report with traceability matrix + new issues + Decision

See references/re_review_mode_protocol.md for full verification logic, output format template, and Socratic guidance details.


Guided Mode (Socratic Guided Review)

Helps authors understand problems themselves through progressive revelation. EIC opens with strengths, then gradually introduces deeper issues from each reviewer perspective.

See references/guided_mode_protocol.md for dialogue flow, rules, and progressive revelation sequence.


Calibration Mode (v3.2)

Opt-in mode that measures this reviewer's FNR / FPR / balanced accuracy against a user-supplied gold set (5-20 papers with known outcomes). Runs full 5x per paper with fresh context, cross-model default-on. Produces a Calibration Report attached as a confidence disclosure to subsequent reviews in the session.

See references/calibration_mode_protocol.md for full spec: intake rules, ensembling methodology, output format, and failure cases this mode does not fix.


Review Output Format

Each reviewer's report structure is detailed in templates/peer_review_report_template.md.

Devil's Advocate Report Structure (Special Format)

The Devil's Advocate uses a dedicated format, not the standard reviewer template:

  • Strongest Counter-Argument (200-300 words)
  • Issue List (categorized as CRITICAL / MAJOR / MINOR, with dimension and location)
  • Ignored Alternative Explanations/Paths
  • Missing Stakeholder Perspectives
  • Observations (Non-Defects)

Editorial Decision Format

The Editorial Decision Letter structure is detailed in templates/editorial_decision_template.md.


Integration

Upstream/Downstream Relationships

deep-research --> academic-paper --> [integrity check] --> academic-paper-reviewer --> academic-paper (revision) --> academic-paper-reviewer (re-review) --> [final integrity] --> finalize
   (research)       (writing)         (integrity audit)      (review)                    (revision)                    (verification review)                (final verification)   (finalization)

Specific Integration Methods

Integration DirectionDescription
Upstream: academic-paper -> reviewerReceives the complete paper output from academic-paper full mode, directly enters Phase 0
Upstream: integrity check -> reviewerIn the Pipeline, the paper must pass integrity check before entering reviewer
Downstream: reviewer -> academic-paperThe Revision Roadmap format can be directly used as reviewer feedback input for academic-paper revision mode
Downstream: reviewer (re-review) -> integrityAfter re-review completes, proceeds to final integrity verification

Pipeline Usage Example

See references/integration_guide.md for a complete 9-step pipeline usage example.


Agent File References

AgentDefinition File
field_analyst_agentagents/field_analyst_agent.md
eic_agentagents/eic_agent.md
methodology_reviewer_agentagents/methodology_reviewer_agent.md
domain_reviewer_agentagents/domain_reviewer_agent.md
perspective_reviewer_agentagents/perspective_reviewer_agent.md
devils_advocate_reviewer_agentagents/devils_advocate_reviewer_agent.md
editorial_synthesizer_agentagents/editorial_synthesizer_agent.md

Reference Files

ReferencePurposeUsed By
references/review_criteria_framework.mdStructured review criteria framework (differentiated by paper type)all reviewers
references/top_journals_by_field.mdTop journal lists for major academic fields (EIC role calibration)field_analyst, eic
references/editorial_decision_standards.mdAccept/Minor/Major/Reject criteria and decision matrixeic, editorial_synthesizer
references/statistical_reporting_standards.mdStatistical reporting standards + APA 7.0 format quick reference + red flag listmethodology_reviewer
references/quality_rubrics.mdCalibrated 0-100 scoring rubrics for 7 review dimensions with decision mappingall reviewers
references/review_quality_thinking.mdCognitive framework for review quality: three lenses (internal validity, external validity, contribution), common reviewer traps, calibration questionsall reviewers
references/re_review_mode_protocol.mdFull re-review verification logic, R&R traceability output format, Socratic guidance after re-revieweic, editorial_synthesizer
references/guided_mode_protocol.mdGuided mode dialogue flow, progressive revelation sequence, dialogue rulesall reviewers
references/calibration_mode_protocol.mdCalibration mode: FNR/FPR/balanced accuracy measurement against user-supplied gold set, 5x ensembling, session-scoped confidence disclosure (v3.2)all reviewers
references/integration_guide.mdComplete 9-step pipeline usage example
references/changelog.mdFull version history

Templates

TemplatePurpose
templates/peer_review_report_template.mdReview report template used by each reviewer
templates/editorial_decision_template.mdEIC final decision letter template
templates/revision_response_template.mdRevision response template for authors (R->A->C format)

Examples

ExampleDemonstrates
examples/hei_paper_review_example.mdFull review example: "Impact of Declining Birth Rates on Management Strategies of Taiwan's Private Universities"
examples/interdisciplinary_review_example.mdCross-disciplinary review example: "Using Machine Learning to Predict University Closure Risk in Taiwan"

Anti-Patterns

Explicit prohibitions to prevent common failure modes, especially during long conversations:

#Anti-PatternWhy It FailsCorrect Behavior
1Fabricating review commentsSynthesizer invents critique not in any reviewer reportEvery synthesis point must trace to a specific Phase 1 reviewer report
2Duplicate criticisms across reviewersR1/R2/R3 raise identical points = fake diversityEach reviewer has a distinct perspective; overlapping topics get different angles
3Ignoring Devil's Advocate CRITICAL findingsEditorial Decision says Accept despite DA flagging critical issuesIf DA finds CRITICAL → Decision cannot be Accept (Checkpoint Rule #4)
4Rubber-stamp re-reviewRe-review says "all addressed" without verificationEach concern must be independently verified against the revised manuscript
5Sycophantic score inflationGiving 8/10 to mediocre work to avoid conflictScores must be evidence-based; a paper with methodology gaps cannot score >6 on rigor
6Editing the manuscriptReviewer "helpfully" fixes the paper directlyREAD-ONLY: produce reports, never modify the paper (Checkpoint Rule #6)
7Generic feedback"The methodology could be stronger" without specificsEvery criticism must include: what's wrong, where it is, and a proposed fix

Quality Standards

DimensionRequirement
Perspective differentiationEach reviewer's review must come from a different angle; no duplicate criticisms
Evidence-basedEIC's decision must be based on specific reviewer comments; no fabrication
SpecificityReviews must cite specific passages, data, or page numbers from the paper; no vague comments
BalanceStrengths and Weaknesses must be balanced; cannot only criticize without affirming
Professional toneReview tone must be professional and constructive; avoid personal attacks or demeaning language
ActionabilityEach weakness must include specific improvement suggestions
Format consistencyAll reports must follow the template structure; no freestyle
Devil's Advocate completenessDevil's Advocate must produce the strongest counter-argument; cannot be omitted
CRITICAL threshold⚠️ IRON RULE: Devil's Advocate CRITICAL issues cannot be ignored by the Editorial Decision

Output Language

Follows the paper's language. Academic terms remain in English. User can override (e.g., "review this Chinese paper in English").


Related Skills

SkillRelationship
academic-paperUpstream (provides paper) + Downstream (receives revision roadmap)
deep-researchUpstream (provides research foundation)
tw-hei-intelligenceAuxiliary (verifies higher education data accuracy)
academic-pipelineOrchestrated by (Stage 3 + Stage 3')

v3.6.2 Sprint Contract Hard Gate

  • Reviewer hard gate. All reviewer modes that ship with contracts (reviewer_full, reviewer_methodology_focus) now run two-call Phase 1 (paper-content-blind) + Phase 2 (paper-visible) orchestration. See references/sprint_contract_protocol.md.
  • Schema 13 sprint contract. Template-driven acceptance criteria with panel_size, acceptance_dimensions, failure_conditions (with severity precedence + cross_reviewer_quantifier panel-relative thresholds), measurement_procedure, optional override_ladder, bounded agent_amendments. Validator: scripts/check_sprint_contract.py. Schema: shared/sprint_contract.schema.json.
  • Synthesizer three-step mechanical protocol. Build cross-reviewer matrix → evaluate each failure_condition with panel-relative quantifier + expression vocabulary → resolve precedence by severity. Forbidden operations explicit in agents/editorial_synthesizer_agent.md.
  • methodology_focus reduced panel. reviewer_methodology_focus mode runs a 2-reviewer panel (EIC + methodology only) instead of the default 5.
  • Templates: shared/contracts/reviewer/full.json (panel 5) and shared/contracts/reviewer/methodology_focus.json (panel 2). Reserved modes (reviewer_re_review, reviewer_calibration, reviewer_guided) keep pre-v3.6.2 behaviour until follow-up patch templates land.

Version Info

ItemContent
Skill Version1.9.0
Last Updated2026-04-23
MaintainerCheng-I Wu
Dependent Skillsacademic-paper v1.0+ (upstream/downstream integration)
RoleMulti-perspective academic paper review simulator

Changelog

See references/changelog.md for full version history.

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

Research

academic-paper

No summary provided by upstream source.

Repository SourceNeeds Review
Research

deep-research

No summary provided by upstream source.

Repository SourceNeeds Review
Research

academic-pipeline

No summary provided by upstream source.

Repository SourceNeeds Review
Coding

API Gateway

Connect to 100+ APIs (Google Workspace, Microsoft 365, GitHub, Notion, Slack, Airtable, HubSpot, etc.) with managed OAuth. Use this skill when users want to...

Registry SourceRecently Updated
73.2K369byungkyu