sec-edgar-pipeline

This pipeline is centered on edgar-analyzer and the EDGAR data sources. The core loop is: configure credentials, create a project with examples, analyze patterns, generate code, run extraction, and export reports.

Safety Notice

This listing is imported from skills.sh public index metadata. Review upstream SKILL.md and repository scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "sec-edgar-pipeline" with this command: npx skills add bobmatnyc/claude-mpm-skills/bobmatnyc-claude-mpm-skills-sec-edgar-pipeline

SEC EDGAR Pipeline

Overview

This pipeline is centered on edgar-analyzer and the EDGAR data sources. The core loop is: configure credentials, create a project with examples, analyze patterns, generate code, run extraction, and export reports.

Setup (Keys + User Agent)

Use the setup wizard to configure required keys:

python -m edgar_analyzer setup

or

edgar-analyzer setup

Required entries:

  • OPENROUTER_API_KEY

  • (Optional) JINA_API_KEY

  • EDGAR user agent string ("Name email@example.com")

End-to-End CLI Workflow

1. Create project

edgar-analyzer project create my_project --template minimal

2. Add examples + project.yaml

projects/my_project/examples/*.json

3. Analyze examples

edgar-analyzer analyze-project projects/my_project

4. Generate extraction code

edgar-analyzer generate-code projects/my_project

5. Run extraction

edgar-analyzer run-extraction projects/my_project --output-format csv

Outputs land in projects/<name>/output/ .

EDGAR-Specific Conventions

  • CIK values are 10-digit, zero-padded (e.g., 0000320193 ).

  • Rate limit: SEC API allows 10 requests/sec. Scripts use ~0.11s delays.

  • User agent is mandatory; include name + email.

Scripted Example (Apple DEF 14A)

edgar/scripts/fetch_apple_def14a.py shows the direct flow:

  • Fetch latest DEF 14A metadata

  • Download HTML

  • Parse Summary Compensation Table (SCT)

  • Save raw HTML + extracted JSON + ground truth

Recipe-Driven Extraction

edgar/recipes/sct_extraction/config.yaml defines a multi-step pipeline:

  • Fetch DEF 14A filings by company list

  • Extract SCT tables with SCTAdapter

  • Validate with sct_validator

  • Write results to output/sct

Report Generation

edgar/scripts/create_csv_reports.py converts JSON results into:

  • executive_compensation_<timestamp>.csv

  • top_25_executives_<timestamp>.csv

  • company_summary_<timestamp>.csv

Troubleshooting

  • No filings found: confirm CIK formatting and filing type (DEF 14A vs DEF 14A/A).

  • API errors: slow down requests and confirm user-agent is set.

  • Extraction errors: regenerate code or use manual ground truth in POC scripts.

Related Skills

  • universal/data/reporting-pipelines

  • toolchains/python/testing/pytest

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

Coding

nodejs-backend-typescript

No summary provided by upstream source.

Repository SourceNeeds Review
Coding

jest-typescript

No summary provided by upstream source.

Repository SourceNeeds Review
Coding

github-actions

No summary provided by upstream source.

Repository SourceNeeds Review
Coding

golang-cli-cobra-viper

No summary provided by upstream source.

Repository SourceNeeds Review