CSV Analyzer
Overview
Comprehensive CSV data analysis and visualization engine. Run the script, then use this guide to interpret results and provide insights to users.
Quick Start
cd ~/.claude/skills/csv-analyzer/scripts export $(grep -v '^#' /path/to/project/.env | xargs 2>/dev/null) python3 analyze_csv.py /path/to/data.csv
Chart Selection Decision Tree
IMPORTANT: Choose charts based on what the user needs to understand:
What is the user trying to understand? │ ├── "What does my data look like?" (Overview) │ └── Run with defaults → overview_dashboard.png │ ├── "Is my data clean?" (Quality) │ └── Check: quality_score, missing_values, duplicates │ └── Show: missing_values.png if problems exist │ ├── "What's the distribution?" (Single Variable) │ ├── Numeric → numeric_distributions.png (histogram + KDE) │ ├── Categorical → categorical_distributions.png (bar chart) │ └── Time-based → time_series.png │ ├── "Are there outliers?" (Anomalies) │ └── box_plots.png → points beyond whiskers are outliers │ ├── "How are variables related?" (Relationships) │ ├── 2 numeric vars → correlation_heatmap.png │ ├── 2-6 numeric vars → pairplot.png (scatter matrix) │ ├── Numeric vs Categorical → violin_plot.png │ └── All numeric → correlation_heatmap.png │ └── "Can I predict X from Y?" (Predictive) └── correlation_heatmap.png → |r| > 0.5 suggests predictive power
How to Interpret Results (For Claude)
Quality Score Interpretation
Score Grade What to Tell User
90-100 A "Your data is excellent quality - ready for analysis"
80-89 B "Good quality data with minor issues worth noting"
70-79 C "Moderate quality - address missing values before critical analysis"
60-69 D "Significant quality issues - recommend data cleaning first"
<60 F "Critical issues - data needs substantial cleaning"
Correlation Interpretation
|r| Value Strength What to Say
0.9 - 1.0 Very Strong "X and Y are very strongly related - almost deterministic"
0.7 - 0.9 Strong "X and Y have a strong relationship - X could help predict Y"
0.5 - 0.7 Moderate "X and Y are moderately correlated - some predictive value"
0.3 - 0.5 Weak "X and Y have a weak relationship - limited predictive power"
0.0 - 0.3 Negligible "X and Y appear unrelated"
Sign matters:
-
Positive: "As X increases, Y tends to increase"
-
Negative: "As X increases, Y tends to decrease"
Skewness Interpretation
Skewness Distribution Shape Recommendation
< -1 Heavy left tail "Most values are high, with some very low outliers"
-1 to -0.5 Mild left skew "Slightly more low outliers than high"
-0.5 to 0.5 Symmetric "Nicely balanced distribution - good for most analyses"
0.5 to 1 Mild right skew "Slightly more high outliers than low"
1 Heavy right tail "Most values are low, with some very high outliers. Consider log transform for modeling."
Outlier Assessment
When reporting outliers:
-
Few outliers (<1%): "A few extreme values that may warrant investigation"
-
Moderate outliers (1-5%): "Notable outliers - check if they're errors or genuine extremes"
-
Many outliers (>5%): "High outlier rate suggests either data issues or a non-normal distribution"
Insight Generation Framework
After running analysis, provide insights in this order:
- Data Overview (Always)
"Your dataset has [rows] records and [cols] columns:
- [n] numeric columns: [list top 3]
- [n] categorical columns: [list top 3]
- Data quality score: [score]/100 ([grade])"
- Key Findings (Pick most relevant)
If quality issues exist:
"I noticed some data quality concerns:
- [X]% missing values in [column] - [recommend: drop/impute/investigate]
- [N] duplicate rows detected - [recommend: keep first/remove all/investigate]"
If strong correlations found:
"Interesting relationships I found:
- [col1] and [col2] are strongly correlated (r=[value]) - [interpretation]
- This suggests [actionable insight]"
If outliers detected:
"I detected outliers in [columns]:
- [column]: [n] values beyond normal range ([min outlier] to [max outlier])
- These could be [data errors / genuine extremes / worth investigating]"
If skewed distributions:
"[Column] has a [right/left]-skewed distribution:
- Most values cluster around [median]
- But there are extreme values up to [max]
- For modeling, consider [log transform / robust methods]"
- Recommendations (Based on findings)
Finding Recommendation
Missing >20% in column "Consider dropping this column or investigating why it's missing"
Missing <5% scattered "Safe to impute with median (numeric) or mode (categorical)"
High correlation (>0.9) "These columns may be redundant - consider keeping only one"
Many outliers "Use robust statistics (median instead of mean) or investigate data collection"
Highly skewed "Apply log transform before linear modeling"
Low quality score "Prioritize data cleaning before analysis"
Multi-Chart Dashboard Requests
When user asks for a "dashboard" or "comprehensive view":
Generate all visualizations
python3 analyze_csv.py data.csv --format html --max-charts 10
Then present charts in this order:
-
overview_dashboard.png - "Here's your data at a glance"
-
correlation_heatmap.png - "Key relationships between variables"
-
numeric_distributions.png - "How your numeric data is distributed"
-
box_plots.png - "Outlier analysis"
-
categorical_distributions.png - "Category breakdowns" (if applicable)
Command Reference
Basic Analysis
python3 analyze_csv.py data.csv
Full Report with All Charts
python3 analyze_csv.py data.csv --format markdown --max-charts 10
Quick Analysis (No Charts)
python3 analyze_csv.py data.csv --no-charts
Large Files (>100MB)
python3 analyze_csv.py huge.csv --sample 50000
Specific Date Columns
python3 analyze_csv.py data.csv --date-columns created_at updated_at
JSON for Programmatic Use
python3 analyze_csv.py data.csv --format json --no-charts
Custom Output Location
python3 analyze_csv.py data.csv --output-dir /path/to/project/.tmp/analysis
Chart Descriptions (For Explaining to Users)
Chart When to Show How to Describe
overview_dashboard.png Always for first look "Here's a bird's eye view of your data"
missing_values.png If missing data exists "This shows where your data has gaps"
numeric_distributions.png When exploring distributions "This shows how your numeric values are spread out"
box_plots.png When checking for outliers "The dots outside the boxes are potential outliers"
correlation_heatmap.png When exploring relationships "Darker colors = stronger relationships"
categorical_distributions.png For category analysis "This shows the breakdown of your categories"
time_series.png For temporal data "Here's how your data changes over time"
pairplot.png For multivariate exploration "Each cell shows how two variables relate"
violin_plot.png Comparing groups "This shows how distributions differ across groups"
Common User Questions → Actions
User Says Action
"Analyze this CSV" Run full analysis, show overview + key insights
"Is my data clean?" Focus on quality_score, missing values, duplicates
"Find patterns" Show correlation_heatmap, highlight strong correlations
"Are there outliers?" Show box_plots, list outlier counts per column
"Compare X across Y" Generate violin_plot for numeric X vs categorical Y
"Show me trends" Generate time_series if datetime column exists
"Create a dashboard" Generate all charts, present organized summary
"What should I clean?" List columns with missing >5%, duplicates, outliers
Output Locations
Charts are saved to:
-
Default: ~/.claude/skills/csv-analyzer/scripts/.tmp/csv_analysis/
-
Custom: Use --output-dir /path/to/project/.tmp/analysis
Always copy charts to user's project .tmp for visibility:
cp ~/.claude/skills/csv-analyzer/scripts/.tmp/csv_analysis/*.png /path/to/project/.tmp/csv_analysis/
Cost
Free - runs entirely locally using pandas, matplotlib, seaborn, scipy.
Dependencies
pip install pandas matplotlib seaborn scipy numpy