Data Analyst Intelligence for AI IDEs
An AI skill that provides structured, professional data analysis workflows with built-in validation - helping AI coding assistants perform data analysis like a careful human analyst.
CrushData AI provides:
- 10 Analysis Workflows - EDA, Dashboard, A/B Test, Cohort, Funnel, Time Series, Segmentation, Data Cleaning, Ad-hoc, KPI Reporting
- 400+ Searchable Patterns - Metrics, SQL, Python, Charts, Database Tips, Common Mistakes
- Context-Building Protocol - Forces AI to ask questions and validate before delivering results
- 4 Industry Modules - SaaS, E-commerce, Finance, Marketing specific metrics
npm install -g crushdataThe -g flag means Global Install:
Local Install (npm install) |
Global Install (npm install -g) |
|
|---|---|---|
| Location | ./node_modules/ in current folder |
System-wide (e.g., %APPDATA%\npm\) |
| Scope | Only available in that project | Available everywhere on your computer |
| Use Case | Libraries for your project | CLI tools you want to run anywhere |
Then in any project:
cd your-project
crushdata init --ai all # All AI IDEs
crushdata init --ai claude # Claude Code onlyWhen you run crushdata init --ai all, the CLI:
-
Creates
.shared/data-analyst/- Contains the BM25 search engine and 13 CSV knowledge databases (~400 rows of data analyst patterns) -
Creates AI IDE config files based on
--aiflag:Flag Creates --ai claude.claude/skills/data-analyst/SKILL.md--ai cursor.cursor/commands/data-analyst.md--ai windsurf.windsurf/workflows/data-analyst.md--ai antigravity.agent/workflows/data-analyst.md--ai copilot.github/prompts/data-analyst.prompt.md--ai kiro.kiro/steering/data-analyst.md--ai allAll of the above -
Your AI IDE automatically detects the config files and enables the
/data-analystcommand
To update the CLI and refresh your project's AI skill files:
npm install -g crushdata@latest
# Update specific IDE (recommended):
crushdata init --ai cursor --force
# Or update everything:
crushdata init --forceCrushData AI now features a Connection Manager to securely handle your data credentials.
Run the connect command to open the management UI:
crushdata connect- Supported Types: CSV, MySQL, PostgreSQL, Shopify, BigQuery, Snowflake
- Private & Secure: Credentials are stored locally on your machine (
~/.crushdata/connections.json). They are never uploaded to any server or included in the npm package.
Note
Persistence: Once you add a connection, you can close the UI (Ctrl+C). The AI IDE reads the saved connection details directly from your local config file, so the server does NOT need to keep running.
crushdata connectionsCrushData AI generates interactive dashboards to visualize your analysis results.
Run the dashboard command to open the local React-based viewer:
# Using installed package (generally faster)
crushdata dashboard
# OR using npx (if not in PATH)
npx crushdata dashboard
Advanced charts visualization (Funnel, Gauge, Radar, etc.)
Standard charts visualization (Line, Bar, Pie, etc.)
- Tier 1 Charts: Line, Bar, Pie, Area, Scatter, Radar (via Recharts)
- Tier 2 Charts: Funnel, Gauge, Heatmap, Sankey, Treemap, Waterfall (via Plotly)
- Auto-Refresh: The dashboard automatically updates when your AI agent writes new data to
reports/dashboards/. - Data Refresh: Use the "Refresh" button π on any chart to re-run the saved SQL/Python query against your data source.
When you ask an AI agent (like Claude or Cursor) to "create a dashboard", it follows this process:
- Analyzes Data: The AI runs SQL/Python to calculate metrics and aggregates.
- Generates JSON: It creates a file at
reports/dashboards/your-topic.jsonusing the CrushData schema. - Visualizes: You run the dashboard command to see the rendered charts instantly.
The AI automatically selects the best chart type (e.g., Line for trends, Bar for comparisons) based on your data.
crushdata init --ai allThe skill activates automatically (Claude) or via slash command (others).
Example Workflow:
- User Request: "Analyze the sales trends in
my-shop-data" - AI Action: The AI checks your saved connections.
- AI Action: The AI runs:
npx crushdata snippet my-shop-data --lang python
- Result: The AI receives the secure code to connect to your data (read-only) and proceeds with analysis.
The skill activates automatically when you request data analysis work. Just chat naturally:
Analyze customer churn for my SaaS product
Use the slash command to invoke the skill:
/data-analyst Analyze customer churn for my SaaS product
Type / in chat to see available commands, then select data-analyst:
/data-analyst Analyze customer churn for my SaaS product
In VS Code with Copilot, type / in chat to see available prompts, then select data-analyst:
/data-analyst Analyze customer churn for my SaaS product
Analyze customer churn for my SaaS product
Create a dashboard for e-commerce analytics
Calculate MRR and ARR from subscription data
Build a cohort retention analysis
Perform A/B test analysis on conversion rates
# Search workflows
python3 .shared/data-analyst/scripts/search.py "EDA" --domain workflow
# Search metrics
python3 .shared/data-analyst/scripts/search.py "churn" --domain metric
# Search SQL patterns
python3 .shared/data-analyst/scripts/search.py "cohort" --domain sql
# Industry-specific
python3 .shared/data-analyst/scripts/search.py "MRR" --industry saas| Domain | Content |
|---|---|
workflow |
Step-by-step analysis processes |
metric |
Metric definitions with formulas |
chart |
Visualization recommendations |
cleaning |
Data quality patterns |
sql |
SQL patterns (window functions, cohorts) |
python |
pandas/polars code snippets |
database |
PostgreSQL, BigQuery, Snowflake tips |
report |
Dashboard UX guidelines |
validation |
Common mistakes to avoid |
| Industry | Key Metrics |
|---|---|
saas |
MRR, ARR, Churn, CAC, LTV, NRR |
ecommerce |
Conversion, AOV, Cart Abandonment |
finance |
Margins, ROI, Cash Flow, Ratios |
marketing |
CTR, CPA, ROAS, Lead Conversion |
- Discovery - AI asks about business context before coding
- Data Profiling - Mandatory checks before analysis
- Data Cleaning (ETL) - Handle missing values/duplicates in
etl/folder - Validation - Verify JOINs, aggregations, and totals
- Sanity Checks - Compare to benchmarks before delivery
To prevent global conflicts, the AI is instructed to:
- Check: Look for existing
venvor.venv. - Create: If missing, run
python3 -m venv venv. - Reports: Save all validation/profiling outputs to
reports/folder. Create if missing.
This prevents the common AI mistakes:
- β Wrong metric definitions
- β Duplicate row inflation
- β Incorrect JOIN types
- β Unreasonable totals
- β Cluttered workspaces (scripts are organized in
analysis/andetl/)
Apache 2.0

