AI assistant skills for building workflows and data pipelines using Dagster.
Compatible with Claude Code, OpenCode, OpenAI Codex, Pi, and other Agent Skills-compatible tools.
Install using the Claude plugin marketplace:
/plugin marketplace add dagster-io/skills
Install using the npx skills command-line:
npx skills add dagster-io/skillsSee full instructions...
Clone the repository and copy skills to your tool's skills directory:
OpenCode:
git clone https://github.com/dagster-io/skills.git
cp -r skills/skills/* ~/.config/opencode/skill/OpenAI Codex:
git clone https://github.com/dagster-io/skills.git
cp -r skills/skills/* ~/.codex/skills/Pi Agent:
git clone https://github.com/dagster-io/skills.git
cp -r skills/skills/* ~/.pi/agent/skills/Expert guidance for building production-quality Dagster projects, covering CLI commands, asset patterns, automation strategies, and implementation workflows.
What you can do:
- Create and scaffold projects, assets, schedules, and sensors
- Understand asset patterns (dependencies, partitions, multi-assets, metadata)
- Implement automation (declarative automation, schedules, sensors)
- Use CLI commands (launch, list, check, scaffold, logs)
- Design project structure and configure environments
- Follow implementation workflows and best practices
- Debug issues and validate project configuration
Example prompts:
Create a new Dagster project called analytics
How do I scaffold a new asset?
Show me how to set up declarative automation
What's the proper way to partition my assets?
Help me debug why my materialization failed
How should I structure my project for multiple pipelines?
Launch all assets tagged with priority=high
Comprehensive catalog of 82+ Dagster integrations organized by category.
What's included:
- AI & ML: OpenAI, Anthropic, Gemini, MLflow, W&B
- ETL/ELT: dbt, Fivetran, Airbyte, dlt, Sling, PySpark
- Storage: Snowflake, BigQuery, Postgres, S3, DuckDB, Weaviate
- Compute: AWS, Azure, GCP, Databricks, Spark, Kubernetes
- BI: Looker, Tableau, PowerBI, Sigma, Hex
- Monitoring: Datadog, Prometheus, Papertrail
- Alerting: Slack, PagerDuty, MS Teams, Discord, Twilio
- Testing: Great Expectations, Pandera
- Other: Pandas, Polars
Example questions:
Which tool should I use for data warehousing?
Does Dagster support dbt?
How do I compare Snowflake vs BigQuery?
What integrations are available for ML?
Production-quality Python coding standards for modern Python.
Use for general Python code quality, not Dagster-specific patterns.
What's included:
- Modern type syntax (list[str], str | None)
- LBYL exception handling patterns
- Pathlib operations
- Python version-specific features (3.10-3.13)
- CLI patterns (Click, argparse)
- Advanced typing patterns
- Interface design (ABC, Protocol)
- API design principles
Example questions:
Is this good Python code?
How should I annotate this function?
What's the difference between LBYL and EAFP?
Should I use pathlib or os.path?
See CONTRIBUTING.md.
