feat(data-engineering): add 6 skills, 2 agents, 1 command for data engineers#216
feat(data-engineering): add 6 skills, 2 agents, 1 command for data engineers#216theondrejivan wants to merge 1 commit intoEveryInc:mainfrom
Conversation
…gineers Add comprehensive data engineering capabilities to make this the best Claude Code plugin for data engineers: Skills (6 new): - dbt: project structure, models, testing, Jinja, incremental, packages - snowflake: SQL patterns, optimization, cost management, Terraform - duckdb: file querying, SQL extensions, Python/dbt integration - databricks: Delta Lake, Unity Catalog, Spark optimization, Terraform - warehouse-architecture: Kimball, Data Vault 2.0, Medallion, SCD types - data-quality: Pandera, Great Expectations, Soda Core, dbt contracts Agents (2 new): - dbt-model-reviewer: SQL quality, ref/source, materialization, testing - data-pipeline-reviewer: idempotency, error handling, credentials Command (1 new): - data-scaffold: scaffold dbt models or dimensional data models Enhanced existing: - performance-oracle: added warehouse SQL optimization section - architecture-strategist: added data warehouse architecture section Bumps version to 2.36.0 (31 agents, 23 commands, 25 skills). Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
|
Added this to cover modern data engineering needs. Feel free to dismiss if it's not the direction you want this plugin to go. |
There was a problem hiding this comment.
💡 Codex Review
Here are some automated review suggestions for this pull request.
Reviewed commit: f692cd7b27
ℹ️ About Codex in GitHub
Your team has set up Codex to review pull requests in this repo. Reviews are triggered when you
- Open a pull request for review
- Mark a draft as ready
- Comment "@codex review".
If Codex has suggestions, it will comment; otherwise it will react with 👍.
Codex can also answer questions or update the PR. Try commenting "@codex address that feedback".
| 5. **Generate source YAML** at `models/staging/<source>/_<source>__sources.yml`: | ||
| ```yaml | ||
| sources: | ||
| - name: <source> | ||
| description: "TODO: Describe the <source> data source" |
There was a problem hiding this comment.
Merge source YAML instead of regenerating it
Step 3 checks for an existing source definition, but step 5 still instructs generating _<source>__sources.yml from a single-table template. In an existing dbt project, running /data-scaffold dbt <source>.<table> for another table in the same source would overwrite prior tables: entries and freshness metadata, which is destructive; this flow should append/merge when the file already exists.
Useful? React with 👍 / 👎.
| 6. **Generate model YAML** at `models/staging/<source>/_<source>__models.yml`: | ||
| ```yaml | ||
| models: | ||
| - name: stg_<source>__<table> | ||
| description: "Cleaned <source> <table> with standardized column names" |
There was a problem hiding this comment.
Append model YAML entries instead of replacing file
Step 6 generates _<source>__models.yml as a fresh document each time, so scaffolding a second model for the same source can erase existing model descriptions and tests in that YAML file. That makes iterative use unsafe in real dbt repos; this should add a new models entry when the file is already present rather than replacing it wholesale.
Useful? React with 👍 / 👎.
| ```markdown | ||
| ```mermaid |
There was a problem hiding this comment.
Close the outer markdown fence in model mode example
The ```markdown fence opened before the Mermaid snippet is never closed, so the remaining model-mode guidance is rendered as code block content instead of normal instructions. This can cause steps 5–7 to be treated as literal sample text rather than executable guidance when the command is followed; close the outer fence (or use nested fencing) to restore correct parsing.
Useful? React with 👍 / 👎.
Summary
dbt,snowflake,duckdb,databricks,warehouse-architecture,data-quality— covering the modern data stack with 21 reference files totaling 10K+ lines of patterns, SQL examples, and best practicesdbt-model-reviewer(SQL quality, ref/source, materialization, testing) anddata-pipeline-reviewer(idempotency, error handling, credential safety)/data-scaffold— scaffold dbt staging models or dimensional data models with ERDsperformance-oracle(+ warehouse SQL optimization for Snowflake/DuckDB/Databricks) andarchitecture-strategist(+ data warehouse architecture: Kimball, SCD, medallion)Bumps version to v2.36.0 (31 agents, 23 commands, 25 skills).
What's Included
dbtsnowflakeduckdbdatabrickswarehouse-architecturedata-qualitySecurity
profiles.ymlexamples always use{{ env_var() }}.gitignoredata-pipeline-reviewerTest Plan
cat plugins/compound-engineering/.claude-plugin/plugin.json | jq .claude /data-scaffold dbt stripe.paymentsdbt-model-reviewer, notperformance-oraclePost-Deploy Monitoring & Validation
No additional operational monitoring required: all changes are additive plugin content (markdown files) with no runtime infrastructure.