Data Engineering
3 crystals in this category
Data engineering is the discipline of designing, building, and maintaining the infrastructure that moves data from source systems to analytical destinations reliably and at scale. Modern data engineers work across the full pipeline lifecycle: ingestion, transformation, quality validation, orchestration, and storage — leveraging tools such as Apache Airflow, dbt, Apache Spark, Apache Kafka, and cloud data warehouses like Snowflake, BigQuery, and ClickHouse. This category provides crystals that accelerate the most time-consuming parts of the daily data engineering workflow: ETL pipeline design, DAG authoring, data quality rule generation, schema migration scripting, data lineage documentation, and dbt model scaffolding. By delegating boilerplate generation to AI, engineers can redirect their attention to architecture decisions and business-logic design rather than repetitive SQL and Python scaffolding. Each crystal in this category is designed for immediate productivity — drop in your table schema or a natural-language description of your pipeline requirements, and receive production-ready code with retry logic, alerting hooks, and unit test stubs included.
Airflow ETL DAG Generator
Generate production-ready Apache Airflow ETL DAG code from a plain-language description of your data sources and transformation logic.
dbt Model Generator
Auto-generate dbt staging, intermediate, and mart model SQL files plus YAML schema definitions from a description of your warehouse tables.
我想求一个skill的晶体
需要非常强大