Search Results: data-pipelines
Found 32 Skills
data-pipelines
Use this skill for data pipeline work — ingestion with dlt, transformations with sqlmesh, analytics with DuckDB/MotherDuck, DataFrames with polars, notebooks with marimo, and project management with uv.
tensorflow-data-pipelines
Create efficient data pipelines with tf.data
flipside
Query blockchain data across 40+ chains, create AI agents for crypto analytics, and build automated data pipelines. Use when working with blockchain data, crypto wallets, DeFi protocols, NFTs, token transfers, or on-chain analytics. Requires the Flipside CLI (https://docs.flipsidecrypto.xyz/get-started/cli).
senior-data-engineer
Data engineering skill for building scalable data pipelines, ETL/ELT systems, and data infrastructure. Expertise in Python, SQL, Spark, Airflow, dbt, Kafka, and modern data stack. Includes data modeling, pipeline orchestration, data quality, and DataOps. Use when designing data architectures, building data pipelines, optimizing data workflows, implementing data governance, or troubleshooting data issues.
data-engineer
Expert in data pipelines, ETL processes, and data infrastructure
data-engineer
Build scalable data pipelines, modern data warehouses, and real-time streaming architectures. Implements Apache Spark, dbt, Airflow, and cloud-native data platforms. Use PROACTIVELY for data pipeline design, analytics infrastructure, or modern data stack implementation.
kafka-engineer
Expert in Apache Kafka, Event Streaming, and Real-time Data Pipelines. Specializes in Kafka Connect, KSQL, and Schema Registry.
pandas
Expert data analysis and manipulation for customer support operations using pandas
apache-airflow-orchestration
Complete guide for Apache Airflow orchestration including DAGs, operators, sensors, XComs, task dependencies, dynamic workflows, and production deployment
airflow-expert
Expert-level Apache Airflow orchestration, DAGs, operators, sensors, XComs, task dependencies, and scheduling
reporting-pipelines
Reporting pipelines for CSV/JSON/Markdown exports with timestamped outputs, summaries, and post-processing.
data-pipeline-engineer
Expert data engineer for ETL/ELT pipelines, streaming, data warehousing. Activate on: data pipeline, ETL, ELT, data warehouse, Spark, Kafka, Airflow, dbt, data modeling, star schema, streaming data, batch processing, data quality. NOT for: API design (use api-architect), ML training (use ML skills), dashboards (use design skills).