Domäne
Data Engineering
Skill-Profil
DAGs, operators, sensors, XComs, dynamic task mapping, KubernetesPodOperator
Rollen
4
wo dieser Skill vorkommt
Stufen
5
strukturierter Entwicklungspfad
Pflichtanforderungen
20
die anderen 0 optional
Data Engineering
Data Orchestration
17.3.2026
Wählen Sie Ihr aktuelles Level und vergleichen Sie die Erwartungen.
Die Tabelle zeigt, wie die Tiefe von Junior bis Principal wächst.
| Rolle | Pflicht | Beschreibung |
|---|---|---|
| Analytics Engineer | Pflicht | Understands basic Airflow concepts: DAGs, operators, and task dependencies. Follows established DAG templates to build simple transformation pipelines. Uses dbt + Airflow integration patterns defined by the team. |
| BI Analyst | Pflicht | Understands basic Airflow DAG structure and scheduling concepts. Monitors scheduled report refresh pipelines and identifies failures. Follows team guidelines for triggering dashboard data updates through Airflow UI. |
| Data Analyst | Pflicht | Understands basic Airflow concepts and DAG scheduling. Monitors data pipeline runs that feed analytical datasets. Follows team documentation to trigger ad-hoc DAG runs for data refresh and extraction tasks. |
| Data Engineer | Pflicht | Creates Airflow DAGs: PythonOperator, BashOperator, task dependencies. Understands execution date, catchup, schedule interval. Monitors runs in Airflow UI. Debugs failed tasks through logs. |
| Rolle | Pflicht | Beschreibung |
|---|---|---|
| Analytics Engineer | Pflicht | Independently builds Airflow DAGs for ELT pipelines with dbt operators and data quality checks. Configures retry policies, SLAs, and alerting for transformation jobs. Optimizes task parallelism and resource pools. |
| BI Analyst | Pflicht | Independently configures Airflow DAGs for scheduled report generation and dashboard data refresh. Implements data quality sensors to validate source data before BI layer updates. Troubleshoots pipeline failures affecting reporting. |
| Data Analyst | Pflicht | Independently builds Airflow DAGs for automated data extraction and cohort preparation pipelines. Implements data validation tasks with Great Expectations integration. Configures scheduling for recurring analytical data refreshes. |
| Data Engineer | Pflicht | Designs Airflow DAGs: dynamic task generation, XCom for data passing, TaskGroups for organization. Uses sensors, hooks for external system integration. Configures connections and variables. |
| Rolle | Pflicht | Beschreibung |
|---|---|---|
| Analytics Engineer | Pflicht | Architects data systems with Apache Airflow. Optimizes for big data. Implements data governance and quality frameworks. |
| BI Analyst | Pflicht | Designs Airflow-based data pipeline architecture for enterprise BI platform. Implements complex dependency graphs across multiple data sources with SLA monitoring. Mentors team on DAG design patterns for reporting workflows. |
| Data Analyst | Pflicht | Designs Airflow pipeline architecture for complex analytical workflows with cross-dataset dependencies. Implements data lineage tracking and audit logging. Optimizes DAG performance for large-scale analytical data processing. |
| Data Engineer | Pflicht | Designs Airflow architecture: KubernetesExecutor for dynamic scaling, custom operators/hooks, DAG factory pattern for generation. Optimizes performance: pool management, priority weight, concurrency. |
| Rolle | Pflicht | Beschreibung |
|---|---|---|
| Analytics Engineer | Pflicht | Defines orchestration strategy for the analytics pipeline: Airflow for coordinating dbt runs, sensors for upstream data dependencies. Implements DAG design standards: idempotency, retry policies, SLA monitoring for analytical models. |
| BI Analyst | Pflicht | Defines BI data pipeline strategy and Airflow platform standards. Establishes DAG development guidelines, code review practices, and deployment workflows for reporting team. Coordinates data freshness SLAs with stakeholders. |
| Data Analyst | Pflicht | Defines analytical data pipeline strategy and Airflow governance standards. Establishes DAG naming conventions, testing requirements, and monitoring practices. Drives adoption of self-service pipeline creation among analyst teams. |
| Data Engineer | Pflicht | Defines Airflow standards: DAG structure, naming conventions, testing requirements, deployment workflow. Chooses between Airflow and alternatives (Dagster, Prefect) by scenario. |
| Rolle | Pflicht | Beschreibung |
|---|---|---|
| Analytics Engineer | Pflicht | Architects enterprise analytics platform orchestration: Airflow/Dagster for multi-project dbt, event-driven triggers, cross-team dependency management. Defines migration strategy to managed orchestration (dbt Cloud, Dagster Cloud). |
| BI Analyst | Pflicht | Defines enterprise data orchestration strategy spanning Airflow, dbt, and BI tools. Evaluates orchestration platforms and migration paths. Shapes organizational data delivery standards and cross-team pipeline governance. |
| Data Analyst | Pflicht | Defines enterprise analytical data orchestration strategy. Shapes organizational standards for pipeline reliability and data delivery guarantees. Evaluates next-gen orchestration tools and drives platform evolution decisions. |
| Data Engineer | Pflicht | Designs orchestration strategy: Airflow for batch, event-driven for real-time, hybrid patterns. Defines multi-team governance, shared infrastructure, cost allocation. |