← All Programs
💻

Data Engineering & ETL Pipelines

Build production data pipelines with Airflow, dbt, and Spark. Extract data from APIs and files, transform it, and load into data warehouses.

20 Projects to Choose From42 Days DurationVirtual / RemoteVerified Certificate
Already registered? Log In

Sample Projects

01

Airflow ETL DAG

Fetch weather API daily. Transform. Load to PostgreSQL. Email on failure.

PythonApache AirflowPostgreSQLdbt
Intermediate
02

dbt Sales Models

Staging → intermediate → mart models for e-commerce sales data. All tested.

PythonApache AirflowPostgreSQLdbt
Intermediate
03

dbt Documentation

Full schema.yml with descriptions + tests. Docs site generated and shared.

PythonApache AirflowPostgreSQLdbt
Intermediate
04

PySpark Data Processing

Process 1M rows with PySpark. Compare performance to Pandas.

PythonApache AirflowPostgreSQLdbt
Intermediate
05

Spark SQL Analytics

5 complex queries with joins + window functions on a large dataset.

PythonApache AirflowPostgreSQLdbt
Intermediate

+15 more projects available after enrollment

What You'll Get

  • Personalised 4-week roadmap PDF with daily tasks
  • Step-by-step implementation guide for your chosen project
  • Curated tools, libraries, and learning resources
  • Submission and evaluation criteria
  • Verified certificate with QR code on completion
  • LinkedIn-shareable certificate
Enroll Now

Build a real project in 4 weeks


  • 📅 42-day program
  • 🏠 100% Virtual / Remote
  • 📁 20 Projects to choose from
  • 🏆 Verified certificate on completion
  • 📄 Personalised roadmap PDF