Month 1 – Foundations of Data Engineering with Snowflake
Goal: Get comfortable with Snowflake basics, SQL, and core data engineering concepts.
Week 1–2: Snowflake Basics
- Introduction to Snowflake cloud data platform
- Account setup, roles, and access management
- Snowflake architecture: storage, compute (warehouses), services layer
- Loading and unloading data (stages, COPY INTO, file formats, Snowpipe)
- Time Travel & Fail-safe
Hands-on labs:
- Create a Snowflake database and schema
- Load CSV/JSON/Parquet into Snowflake
- Query semi-structured data using VARIANT
Week 3–4: SQL & Data Modeling in Snowflake
- Advanced SQL in Snowflake: CTEs, window functions, clustering, materialized views
- Normalization vs. denormalization
- Star schema, Snowflake schema, and Data Vault modeling
- Partitioning, clustering, and performance tuning basics
- Intro to data governance: masking policies, RBAC
Hands-on labs:
- Create dimension & fact tables in Snowflake
- Optimize queries with clustering keys
- Implement row-level security
Month 2 – dbt Core + dbt Cloud with Snowflake
Goal: Build data transformations with dbt, integrate with version control, and apply testing.
Week 5–6: dbt Foundations
- What is dbt? Role in the modern data stack
- Installing dbt Core & connecting to Snowflake
- dbt project structure (models, seeds, snapshots, macros)
- Writing modular SQL models in dbt
- Jinja templating basics in dbt
Hands-on labs:
- Set up dbt project with Snowflake connection
- Build staging and marts layers in dbt
- Implement snapshots for slowly changing dimensions (SCD Type 2)
Week 7–8: dbt Advanced Concepts
- Testing & documentation in dbt
- dbt sources, exposures, and lineage
- dbt packages (dbt-utils, dbt-expectations)
- Macros, hooks, and operations
- dbt Cloud vs dbt Core: orchestration & deployment options
Hands-on labs:
- Implement unit tests on transformations
- Generate dbt docs and lineage graph
- Use dbt packages to extend transformations
Month 3 – End-to-End Data Engineering Project
Goal: Build, orchestrate, and deploy a full production-grade data pipeline.
Week 9–10: Orchestration + CI/CD
- Scheduling dbt runs with dbt Cloud or Airflow
- CI/CD with GitHub/GitLab for dbt models
- Implementing dbt in a modern data workflow (CI/CD pipelines, pull request reviews)
- Logging, monitoring, and observability with dbt + Snowflake
Hands-on labs:
- Build CI/CD pipeline for dbt project
- Automate nightly dbt runs on Snowflake warehouse
- Integrate alerts for failed jobs
Week 11–12: Capstone Project
- Project Objective:
- Ingest raw data into Snowflake (CSV/JSON/Parquet via Snowpipe or COPY).
- Transform data with dbt into staging → core → analytics marts.
- Apply testing, documentation, and CI/CD pipeline.
- Build a dashboard (e.g., with Tableau or Power BI) connected to Snowflake marts.
- Deliverables:
- Documented dbt project repo (GitHub/GitLab)
- Snowflake schema with marts ready for BI
- Final project presentation/report
🎯 Outcomes After 3 Months
By the end of this course, you will be able to:
- Design data models and schemas in Snowflake.
- Ingest & Transform data with dbt.
- Test, Document, and Deploy data pipelines using CI/CD best practices.
- Implement Governance and security features in Snowflake.
- Build a Real Project showcasing your skills for portfolios or interviews.