Snowflake SnowPro Advanced – Data Engineer

Course Summary
The SnowPro Advanced: Data Engineer certification is designed for professionals responsible for developing and maintaining data pipelines, transforming raw data into trusted assets, and implementing engineering best practices on the Snowflake Data Cloud. This training goes beyond SnowPro Core to focus on advanced ETL/ELT design, automation, orchestration, data modeling, and performance tuning — essential for building scalable and efficient data engineering workflows in Snowflake.
This course prepares participants for real-world responsibilities as well as the SnowPro Advanced: Data Engineer certification exam, positioning them as specialists in modern data engineering using Snowflake.
Why Choose This Course
As organizations shift to cloud-native data platforms, skilled Snowflake data engineers are in high demand to build pipelines that are reliable, high-performing, and cost-efficient. This course:
Focuses on real-time and batch data engineering practices
Equips learners to design enterprise-scale ETL/ELT pipelines on Snowflake
Covers performance optimization, automation, and DevOps integration
Includes advanced use of SQL, Snowpipe, Streams, Tasks, and third-party orchestration tools
Prepares participants to take and pass the SnowPro Advanced: Data Engineer exam
Led by certified trainers with hands-on Snowflake engineering experience, the course is rooted in practical application and industry relevance.
Who Should Enroll
This course is ideal for:
Experienced data engineers working on Snowflake or other cloud data platforms
ETL/ELT developers looking to upskill for the cloud
Professionals preparing for the SnowPro Advanced: Data Engineer certification
Cloud engineers or DevOps practitioners supporting data pipeline infrastructure
Data professionals seeking to deepen their SQL automation, scheduling, and optimization skills
Skills You Will Gain
Designing and managing complex ELT pipelines in Snowflake
Advanced use of Streams, Tasks, and Snowpipe for real-time ingestion
Handling semi-structured data (JSON, Parquet, Avro) and variant types
Performance tuning using clustering, caching, and query optimization
Automating workflows using tools like Airflow, dbt, or CI/CD pipelines
Monitoring, troubleshooting, and optimizing data flows across workloads
Career Benefits
After completing this training, you will:
Be ready to pass the SnowPro Advanced: Data Engineer certification exam
Qualify for roles like Senior Data Engineer, Snowflake Data Engineer, or Cloud ETL Specialist
Gain credibility in managing large-scale data pipelines for analytics, AI/ML, and business intelligence
Be prepared to work in multi-cloud, cross-platform data environments with Snowflake at the center
Build a skill set that aligns with the modern data stack used in enterprises globally
Engineer Smarter Data Pipelines with Snowflake
Data engineers power the future of analytics and AI. This course empowers you to build resilient, efficient, and high-performance data pipelines using Snowflake’s cloud-native features.
Get certified. Get noticed. Get ahead.
Snowflake SnowPro Advanced - Data Engineer Syllabus
Modules
Data Movement - 28%
- Given a data set, load data into Snowflake.
Outline considerations for data loading
Define data loading features and potential impact
- Ingest data of various formats through the mechanics of Snowflake.
Required data formats
Outline Stages
- Troubleshoot data ingestion.
Identify causes of ingestion errors
Determine resolutions for ingestion errors
- Design, build and troubleshoot continuous data pipelines.
Stages
Tasks
Streams
Snowpipe (for example, Auto ingest as compared to Rest API)
- Analyze and differentiate types of data pipelines.
Create User-Defined Functions (UDFs) and stored procedures including Snowpark
Design and use the Snowflake SQL API
- Install, configure, and use connectors to connect to Snowflake.
- Design and build data sharing solutions.
Implement a data share
Create a secure view
Implement row level filtering
- Outline when to use External Tables and define how they work.
Partitioning external tables
Materialized views
Partitioned data unloading
Performance Optimization - 22%
- Troubleshoot underperforming queries.
Identify underperforming queries
Outline telemetry around the operation
Increase efficiency
Identify the root cause
- Given a scenario, configure a solution for the best performance.
Scale out as compared to scale up
Virtual warehouse properties (for example, size, multi-cluster)
Query complexity
Micro-partitions and the impact of clustering
Materialized views
Search optimization service
Query acceleration service
- Outline and use caching features.
- Monitor continuous data pipelines.
Snowpipe
Tasks
Streams
Storage and Data Protection - 10%
- Implement data recovery features in Snowflake.
Time Travel
Fail-safe
- Outline the impact of Streams on Time Travel.
- Use System Functions to analyze Micro-partitions.
Clustering depth
Cluster keys
- Use Time Travel and Cloning to create new development environments.
Clone objects
Validate changes before promoting
Rollback changes
Security - 10%
- Outline Snowflake security principles.
Authentication methods (Single Sign-On (SSO), Key pair Authentication, Username/Password, Multi-factor Authentication (MFA))
Role Based Access Control (RBAC)
Column Level Security and how data masking works with RBAC to secure sensitive data
- Outline the system defined roles and when they should be applied.
The purpose of each of the system defined roles including best practices usage in each case
The primary differences between SECURITYADMIN and USERADMIN roles
The difference between the purpose and usage of the USERADMIN/SECURITYADMIN roles and SYSADMIN
- Manage Data Governance.
Explain the options available to support column level security including Dynamic Data Masking and External Tokenization
Explain the options available to support row level security using Snowflake Row Access Policies
Use DDL required to manage Dynamic Data Masking and Row Access Policies
Use methods and best practices for creating and applying masking policies on data
Use methods and best practices for Object Tagging
Data Transformation
- Define User-Defined Functions (UDFs) and outline how to use them.
Snowpark UDFs (for example, Java, Python, Scala)
Secure UDFs
SQL UDFs
JavaScript UDFs
User-Defined Table Functions (UDTFs)
- Define and create External Functions.
Secure external functions
Integration requirements
- Design, build, and leverage Stored Procedures.
Snowpark stored procedures (for example, Java, Python, Scala)
SQL Scripting stored procedures
JavaScript stored procedures
Transaction management
- Handle and transform semi-structured data.
Traverse and transform semi-structured data to structured data
Transform structured data to semi-structured data
Understand how to work with unstructured data
- Use Snowpark for data transformation.
Understand Snowpark architecture
Query and filter data using the Snowpark library
Perform data transformations using Snowpark (for example, aggregations)
Manipulate Snowpark DataFrames