Thank you for sending your enquiry! One of our team members will contact you shortly.
Thank you for sending your booking! One of our team members will contact you shortly.
Course Outline
Foundations of Data Warehousing
- Warehouse purpose, components, and architecture
- Data marts, enterprise warehouses, and lakehouse patterns
- OLTP vs OLAP fundamentals and workload separation
Dimensional Modeling
- Facts, dimensions, and grain
- Star schema vs snowflake schema
- Slowly Changing Dimensions types and handling
ETL and ELT Processes
- Extraction strategies from OLTP and APIs
- Transformations, data cleansing, and conformance
- Load patterns, orchestration, and dependency management
Data Quality and Metadata Management
- Data profiling and validation rules
- Master and reference data alignment
- Lineage, catalogs, and documentation
Analytics and Performance
- Cubing concepts, aggregates, and materialized views
- Partitioning, clustering, and indexing for analytics
- Workload management, caching, and query tuning
Security and Governance
- Access control, roles, and row-level security
- Compliance considerations and auditing
- Backup, recovery, and reliability practices
Modern Architectures
- Cloud data warehouses and elasticity
- Streaming ingestion and near real-time analytics
- Cost optimization and monitoring
Capstone: From Source to Star Schema
- Modeling a business process into facts and dimensions
- Building an end-to-end ETL or ELT workflow
- Publishing dashboards and validating metrics
Summary and Next Steps
Requirements
- An understanding of relational databases and SQL
- Experience with data analysis or reporting
- Basic familiarity with cloud or on-premises data platforms
Audience
- Data analysts transitioning to data warehousing
- BI developers and ETL engineers
- Data architects and team leads
35 Hours
Testimonials (3)
I liked that it was practical. Loved to apply the theoretical knowledge with practical examples.
Aurelia-Adriana - Allianz Services Romania
Course - Python and Spark for Big Data (PySpark)
The fact that we were able to take with us most of the information/course/presentation/exercises done, so that we can look over them and perhaps redo what we didint understand first time or improve what we already did.
Raul Mihail Rat - Accenture Industrial SS
Course - Python, Spark, and Hadoop for Big Data
The combination of theory and practice with tools like Databricks
Graciela Saud - Servicio de Impuestos Internos
Course - Spark for Developers
Machine Translated