Data engineering practice, including building data pipelines (ELT) from a variety of sources.
-
Updated
Feb 13, 2023 - Python
Data engineering practice, including building data pipelines (ELT) from a variety of sources.
Building an ETL pipeline that extracts data from S3, stages them in Redshift.
Building a next-generation hybrid data pipeline architecture that combines the power of Microsoft Fabric, Azure Cloud, and Power BI. This pipeline is engineered to tackle the challenges of real-time data ingestion, multi-layered processing, and analytics, delivering business-critical insights.
Summary/Notes of Snowflake cloud data warehouse. (Complete ✅)
moved, cleaned, and transformed data stored in S3 as json to Redshift.
Building a cloud data warehouse with AWS Redshift.
This project builds a cloud-based ETL pipeline for Sparkify to move data to a cloud data warehouse. It extracts song and user activity data from AWS S3, stages it in Redshift, and transforms it into a star-schema data model with fact and dimension tables, enabling efficient querying to answer business questions.
Add a description, image, and links to the cloud-data-warehouse topic page so that developers can more easily learn about it.
To associate your repository with the cloud-data-warehouse topic, visit your repo's landing page and select "manage topics."