ddgope/Data-Pipelines-with-Airflow
This project helps me to understand the core concepts of Apache Airflow. I have created custom operators to perform tasks such as staging the data, filling the data warehouse, and running checks on the data quality as the final step. Automate the ETL pipeline and creation of data warehouse using Apache Airflow. Skills include: Using Airflow to automate ETL pipelines using Airflow, Python, Amazon Redshift. Writing custom operators to perform tasks such as staging data, filling the data warehouse, and validation through data quality checks. Transforming data from various sources into a star schema optimized for the analytics team’s use cases. Technologies used: Apache Airflow, S3, Amazon Redshift, Python.
Python
Stargazers
- alimesutk
- aoguedaoGeorge Mason University
- BaekSeSI Analytics
- BobbyAxelrods
- christophercronk
- devgoalposts
- digital-hero
- JamesLauer
- jimbrig@noclocks
- jmhcodes
- kiranwale1989
- LocNguyenHuuST Engineering
- magicwindOriente
- marcolg2404Buk
- mertezztretton37
- Nazareno95NTT DATA
- nelsonic1Halifax, NS
- pawan-shivhare
- prakass1Nordcloud
- r4phael@intelligentagents
- rmania
- shizidushu
- SmendowskiAGH University of Science and Technology
- steveding1
- TaiChiTiger
- trijuhari
- TwilightInSundayThailand
- ungaro
- v-Vahe
- venclov
- WhiskersReneeWeScale AI
- wynnemloHong Kong
- yangrong688
- yaseen221
- zkanBearCamp (  ̄(エ) ̄)
- zzszaSeoul, republic of korea