Design and implement Apache Airflow scheduling and dependency framework
Tune performance and benchmark against on-prem solution
Use Jira for task tracking and GitHub for code management, reviews, and pull requests
Experience working with Oracle RDS
Experience working with one or more AWS Cloud services such as S3 storage, Managed Airflow (MWAA), Data Migration Service (DMS) in support of building data pipelines
Experience working with a variety of backend data sources
Advanced SQL coding skills and ability to translate Oracle PL SQL and Stored Procedure code to alternative SQL platforms such as Snowflake
Familiarity with data warehouse and data mart concepts such as normalization, facts, dimensions, slowly changing dimensions
Familiarity with Change Data Capture (CDC) concepts and implementations
Understanding of common file formats such as JSON, XML, CSV
Basic experience using scripting tools to automate tasks (e.g.: Python, Windows PowerShell, bash)
Ability to write unit test scripts and be able to validate migrated ELT/ETL code