Choosing Capgemini means choosing a company where you will be empowered to shape your career in the way you’d like, where you’ll be supported and inspired by a collaborative community of colleagues around the world, and where you’ll be able to reimagine what’s possible. Join us and help the world’s leading organizations unlock the value of technology and build a more sustainable, more inclusive world.Your Role: Design, develop, and deploy Apache Airflow DAGs to orchestrate Databricks jobs.Build and maintain ETL workflows to extract, transform, and load data. Optimize Airflow DAGs for performance, scalability, and reliability. Monitor Airflow and Databricks job runs, and troubleshoot execution failures. Refactor and maintain reusable workflow components in python. Collaborate with platform, DevOps, and other data engineering teams. Skills and experience Hands-on experience with Apache Airflow 2.0, Databricks, PySpark, and SparkSQL. Strong Python programming skills. Experience designing and developing ETL processes and data pipelines in Databricks. Solid understanding of scheduling, dependencies, and error handling in Airflow.
Stand Out From the Crowd
Upload your resume and get instant feedback on how well it matches this job.
Job Type
Full-time
Career Level
Mid Level
Education Level
No Education Listed