Databricks Data Engineer

LED FastStart
3d$88,200 - $214,600Remote

About The Position

Do you want to take your career to the next level? Are you ready for the responsibility of working with high-profile clients? Are you passionate about building and supporting modern data platforms in the cloud? CGI is looking for a highly skilled Senior Databricks Data Engineer to join our team in St. Louis, MO, where you will design, build, and optimize scalable data pipelines and Lakehouse solutions on the Databricks platform. This role is hands-on and engineering-focused, requiring strong Spark expertise, deep Databricks knowledge, and the ability to deliver reliable, high-performance data products that support analytics, reporting, and enterprise decision-making. The ideal candidate excels in modern cloud data engineering, loves solving complex problems, and thrives in fast-paced, collaborative environments. Growth at CGI is driven by your goals, so if you're looking for an inclusive place where you're empowered to chart your own path, then we'd love to meet you. This role is remote and open to candidates living in the United States. We are accepting applications through January 31, 2026.

Requirements

  • 7+ years of experience in data engineering, including 2+ years of direct experience building data pipelines on Databricks using Apache Spark, PySpark and SQL. Strong proficiency with Delta Lake, Lakehouse architecture, and Databricks SQL. Experience managing Databricks clusters, notebooks, jobs, Repos, and Workflows.
  • Data Engineering & ETL Development - Ability to design, build, and maintain scalable ETL pipelines using Databricks and Spark.
  • Cloud Platforms - Hands-on experience with Azure, AWS, or GCP integrations with Databricks. Familiarity with cloud-native storage (ADLS, S3, GCS) and integration with Databricks.
  • Data Modeling & Architecture Foundations - Experience creating data models to support analytics, reporting, and downstream consumption. Understanding of medallion/ Lakehouse data design patterns.
  • Data Governance & Security - Experience implementing governance with Unity Catalog, role-based access control, and data lineage.
  • Performance Optimization- Ability to optimize Spark jobs, cluster configurations, and storage for cost and speed.
  • Demonstrated Experience- Proven track record of successfully delivering multiple complex, enterprise-scale projects leveraging Databricks, including end-to-end solution design, implementation, and optimization.

Responsibilities

  • Build and optimize scalable data pipelines and Lakehouse solutions using Databricks, leveraging core platform capabilities such as Delta Lake, Delta Live Tables, Unity Catalog, and modern data processing patterns. Troubleshoot, debug, and optimize data pipelines, resolving complex issues across Spark jobs, cluster configurations, and cloud integrations.
  • Implement end-to-end data engineering solutions including ingestion, transformation, modeling, and quality frameworks to support analytics and business reporting.
  • Collaborate closely with data architects and cross-functional teams to translate business and technical requirements into robust, production-ready Databricks workflows.
  • Provide technical expertise in Databricks and Apache Spark, contributing to design discussions and ensuring best practices for performance, reliability, and cost optimization.
  • Provide coaching to junior engineers, and promote engineering excellence in coding standards, testing, and automation.
  • Work directly with stakeholders to understand data needs and deliver high-quality solutions that improve decision-making and operational efficiency.

Benefits

  • Competitive compensation
  • Comprehensive insurance options
  • Matching contributions through the 401(k) plan and the share purchase plan
  • Paid time off for vacation, holidays, and sick time
  • Paid parental leave
  • Learning opportunities and tuition assistance
  • Wellness and Well-being programs
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service