About The Position

The Senior Data Engineer designs, builds, and maintains scalable data pipelines and Lakehouse solutions on the Databricks platform to support enterprise data and AI initiatives. This role partners closely with architects, analytics, and data science teams to deliver high‑quality, reliable data products. The engineer operates with significant autonomy and contributes deep technical expertise across the full data lifecycle.

Requirements

  • 5 Years of experience designing, developing, and deploying data solutions on the Databricks platform.
  • Proficiency in Python, including PySpark, and SQL.
  • Hands‑on experience with Spark, Delta Lake, and Lakehouse architectures.
  • Experience implementing data quality, governance, and security practices across data pipelines.
  • Strong problem‑solving, collaboration, and communication skills.

Nice To Haves

  • Familiarity with machine learning concepts, tools, and libraries such as TensorFlow, PyTorch, Scikit‑learn, and MLflow is a plus.
  • Experience configuring and integrating external AI models and working with AI governance and monitoring tools is a plus.
  • Experience with asynchronous programming patterns in Python for building scalable data or AI workloads is a plus.

Responsibilities

  • Build and maintain scalable data pipelines, ETL and ELT processes, and data models within the Databricks platform.
  • Design, develop, and deploy data and AI solutions using Databricks, Spark, Delta Lake, and related technologies.
  • Develop batch and streaming pipelines using tools such as Databricks Workflows and Azure Data Factory.
  • Design logical data flow diagrams and normalized schemas, implementing Lakehouse patterns such as the Medallion Architecture (Bronze, Silver, Gold layers).
  • Ensure data quality, integrity, security, and governance throughout the data lifecycle, including use of Unity Catalog.
  • Optimize Spark jobs and data transformations through effective partitioning, caching, and join strategies.
  • Monitor pipeline execution, identify failures, and troubleshoot complex data processing issues.
  • Collaborate with data architects, analysts, data scientists, and business stakeholders to understand requirements and deliver solutions.
  • Support documentation of data processes, standards, and data flows.
© 2026 Teal Labs, Inc
Privacy PolicyTerms of Service