About The Position

Build and maintain large -scale data pipelines with strict SLAs. Design shared libraries in Scala and Python to standardize data logic. Develop foundational datasets from clickstream and telemetry data. Ensure data quality, reliability, and operational efficiency. Partner with product, engineering, and analytics teams. Define and document data standards and best practices. Participate actively in Agile and Scrum ceremonies. Communicate technical outcomes clearly to business stakeholders. Maintain detailed technical and data governance documentation.

Requirements

  • Advanced SQL expertise
  • Strong Scala development experience
  • Python for data engineering
  • Apache Spark in production
  • Airflow for orchestration
  • Databricks platform experience
  • Cloud data storage experience (S3 or equivalent)
  • 5+ years of data engineering experience.
  • Strong problem -solving and algorithmic skills.
  • Expert -level SQL with complex analytical queries.
  • Hands -on experience with distributed systems at scale.
  • Experience supporting production data platforms.
  • Self -starter who can define outcomes and drive solutions.
  • Ability to translate technical concepts for non -technical audiences.
  • Bachelor’s degree or equivalent experience.

Responsibilities

  • Build and maintain large -scale data pipelines with strict SLAs.
  • Design shared libraries in Scala and Python to standardize data logic.
  • Develop foundational datasets from clickstream and telemetry data.
  • Ensure data quality, reliability, and operational efficiency.
  • Partner with product, engineering, and analytics teams.
  • Define and document data standards and best practices.
  • Participate actively in Agile and Scrum ceremonies.
  • Communicate technical outcomes clearly to business stakeholders.
  • Maintain detailed technical and data governance documentation.
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service