Senior Data Engineer (Streaming & APIs)

Umanist StaffingGlendale, CA
Onsite

About The Position

Job Title: Sr Data Engineer Location: Glendale, CA – Onsite 4 days a week Platform / Stack You will work with technologies that include Python, AWS, Spark, Snowflake, Databricks, and Airflow. What You'll Do As a Sr Data Engineer: Contribute to maintaining, updating, and expanding existing Core Data platform data pipelines Build and maintain APIs to expose data to downstream applications Develop real-time streaming data pipelines Tech stack includes Airflow, Spark, Databricks, Delta Lake, and Snowflake Collaborate with product managers, architects, and other engineers to drive the success of the Core Data platform Contribute to developing and documenting both internal and external standards and best practices for pipeline configurations, naming conventions, and more Ensure high operational efficiency and quality of the Core Data platform datasets to ensure our solutions meet SLAs and project reliability and accuracy to all our stakeholders (Engineering, Data Science, Operations, and Analytics teams)

Requirements

  • 5+ years of data engineering experience developing large data pipelines
  • Proficiency in at least one major programming language (e.g. Python, Java, Scala)
  • Hands-on production environment experience with distributed processing systems such as Spark
  • Hands-on production experience with data pipeline orchestration systems such as Airflow for creating and maintaining data pipelines
  • Experience with at least one major Massively Parallel Processing (MPP) or cloud database technology (Snowflake, Databricks, Big Query).
  • Experience in developing APIs with GraphQL
  • Advanced understanding of OLTP vs OLAP environments
  • Strong background in at least one of the following: distributed data processing or software engineering of data services, or data modeling.

Responsibilities

  • Contribute to maintaining, updating, and expanding existing Core Data platform data pipelines
  • Build and maintain APIs to expose data to downstream applications
  • Develop real-time streaming data pipelines
  • Collaborate with product managers, architects, and other engineers to drive the success of the Core Data platform
  • Contribute to developing and documenting both internal and external standards and best practices for pipeline configurations, naming conventions, and more
  • Ensure high operational efficiency and quality of the Core Data platform datasets to ensure our solutions meet SLAs and project reliability and accuracy to all our stakeholders (Engineering, Data Science, Operations, and Analytics teams)
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service