Data Platform Engineer, Data Capture

UpstartSan Mateo, CA
82d

About The Position

The Data Engineering team (Part of Upstart's core platform vertical) provides developer tools, frameworks, and scalable data infrastructure as shared services across Upstart. The team's primary objective is to provide Data Analysts, Software Engineers, and ML scientists access to high-quality data and developer tools to create business metrics for respective product verticals. You will join the subdivision focusing on Data Capture. As the Data Platform Engineer joining this team, you will get the opportunity to contribute to 3 areas: Expand the Developer tools that capture business events, operational data and third party data into the Lakehouse; Expand the self-serve and orchestration framework that developers use to build pipelines; Improve the observability and SLAs for the above services and raw layer data in the Lakehouse. We would love to hear from you if you are passionate about building data products!

Requirements

  • A bachelor's degree in Computer Science, Data Science, Engineering, or a related field.
  • 3+ years of experience in data engineering or related fields, with a strong focus on data quality, governance, and data infrastructure.
  • Proficiency in data engineering tech stack; Databricks / PostgreSQL / Python / Spark / Kafka/ Streaming / SQL / AWS / Airflow/ Airbyte/ Fivetran / DBT / EKS/ containers and orchestration (Docker, Kubernetes) and others.
  • Ability to approach problems with first principles thinking, embrace ambiguity, and enjoy collaborative work on complex solutions.

Nice To Haves

  • Strong foundation in algorithms and data structures and their real-world use cases.
  • Experience and understanding of distributed systems, data architecture design, and big data technologies (e.g., Spark Streaming / Batch, Kafka Streaming, Lakehouse, Databricks, Redshift, Airbyte, Fivetran, EKS).
  • Experience with AWS technologies ( e.g., AWS RDS, S3, Redshift, IAM etc.).
  • Knowledge of data ingestion, software engineering and data security best practices.
  • Good knowledge of DevOps engineering using Continuous Integration/Delivery tools like Kubernetes, Jenkins, Terraform, etc., thinking about automation, alerting, monitoring, security, and other declarative infrastructure.
  • Ability to explain complex concepts in easy-to-understand ways and navigate environments where problems are not well-defined (and evolve quickly).

Responsibilities

  • Build and Improve the tools owned by the Capture team
  • Own the service/raw layer data observability and SLAs
  • Participate in planning and prioritization by collaborating with stakeholders across the various product verticals and functions (ML, Analytics, Finance) to ensure our architecture aligns with the overall business objectives.
  • Collaborate with stakeholders such as Software Engineering, Machine Learning, Machine Learning Platform and Analytics teams to ingest data into the Lakehouse and adopt the developer tools
  • Participate in code reviews and architecture discussions to exchange actionable feedback with peers.
  • Contribute to engineering best practices and mentor junior team members.
  • Help break down complex projects and requirements into sprints.
  • Continuously monitor and improve data platform performance, reliability, and security.
  • Stay up-to-date with emerging technologies and industry best practices in data engineering.

Stand Out From the Crowd

Upload your resume and get instant feedback on how well it matches this job.

Upload and Match Resume

What This Job Offers

Career Level

Mid Level

Industry

Credit Intermediation and Related Activities

Education Level

Bachelor's degree

Number of Employees

1,001-5,000 employees

© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service