Senior Data Engineer, Engineering

WorkWhileSan Francisco, CA
6dRemote

About The Position

WorkWhile is seeking a Senior Data Engineer to join the Engineering team and lead the development and operation of our data platform powering analytics, ML models, and operational intelligence across the WorkWhile staffing marketplace. In this role, you will be responsible for designing, building, and maintaining data pipelines and infrastructure on Google Cloud Platform that enable our Data Science team to train and deploy models at scale - from demand forecasting and worker reliability prediction to dynamic pricing and retention modelling. You will own and evolve our Dagster-based orchestration platform, BigQuery data warehouse, and the integrations that connect our PostgreSQL production database to our analytical layer. You will work closely with Data Science, Product Engineering, and Operations to ensure reliable, well-modelled, high-quality data is available where and when it’s needed. Your work will directly improve how WorkWhile matches workers with opportunities, predicts marketplace dynamics, and delivers a better experience for workers and businesses alike. Expect to work in a fast-paced, collaborative environment where your contributions directly support WorkWhile’s mission to make hourly work more accessible, reliable, and rewarding for everyone.

Requirements

  • Bachelor’s degree in Computer Science, Data Engineering, Mathematics, or equivalent experience
  • 5+ years in data engineering or data platform engineering
  • Experience with Dagster or similar orchestration tools (Airflow, Prefect)
  • Expertise in SQL, with the ability to write and optimize complex analytical queries across BigQuery and PostgreSQL
  • Proficiency building data pipelines in Python
  • Experience maintaining data warehouses on BigQuery, Snowflake, or Redshift
  • Hands-on experience with Google Cloud Platform services
  • Familiarity with ML workflows and the ability to collaborate with Data Scientists on feature engineering, training pipelines, and model serving
  • Experience with infrastructure-as-code (Terraform) and containerized deployments (Docker, ECS, Cloud Run, Kubernetes, etc.)
  • Proficiency with data quality frameworks, monitoring, and observability tooling
  • Strong collaboration skills and a track record of partnering effectively with Data Science and Product Engineering teams
  • Passion for building reliable, well-tested data systems - you care about code quality (linting, type checking, CI) as much as pipeline uptime

Nice To Haves

  • Experience with reverse ETL tools like Hightouch or Census
  • Familiarity with BI tools such as Metabase or Looker
  • Background in marketplace, staffing, or gig economy data domains

Responsibilities

  • Build and optimize data pipelines that ingest, transform, and model data from PostgreSQL, Amplitude, and external sources into BigQuery
  • Own BigQuery data warehouse architecture: dataset organization, table design, partitioning, clustering, and query performance optimization
  • Work to improve Ops ML platform capabilities and processes, partnering with the Data Science team to support efficient and reliable ML training and pipelines
  • Work on reverse ETL workflows and API integrations that push model predictions back into production systems
  • Support analytics by ensuring clean, performant datasets are available for self-serve reporting
  • Collaborate with Engineering on Terraform-managed GCP infrastructure
  • Optimize Cloud Tasks and Cloud Scheduler configurations for data refresh jobs and materialized view maintenance

Benefits

  • Remote-friendly work culture with office hubs in SF, NY, Seattle & Toronto
  • In-person company offsites
  • Medical, dental, & vision coverage
  • Flexible time off
  • 401(k) with employer match
  • WFH stipend to support your home office needs
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service