As a Senior Data Engineer, you will own the data infrastructure that powers real-time experiences for our members. You will build and scale pipelines that move data from dozens of upstream services [across Kafka event streams and transactional databases] into a unified data platform that serves both real-time APIs and analytical workloads on Databricks.Your work will directly enable AI-powered coaching assistants and physical therapy tools that use live member data [including engagement logs and clinical data] to generate personalized recommendations. You will work at the intersection of data engineering and AI, building the reliable, low-latency data foundation that these systems depend on. You will work in a modern stack: Python, Flink and PySpark for pipeline development, Kafka for event streaming, Delta Lake for scalable storage, and Aurora PostgreSQL for operational data. This is a high-ownership role. You will work closely with application engineers, data scientists, and AI teams across the organization, defining how data flows from the moment it is created to the moment it is consumed. You will also help establish the standards and practices that enable product teams to take ownership of their own data in a HIPAA-compliant environment. If you are excited about building the data infrastructure behind AI systems that have a direct impact on people's health, this role is for you. Our tech stack: Python, SQL, dbt, Airflow, PostgreSQL, MySQL, REST, Aptible, Docker, Tonic.ai, Terraform, Spark, Kafka, Flink, Fivetran, Databricks, AWS (S3, Lambda, Kinesis, RDS, Glue).
Stand Out From the Crowd
Upload your resume and get instant feedback on how well it matches this job.
Job Type
Full-time
Career Level
Senior
Number of Employees
501-1,000 employees