Senior Staff Engineer, Data Infrastructure

ArcherSan Jose, CA
2hHybrid

About The Position

Archer is an aerospace company based in San Jose, California building an all-electric vertical takeoff and landing aircraft with a mission to advance the benefits of sustainable air mobility. We are designing, manufacturing, and operating an all-electric aircraft that can carry four passengers while producing minimal noise. Our sights are set high and our problems are hard, and we believe that diversity in the workplace is what makes us smarter, drives better insights, and will ultimately lift us all to success. We are dedicated to cultivating an equitable and inclusive environment that embraces our differences, and supports and celebrates all of our team members. Senior Staff Engineer, Data Infrastructure (Hybrid-San Jose, CA) The Mission We are looking for a heavy-hitter to build the "Data Backbone" of our company. You will be responsible for the architecture, scaling, and reliability of the infrastructure that powers our Data Engineering and ML teams. Your goal is to provide a seamless, self-service environment where data scientists can go from a JupyterHub notebook to a massive Ray cluster or Trino query without worrying about the underlying hardware.

Requirements

  • The "Data-Aware" Engineer: You understand that scaling a database or a Ray cluster is different from scaling a stateless API. You know how to handle persistent volumes and data gravity.
  • Senior Leadership: You’ve spent time in the trenches. You’ve been on-call for 2:00 AM outages and have built the automation to ensure those outages never happen twice.
  • Tooling Polyglot: You don't just use tools; you contribute to them. You are comfortable writing Go or Python to extend Kubernetes Operators or automate data workflows.
  • Self-Directed: You thrive in ambiguity. You can take a high-level requirement ("Make Trino faster") and turn it into a multi-week infrastructure roadmap.
  • Bachelor's degree in Computer Science, Engineering, or a related field, or equivalent practical experience.

Responsibilities

  • Data Plane Ownership: Architect and manage the lifecycle of high-throughput data tools including Trino, Ray, and JupyterHub on Kubernetes.
  • GitOps & Automation: Drive a "zero-manual-touch" philosophy using ArgoCD and Terraform to manage complex, stateful data environments.
  • Observability at Scale: Build high-cardinality monitoring systems using VictoriaMetrics and Vector to track pipeline health, data ingestion rates, and system performance.
  • ML Lifecycle Support: Maintain and optimize MLflow for model tracking, ensuring it integrates deeply with our compute and storage layers.
  • Engineering Sovereignty: As a self-starter, you will identify performance bottlenecks in data processing and proactively implement infrastructure-level optimizations.
  • Reliability: Participate in on-call rotations for the data stack, treating "data downtime" with the same urgency as a site outage.
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service