About The Position

Build the data foundation that powers Auger’s Supply Chain OS, AI systems, and execution workflows. Auger is building an operating system for supply chain teams. Our customers rely on Auger to understand reality and change it: reporting, AI-powered decision support, and write-back execution systems that operate at scale. This role is data-centric software engineering at a very high bar: you own substantial parts of the transformation layer that turns messy, customer-shared data into a unified, production-grade ontology powering analytics, AI workflows, and execution systems. This is not a “move data from A to B” role. You are expected to own semantic correctness, operability, and durability for the systems you touch—while working within platform standards and influencing improvements across the team.

Requirements

  • Degree in Computer Science, Mathematics, Statistics, or another data-intensive discipline, with substantive principal-level engineering experience.
  • 10+ years of professional development experience, including 8+ years hands-on across SQL and Python (strong familiarity with at least one large-scale batch engine such as Spark; Scala/Flink/Beam a plus).
  • 8+ years in data management (structured and semi-structured), modern warehouses/lakehouses, and schema design in complex domains — with proven production ownership at scale: on-call, incidents, lakehouse architecture, ETL hardening, and long-term reliability improvements.
  • Agent-native fluency: AI-assisted development for data work with rigorous validation — you know when generated SQL or pipelines are wrong and how to prove they are right.
  • Experience defining data quality, observability, anomaly detection, or reliability standards and getting teams to adopt them.
  • Technical leadership: lead through ambiguity, set direction for frameworks and conventions, mentor others, and communicate clearly with technical and non-technical partners — including crisp written technical strategy when stakes are high.

Nice To Haves

  • A plus if you have prior experience in supply chain, planning, or fulfillment domains.

Responsibilities

  • Own the evolution of the data lifecycle from ingestion through production-ready ontology, including standards for medallion-style lakehouse pipelines, schema contracts, and clear promotion paths from development to production.
  • Practice test-driven engineering for data: define what "correct" means for critical datasets, encode that in tests and checks, and use failures to drive fixes and systemic improvements—not one-off patches.
  • Shape technical direction for the data platform: boundaries between layers, evolution strategies, and correctness guarantees downstream systems can rely on.
  • Drive pragmatic data governance engineers will adopt: producer/consumer contracts, schema evolution norms, ownership expectations.
  • Represent data engineering in architecture and design discussions; make tradeoffs explicit among freshness, cost, complexity, and risk, and document decisions others can execute.
  • Institutionalize observability and reliability engineering for data: SLIs/SLOs where appropriate, monitoring, incident response, post-incident learning, backfill/replay strategy, and elimination of recurring failure modes.
  • Design and implement reusable, agentic frameworks across heterogeneous customer data sources so the team can rapidly discover schemas and semantics, draft transformation logic, generate and iterate on SQL, and run end-to-end troubleshooting workflows in a consistent, auditable way.
  • Mentor engineers on AI-native discipline and operational rigor, and set expectations for quality bars on AI-assisted output.
  • Partner across product, science, and data-tools platform to translate ambiguous needs into durable designs—aligning semantics, delivery, and operability with what customers experience in the product.
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service