Data Engineer — Analytics Infrastructure (Foundational Hire)

Vast.aiLos Angeles, CA
69d$140,000 - $190,000

About The Position

Vast.ai’s cloud powers AI projects and businesses all over the world. We are democratizing and decentralizing AI computing—reshaping our future for the benefit of humanity. We are a growing and highly motivated team dedicated to an ambitious technical plan. Our structure is flat, our ambitions are out‑sized, and leadership is earned by shipping excellence. We seek a data engineer with strong intrinsic drive, a true passion for uncovering insights from data, and a mix of analytical, programming, and communication skills. This is a foundational role: you’ll own the 0→1 build of our data platform—ingestion, modeling, governance, and self‑serve analytics in QuickSight—for Marketing, Sales, Accounting, and leadership. We’re hiring a Data Engineer to build and own the end‑to‑end data platform at Vast.ai. This is a hands‑on role for a builder who can move fast: designing schemas, implementing ELT/ETL, hardening data quality, and enabling secure, governed access to data across the company.

Requirements

  • 3+ years (typically 3–6) in a Data Engineering role building production ELT/ETL on a cloud platform (AWS strongly preferred).
  • Expert SQL and solid Python for data processing/automation.
  • Proven experience designing data models (staging, marts, star schemas) and standing up a warehouse/lakehouse.
  • Orchestration, scheduling, and operational ownership (SLAs, alerting, runbooks).
  • Experience enabling a BI layer (ideally QuickSight) with secure, governed datasets.
  • Strong collaboration and communication; able to gather requirements from non‑technical stakeholders and translate to data contracts.

Nice To Haves

  • Marketing/Sales/RevOps data (CRM, ads, attribution), Accounting/Finance integrations, or product telemetry/event pipelines.
  • Stream processing (Kafka/Kinesis), CDC, or near‑real‑time ingestion.
  • Data privacy/security best practices (e.g., CPRA), partitioning/performance tuning, and cost management on AWS.

Responsibilities

  • Own the data pipeline: design, build, and operate batch/streaming ingestion from product, billing, CRM, support, and marketing/ad platforms into a central warehouse.
  • Model the data: create clean, well‑documented staging and business marts (dimensional/star schemas) that map to the needs of Marketing, Sales, Accounting/Finance, and Operations.
  • Enable: publish certified datasets with row‑/column‑level security, manage refresh SLAs, and make it easy for teams to self‑serve.
  • Collaborate cross‑functionally: intake requirements, translate them into data contracts and models, and partner with Engineering on event/telemetry capture.
  • Document & scale: maintain clear docs, lineage, and a pragmatic data catalog so others can discover and trust the data.

Benefits

  • Comprehensive health, dental, vision, and life insurance
  • 401(k) with company match
  • Meaningful early-stage equity
  • Onsite meals, snacks, and close collaboration with founders/tech leaders
  • Ambitious, fast-paced startup culture where initiative is rewarded
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service