Data Infrastructure

Roboflow
1dRemote

About The Position

Our mission is to make the world programmable. Sight is one of the key ways we understand the world, and soon this will be true for the software we use, too. We’re building the tools, community, and resources needed to make the world programmable with artificial intelligence. Roboflow simplifies building and using computer vision models. Today, over 1M+ developers, including those from half the Fortune 100, use Roboflow’s machine learning open source and hosted tools. That includes counting cells to accelerate cancer research, improving construction site safety, digitizing floor plans, preserving coral reef populations, guiding drone flight, and much more. Our team is small relative to our impact, and we believe our user success is our success (not the inverse). A team member summarized: “Roboflow is a company full of giant brains and tiny egos.” We find software has a multiplier effect on all roles (not only product and engineering), so Roboflow employs developers across the company in design, sales, customer support, marketing, and beyond. We’re supported by great customers and investors, having raised over 63 million from Google Ventures, Y Combinator, Craft Ventures, Sam Altman, Lachy Groom, amongst other leading software investors. We're looking for a Data Infrastructure engineer to centralize our data engineering capabilities while empowering teams across the organization to leverage data for their specific domains. Reporting to the Business Infrastructure Lead, this role sits at the intersection of infrastructure engineering and platform enablement; you'll be responsible for both the technical excellence of our data systems and ensuring every team can effectively access and utilize data for their needs. Our current data stack powers critical business functions but needs thoughtful stewardship and evolution. You'll own our point of view on data services, lead the buildout of our data pipeline architecture, and ensure we're maximizing value while optimizing GCP utilization and spend. You are an analytical builder who wants to be an important part of an exceptional team that focuses on using Roboflow's computer vision tools to impact and improve every industry. You have high agency and a bias toward action.

Requirements

  • Strong foundation in data engineering principles and patterns; experience building scalable, maintainable data pipelines
  • Deep understanding of modern data warehouse architecture and analytical database design
  • Hands-on experience with BigQuery and GCP data services (or equivalent cloud data platforms)
  • Proficiency with data transformation frameworks (dbt strongly preferred)
  • Experience with data visualization tools (Hex preferred)
  • Track record of balancing technical excellence with pragmatic cost optimization
  • Ability to translate business requirements into robust data solutions
  • Experience enabling non-technical teams to self-serve analytics
  • Strong communication skills and comfort working across organizational boundaries

Nice To Haves

  • Experience with workflow orchestration tools beyond GitHub Actions (Airflow, Prefect, Dagster, etc.)
  • Knowledge of real-time data processing and streaming architectures
  • Background in data governance and quality frameworks
  • Experience with reverse ETL or operational analytics patterns
  • Familiarity with BI tools and semantic layers
  • Understanding of privacy and compliance considerations in data systems

Responsibilities

  • Own the data platform: Provide stewardship for our BigQuery and dbt infrastructure, improving reliability, performance, and developer experience
  • Enable teams: Build self-service capabilities and tooling that allow product, sales, marketing, and engineering teams to access and analyze data independently
  • Design data architecture: Define and implement our philosophy on data services, including pipeline patterns, data modeling standards, and integration approaches
  • Optimize infrastructure: Deeply understand GCP and BigQuery to architect cost-effective solutions that scale with our growth
  • Build collaboratively: Work with stakeholders across the company to understand data needs and build pipelines that serve multiple use cases
  • Establish best practices: Create documentation, templates, and standards that make it easy for teams to work with data correctly
  • Act as advisor to team who produce and consume data with the goal of ensuring best practices
  • Drive data quality: Implement monitoring, testing, and validation frameworks to ensure data reliability

Benefits

  • $4000/yr Travel Stipend to travel anywhere anytime to work alongside other Roboflowers
  • $350/mo Productivity stipend to spend on things that make your work environment more productive, like high-speed internet at home or a co-working space
  • Cover up to 100% of your health insurance costs for you and your partner or family
  • Remote first/flexible schedule allowing you to work collaboratively with other team members and asynchronously
  • Unlimited PTO- with an annual 2 week minimum, we encourage you to take time off for yourself
  • 12 weeks parental leave
  • Equity in the company so we are all invested in the future of computer vision
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service