About The Position

At Snowflake, we are powering the era of the agentic enterprise, seeking AI-native thinkers across every function who are energized by the opportunity to reinvent how they work. Observe by Snowflake is an AI-powered observability platform built on the Snowflake AI Data Cloud and engineered for scale. It ingests and stores logs, metrics, traces, and events on an open, scalable data lakehouse using open formats like Apache Iceberg, at dramatically lower cost. A dynamic Context Graph and chat-based AI SRE provide rich context and automated workflows so teams can move from detection to root cause and resolution 10x faster. Leading engineering teams rely on Observe to troubleshoot hundreds of terabytes of telemetry daily while maintaining reliability at enterprise scale. As part of Snowflake, Observe combines startup-style ownership and velocity with the global reach, operational excellence, and ecosystem of one of the world's leading data platforms. We are hiring a Senior Software Engineer for the Observe Data Management team. This team owns the core pipelines that ingest and process over 1 petabyte of telemetry data per day, which is the foundational infrastructure powering Observe's entire observability stack. You'll be working at the intersection of massive scale, open-source innovation, and real-world reliability challenges for enterprise customers around the globe.

Requirements

  • 5+ years of software engineering experience with deep expertise in distributed systems
  • Proficiency in Go and/or C++, with an ability to write high-performance, production-grade systems code
  • Demonstrated experience designing and operating large-scale data ingestion or stream processing pipelines
  • A strong sense of user empathy and product intuition — you think beyond APIs and care about end-to-end data onboarding and management experience.
  • Hands-on experience building and running services across major cloud providers (AWS and/or Azure)
  • Strong fundamentals in systems programming: concurrency, memory management, networking, and I/O
  • A track record of solving hard infrastructure or platform engineering problems at scale
  • B.S. in Computer Science, Engineering, or equivalent practical experience

Nice To Haves

  • Experience with OpenTelemetry SDKs, instrumentation, or ecosystem tooling
  • Prior open-source contributions or project maintainership
  • Familiarity with Apache Iceberg or other open table formats and data lakehouse architectures
  • Background in observability, monitoring, or SRE
  • Experience with multi-cloud data infrastructure or telemetry platforms at petabyte scale

Responsibilities

  • Design, build, and scale high-throughput data ingestion and processing pipelines handling petabyte-scale telemetry — logs, metrics, traces, and events
  • Develop performance-critical, distributed systems components in Go and/or C++ that operate reliably across AWS and Azure
  • Contribute to OpenTelemetry and drive Observe's open-source strategy, including external community engagement and upstream contributions
  • Architect solutions that maintain enterprise-grade availability and low latency under extreme data volumes
  • Collaborate with SRE, product, and platform teams to define data reliability standards and improve detection-to-resolution times for customers
  • Debug and resolve complex distributed systems issues at the deepest layers of the stack
  • Help shape the technical roadmap for the Data Management team and mentor engineers across the organization
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service