TRM Labs-posted 5 months ago
$190,000 - $220,000/Yr
Full-time • Senior
101-250 employees

TRM Labs is a blockchain intelligence company committed to fighting crime and creating a safer world. By leveraging blockchain data, threat intelligence, and advanced analytics, our products empower governments, financial institutions, and crypto businesses to combat illicit activity and global security threats. At TRM, you'll join a mission-driven, fast-paced team made up of experts in law enforcement, data science, engineering, and financial intelligence, tackling complex global challenges daily. Whether analyzing blockchain data, developing cutting-edge tools, or collaborating with global organizations, you'll have the opportunity to make a meaningful and lasting impact. We’re building the foundational data infrastructure powering next-generation analytics at scale. As part of our mission, we’re architecting a modern data lakehouse to support complex workloads, real-time data pipelines, and secure data governance—at petabyte scale. We are looking for a Senior Data Engineer to help us design, implement, and scale core components of our lakehouse architecture.

  • Architect and scale a high-performance data lakehouse on GCP, leveraging technologies like StarRocks, Apache Iceberg, GCS, BigQuery, Dataproc, and Kafka.
  • Design, build, and optimize distributed query engines such as Trino, Spark, or Snowflake to support complex analytical workloads.
  • Implement metadata management in open table formats like Iceberg and data discovery frameworks for governance and observability using Iceberg compatible catalogs.
  • Develop and orchestrate robust ETL/ELT pipelines using Apache Airflow, Spark, and GCP-native tools (e.g., Dataflow, Composer).
  • Collaborate across departments, partnering with data scientists, backend engineers, and product managers to design and implement.
  • 5+ years of experience in data or software engineering, with a focus on distributed data systems and cloud-native architectures.
  • Proven experience building and scaling data platforms on GCP, including storage, compute, orchestration, and monitoring.
  • Strong command of one or more query engines such as Trino, Presto, Spark, or Snowflake.
  • Experience with modern table formats like Apache Hudi, Iceberg, or Delta Lake.
  • Exceptional programming skills in Python, as well as adeptness in SQL or SparkSQL.
  • Hands-on experience orchestrating workflows with Airflow and building streaming/batch pipelines using GCP-native services.
  • Generous PTO
  • Holidays
  • Parental Leave for full time employees
  • Participation in TRM’s equity plan
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service