Data Engineer

Data SocietyWashington, DC
Remote

About The Position

The Data Engineer is responsible for building scalable, performant data pipelines that power critical operational and analytical applications. The engineer must be able to work closely with our data science teams to build the supporting data scaffolding to orchestrate, test, and monitor data systems. This is a customer-facing role within a cross-functional team so the ability to manage timelines, work both autonomously and collaboratively, and communicate effectively are a must. As this position will help to support federal contracts with security requirements, you must be a US Citizen to qualify. Occasional travel to client offices in Richmond, VA will be required, and preference will be given to local DMV area candidates. This is a full-time, remote/work from home, benefits eligible position.

Requirements

  • Advanced Degree in Statistics, Applied Mathematics, Data Science, Computer Science, Operations Research or other closely related other quantitative or mathematical disciplines.
  • 5+ years of data and analytics engineering in cloud environments
  • Expertise in SQL, Python, and schema design with experience in data cataloging and governance tools
  • Experienced with data transformation and ETL best practices
  • Experienced with data orchestration tools like Airflow, transformation frameworks like dbt, and cloud deployment tools like Terraform.
  • Demonstrated exceptional oral and written communication skills.
  • The ability to work independently and in a team environment.
  • The ability to work effectively across functions, levels and disciplines.
  • Strong problem solving and critical thinking skills.
  • Superior team-working skills, and a desire to learn, contribute, and explore.

Nice To Haves

  • Experience with Snowflake, Databricks, Kafka, Flume, Spark, or Flink is a plus

Responsibilities

  • Ability to build a full data pipeline from data ingestion to processing/transformation to load to visualization and analysis.
  • Design and manage large-scale data warehouses, lakehouses, and/or data marts
  • Build and optimize data transformation pipelines using tools like dbt to support data flow from ingestion through analytics
  • Champion data governance principles and quality standards, ensuring data lineage, documentation, and metadata are maintained
  • Create efficient, performant SQL-based data queries and Python-based data processing jobs
  • Demonstrate ability to balance computational load, performance, and cost
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service