Data Engineer

SuvodaConshohocken, PA

About The Position

Suvoda is seeking a recent graduate with a passion for data and a strong technical foundation to join their Product Development team. This is an opportunity to launch a data engineering career at Suvoda, a global clinical trial technology company where your work directly helps advance human health. As a Data Engineer, you will join a collaborative team building a modern, cloud-native data platform. You will help build domain-oriented data products using GraphQL APIs and contribute to near real-time reporting through AWS DMS replication to Aurora PostgreSQL.

Requirements

  • Bachelor’s degree in technical fields such as Computer Science or Mathematics.
  • Strong foundational skills in Python and SQL — you’re comfortable writing queries and scripts from scratch.
  • Academic or project experience with data engineering concepts: pipelines, data modeling, or ETL/ELT workflows.
  • Exposure to cloud platforms, particularly AWS (coursework, personal projects, or internships all count).
  • Familiarity with data structures, distributed systems, or database design — you understand the “why” behind the tools.
  • A collaborative mindset and strong communication skills — you ask good questions and explain your thinking clearly.
  • A genuine enthusiasm for data and a desire to keep learning in a fast-moving environment.

Nice To Haves

  • Master’s degree or relevant certifications.
  • Experience with event-driven architectures (e.g., Kafka, Kinesis).
  • Familiarity with data cataloging and metadata management tools.
  • Awareness of data privacy and compliance standards (e.g., GDPR, HIPAA).
  • Exposure to agile development and DevOps practices.

Responsibilities

  • Assist in implementing a data mesh architecture using GraphQL APIs to expose domain-owned data products.
  • Help build and maintain an AWS-based data lake using S3, Glue, Lake Formation, Athena, and Redshift.
  • Develop and maintain ETL/ELT pipelines using AWS Glue and PySpark for batch and streaming data workloads.
  • Support AWS DMS pipelines to replicate data into Aurora PostgreSQL for near real-time analytics.
  • Follow best practices for data governance, quality, observability, and API design.
  • Collaborate with product, engineering, and analytics teams to deliver reliable data solutions.
  • Contribute to CI/CD and automation efforts for data infrastructure and pipelines.
  • Stay curious — explore emerging tools and share ideas that could make our platform even better.
© 2026 Teal Labs, Inc
Privacy PolicyTerms of Service