Data Engineer

San R&D Business Solutions LLCMilton, GA
2dOnsite

About The Position

We are seeking an experienced Data Engineer to join our team. The ideal candidate will have strong expertise in Python, SQL, and Google Cloud Platform (GCP), with hands-on experience building and maintaining scalable data pipelines. This role involves working with large datasets, optimizing data workflows, and collaborating closely with data scientists, analysts, and engineering teams to deliver reliable, high-quality data solutions.

Requirements

  • Bachelor’s degree in Computer Science, Engineering, or a related field (or equivalent experience)
  • Proven experience as a Data Engineer
  • Strong proficiency in Python for data processing and automation
  • Advanced SQL skills, including complex query writing and performance optimization
  • Hands-on experience with Google Cloud Platform (GCP)
  • Strong experience with BigQuery, Apache Spark, Apache Beam, GCP Dataflow
  • Working knowledge of Java for data processing or pipeline development
  • Solid understanding of data warehousing concepts and data modeling
  • Strong problem-solving skills and attention to detail
  • Excellent communication skills and ability to work in a team-oriented environment

Nice To Haves

  • Experience with data governance, data quality, and data security best practices
  • Familiarity with CI/CD pipelines and version control systems (e.g., Git)
  • Experience working in cloud-native data architectures
  • Exposure to additional cloud platforms (AWS or Azure)
  • Experience supporting large-scale, enterprise data environments

Responsibilities

  • Design, build, and maintain scalable, reliable data pipelines to process large volumes of data
  • Develop, optimize, and maintain ETL/ELT workflows using Python, SQL, and cloud-native tools
  • Write and optimize complex SQL queries for data transformation and analysis
  • Work extensively with BigQuery for data storage, processing, and performance optimization
  • Develop and maintain data pipelines using Apache Spark, Apache Beam, and GCP Dataflow
  • Implement data validation, quality checks, and monitoring to ensure data accuracy and integrity
  • Troubleshoot and resolve pipeline failures, performance issues, and data inconsistencies
  • Collaborate with data scientists, analysts, and business stakeholders to understand data requirements
  • Apply best practices in data engineering, data modeling, and cloud architecture
  • Stay current with emerging data engineering tools and technologies
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service