Build distributed, scalable, and reliable data pipelines that ingest and process data at scale and in real-time. Create metrics and apply business logic using Spark, Scala, R, Python, and/or Java. Model, design, develop, code, test, debug, document and deploy an application to production through standard processes. Harmonize, transform, and move data from a raw format to consumable, curated views. Analyze, design, develop, and test applications. Contribute to the maturation of Data Engineering practices, which may include providing training and mentoring to others. Perform Data Designer activities to transform raw data to meaningful datasets and extracts, such as business logic design, source-to-target mappings, data sourcing strategy, and transformation rules. Apply strong Data Governance principles, standards, and frameworks to promote data consistency and quality while effectively managing and protecting the integrity of corporate data. Collaborate with cross-functional teams including data scientists, analysts, and business stakeholders to gather requirements and deliver data solutions aligned with business goals. Document technical specifications, data flow diagrams, and operational procedures to support knowledge transfer and audit requirements. Develop ETL/ELT workflows using Python, PySpark, and SQL to transform raw data into structured formats suitable for downstream consumption. Ensure data quality and integrity through the implementation of data validation, monitoring, and alerting mechanisms using tools like such as AWS CloudWatch, Glue DataBrew, and custom scripts.
Stand Out From the Crowd
Upload your resume and get instant feedback on how well it matches this job.
Job Type
Full-time
Career Level
Entry Level
Number of Employees
251-500 employees