Jr Data Engineer

HalvikWashington, DC
1d

About The Position

Halvik Corp delivers a wide range of services to 13 executive agencies and 15 independent agencies. Halvik is a highly successful WOB business with more than 50 prime contracts and 500+ professionals delivering Digital Services, Advanced Analytics, Artificial Intelligence/Machine Learning, Cyber Security and Cutting-Edge Technology across the US Government. Be a part of something special! Role and Responsibilities: Analyze Informatica workflows and mappings to understand source-to-target logic, transformations, dependencies, and scheduling order. Convert Informatica mappings into Databricks pipelines using PySpark / Spark SQL Implement data ingestion from on-prem and cloud sources into Databricks (landing → bronze → silver) Adapt existing ETL logic to align with a new enterprise data model, identifying gaps and required transformation changes Support unit testing, reconciliation, and data validation between legacy and modern pipelines Document migration logic, assumptions, and deviations from legacy behavior Collaborate with senior engineers, data architects, and QA teams during iterative migration cycles

Requirements

  • Working knowledge of Informatica PowerCenter - Mappings, workflows, sessions. Common transformations (Source Qualifier, Expression, Lookup, Joiner, Aggregator, Router, Filter)
  • Basic to intermediate experience with Databricks: PySpark, Spark, SQL, Notebooks and jobs
  • Strong SQL fundamentals (joins, aggregations, window functions)
  • Understanding of ETL / ELT concepts, data warehousing, and batch processing
  • Data Modeling & Analysis Skills
  • Understand of report development.
  • Data Modeling & Analysis Skills
  • Interpret legacy data models and mapping documents.
  • Identify how legacy fields map (or do not map) to a new target data model
  • Flag missing logic, derived fields, or transformation gaps early
  • Strong attention to detail when validating row counts, aggregates, and business rules

Nice To Haves

  • Exposure to Informatica-to-Databricks migration or similar modernization efforts.
  • Familiarity with Delta Lake, medallion architecture (Bronze / Silver / Gold).
  • Basic understanding of AWS (S3, IAM concepts)
  • Experience reading or generating code from Informatica XML exports
  • Bachelor's Degree in a related area or equivalent experience.

Responsibilities

  • Analyze Informatica workflows and mappings to understand source-to-target logic, transformations, dependencies, and scheduling order.
  • Convert Informatica mappings into Databricks pipelines using PySpark / Spark SQL
  • Implement data ingestion from on-prem and cloud sources into Databricks (landing → bronze → silver)
  • Adapt existing ETL logic to align with a new enterprise data model, identifying gaps and required transformation changes
  • Support unit testing, reconciliation, and data validation between legacy and modern pipelines
  • Document migration logic, assumptions, and deviations from legacy behavior
  • Collaborate with senior engineers, data architects, and QA teams during iterative migration cycles

Benefits

  • Company-supported medical, dental, vision, life, STD, and LTD insurance
  • Benefits include 11 federal holidays and PTO
  • Eligible employees may receive performance-based incentives in recognition of individual and/or team achievements.
  • 401(k) with company matching
  • Flexible Spending Accounts for commuter, medical, and dependent care expenses
  • Tuition Assistance
  • Charitable Contribution matching
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service