Data Engineer

Nimble SolutionsChesterfield, MO
Hybrid

About The Position

Data Engineer Chesterfield Office Hybrid or Remote Why You'll Want to Join! Join a leading Revenue Cycle Management (RCM) company dedicated to transforming healthcare data into actionable insights. We leverage cutting-edge technology to streamline financial and operational processes, improving efficiency and patient outcomes. We are looking for a Data Engineer to help optimize data pipelines and build a next-generation data infrastructure incorporating technologies such as Microsoft Fabric, Azure Synapse, Databricks, and Snowflake . Position Overview Lead the modernization of our data infrastructure as a Data Engineer for nimble. You'll architect scalable cloud-native pipelines using Microsoft Fabric and Databricks to transform healthcare data—claims, EMR/EHR, HL7/FHIR—into actionable insights that drive revenue cycle optimization and clinical outcomes. Why This Role Matters Healthcare data engineering is mission-critical: clean, governed data flows directly impact financial accuracy, compliance, and the decisions that improve patient care. Your ETL/ELT pipelines enable our analytics and data science teams to unlock the full potential of healthcare data.

Requirements

  • 3+ years of professional data engineering or ETL/ELT development experience
  • Expert-level SQL skills with proven optimization experience
  • Proficiency in Python, Scala, or similar data processing languages
  • Hands-on experience with cloud data platforms (Azure Synapse, Snowflake, Databricks, or equivalent)
  • Understanding of healthcare data standards (HL7, FHIR, claims data structures)
  • Strong grasp of data modeling, normalization, and schema design
  • Experience with data versioning, CI/CD pipelines, and data quality frameworks

Nice To Haves

  • Experience with Microsoft Fabric or Azure Data Factory
  • Knowledge of HIPAA compliance and healthcare data security
  • Background in healthcare, RCM, or claims processing
  • Experience with dbt (data build tool) or equivalent transformation frameworks
  • Exposure to dimensional modeling and data warehousing best practices

Responsibilities

  • Design, build, and optimize ETL/ELT pipelines using Azure Synapse, Databricks, and Snowflake
  • Develop robust data models and schemas for healthcare datasets, including claims, EMR/EHR, HL7, and FHIR standards
  • Write and optimize SQL queries for performance across large healthcare datasets
  • Implement data governance, quality frameworks, and HIPAA compliance controls
  • Collaborate with analytics, data science, and business teams to define data requirements
  • Monitor and troubleshoot data pipeline health and performance
  • Develop Python or Scala code for complex transformations and data processing
  • Support Power BI and analytics teams with data modeling and performance optimization
  • Document data lineage, transformations, and technical architecture

Stand Out From the Crowd

Upload your resume and get instant feedback on how well it matches this job.

Upload and Match Resume

What This Job Offers

Job Type

Full-time

Career Level

Mid Level

Education Level

No Education Listed

Number of Employees

101-250 employees

© 2026 Teal Labs, Inc
Privacy PolicyTerms of Service