Data Engineer II

SAGIS PLLCHouston, TX
23h$110,000 - $130,000

About The Position

Sagis Diagnostics is a growing pathology and laboratory diagnostics organization committed to delivering accurate, timely, and patient-centered diagnostic services. Our data team is at the heart of our operational intelligence, building the pipelines, platforms, and reporting infrastructure that enable clinical and business decisions at scale. Data Engineer Level II We are seeking a seasoned Data Engineer with deep hands-on experience designing, building, and maintaining complex ETL pipelines and data workflows. This role sits within our Data Architecture team and is central to evolving our data platform across cloud, on-premise, and hybrid environments. The ideal candidate is a battle-tested practitioner, someone whose expertise was earned in production, not just in coursework. Design, build, and maintain scalable ETL/ELT pipelines that move data across disparate source systems including LIS (Laboratory Information Systems), billing platforms, and cloud data stores. Develop and optimize data workflows in Databricks using PySpark, Delta Lake, and Unity Catalog, following medallion architecture (Bronze → Silver → Gold) patterns. Architect and manage data integrations with Azure Data Factory, Azure SQL Managed Instance, and Azure Blob/ADLS Gen2 storage. Collaborate closely with the Director of Data Architecture to plan, execute, and document major pipeline and schema changes. Build and maintain robust Python-based automation scripts, data quality checks, and monitoring routines. Support reporting and analytics platforms including Power BI and Zoho Analytics, ensuring clean, performant data models and dataset refresh pipelines. Participate in schema design and maintenance for SQL Server environments including stored procedures, views, and schema-level organization. Contribute to data governance practices: access controls, Unity Catalog permissions, lineage documentation, and change management. Engage with internal stakeholders (clinical, billing, operations) to translate business requirements into data engineering solutions. Identify and resolve data quality, latency, and pipeline reliability issues proactively.

Requirements

  • ETL & Data Pipelines 5+ years of hands-on experience building production-grade ETL/ELT pipelines.
  • Proficiency with pipeline orchestration and scheduling (Azure Data Factory, Apache Airflow, or similar).
  • Strong understanding of data modeling — relational, dimensional, and lakehouse patterns.
  • Databricks Significant, production-level Databricks experience — Unity Catalog, Delta Lake, PySpark, and cluster/job management.
  • Familiarity with medallion architecture and incremental/merge data loading strategies.
  • Azure Ecosystem Demonstrated experience with Azure Data Factory for pipeline orchestration and data movement.
  • Proficiency with Azure SQL Managed Instance and Azure Blob / ADLS Gen2.
  • Familiarity with Azure networking, managed identities, and role-based access controls.
  • SQL & Database Engineering Expert-level SQL skills including complex joins, window functions, CTEs, stored procedures, and performance tuning.
  • Experience working within structured schema environments (SQL Server / Azure SQL MI).
  • Python Strong Python skills for data pipeline development, automation, and API integrations.
  • Experience with libraries such as pandas, pyodbc, requests, and Azure SDKs.

Nice To Haves

  • Experience with Power BI including dataset modeling, dataflows, and gateway configuration.
  • Experience with Zoho Analytics including workspace management, custom queries, and report/dashboard development.
  • Healthcare industry experience, particularly in laboratory diagnostics, pathology, or revenue cycle management.
  • Exposure to medical coding systems including CPT (Current Procedural Terminology) and ICD-10 code structures and their use in clinical billing workflows.
  • Familiarity with Laboratory Information Systems (LIS) or billing platforms such as VitalAxis or HealthPac.
  • Experience with Microsoft Graph API or SharePoint integration for collaborative data workflows.
  • Snowflake experience for hybrid or partner data ingestion scenarios.
  • Exposure to Databricks Partner Connect, external table sharing, or cross-cloud data patterns.

Responsibilities

  • Design, build, and maintain scalable ETL/ELT pipelines that move data across disparate source systems including LIS (Laboratory Information Systems), billing platforms, and cloud data stores.
  • Develop and optimize data workflows in Databricks using PySpark, Delta Lake, and Unity Catalog, following medallion architecture (Bronze → Silver → Gold) patterns.
  • Architect and manage data integrations with Azure Data Factory, Azure SQL Managed Instance, and Azure Blob/ADLS Gen2 storage.
  • Collaborate closely with the Director of Data Architecture to plan, execute, and document major pipeline and schema changes.
  • Build and maintain robust Python-based automation scripts, data quality checks, and monitoring routines.
  • Support reporting and analytics platforms including Power BI and Zoho Analytics, ensuring clean, performant data models and dataset refresh pipelines.
  • Participate in schema design and maintenance for SQL Server environments including stored procedures, views, and schema-level organization.
  • Contribute to data governance practices: access controls, Unity Catalog permissions, lineage documentation, and change management.
  • Engage with internal stakeholders (clinical, billing, operations) to translate business requirements into data engineering solutions.
  • Identify and resolve data quality, latency, and pipeline reliability issues proactively.

Benefits

  • Comprehensive Benefits: Medical, Dental, and Vision with Blue Cross/Blue Shield.
  • Company-paid Short-Term and Long-Term Disability, Basic Life, and EAP.
  • Voluntary Accident, Critical Illness, and Life Insurance options.
  • Work-Life Balance: Flexible Scheduling and Fair Wages.
  • Paid Time Off, Floating Holidays, and Scheduled Holidays.
  • Paid Parental Leave.
  • 401(k) Contribution.
  • Free Parking.
  • Wellness Initiatives and a collaborative team culture.
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service