Big Data Engineer - Medicare/Medicaid

Molina HealthcareOmaha, NE
248d$77,969 - $155,508

About The Position

We are seeking a highly skilled and forward-thinking Big Data Engineer to join our healthcare data team. This role encompasses the end-to-end design, development, and management of large-scale data systems tailored for healthcare analytics. The ideal candidate will be responsible for architecting and maintaining robust, scalable, and secure data pipelines that support critical decision-making across the organization. This position requires deep technical expertise in modern Big Data tools, real-time and batch data integration, and a strong understanding of data governance and compliance in healthcare environments.

Requirements

  • Minimum 3 years of hands-on experience in Big Data engineering, data integration, and pipeline development.
  • Proficiency in Python, Java, or Scala for data transformation and system scripting.
  • Expertise in Big Data tools: Spark, Hive, Impala, Presto, Phoenix, Kylin, and Hadoop (HDFS, YARN).
  • Experience building real-time stream-processing systems using Kafka, Storm, or Spark Streaming.
  • Strong knowledge of NoSQL databases like HBase and MemSQL, and traditional RDBMS including PostgreSQL, Oracle, and SQL Server.
  • Skilled in ETL design and development using tools such as Talend or Informatica.
  • Demonstrated experience in deploying and monitoring big data infrastructure with Ambari, Cloudera Manager, and Zookeeper.
  • Solid understanding of data warehousing, data validation, data quality checks, metadata management, and governance.

Nice To Haves

  • 5+ years of progressive experience in Big Data engineering or analytics.
  • Prior experience working in the healthcare industry with familiarity in clinical, claims, or care management data.
  • Experience with cloud platforms (AWS, Azure) and containerization tools (Docker, Kubernetes).

Responsibilities

  • Architect and implement scalable, high-performance Big Data solutions that support structured and unstructured data from diverse sources.
  • Build and manage batch and real-time data ingestion/extraction pipelines using tools like Kafka, Spark Streaming, and Talend.
  • Develop reusable and efficient ETL frameworks using Python/Scala for high-volume data transformation and movement.
  • Design and optimize data models to support analytical and operational use cases, including healthcare claims and utilization data.
  • Collaborate with cross-functional teams, including data scientists, analysts, and business partners, to translate requirements into robust data products.
  • Deploy, monitor, and troubleshoot Hadoop-based infrastructure using tools such as Cloudera Manager, Ambari, and Zookeeper.
  • Enforce data quality, security, and compliance standards using tools such as Kerberos, Ranger, and Sentry.
  • Implement web services and APIs (REST/SOAP) to enable seamless integration with applications and visualization platforms.
  • Contribute to data governance initiatives, including metadata management, lineage tracking, and quality assurance.

Benefits

  • Competitive benefits and compensation package.
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service