Sr Data Engineer - Databricks

McKessonColumbus, OH
Hybrid

About The Position

McKesson is an impact-driven, Fortune 10 company that touches virtually every aspect of healthcare. We are known for delivering insights, products, and services that make quality care more accessible and affordable. Here, we focus on the health, happiness, and well-being of you and those we serve – we care. What you do at McKesson matters. We foster a culture where you can grow, make an impact, and are empowered to bring new ideas. Together, we thrive as we shape the future of health for patients, our communities, and our people. If you want to be part of tomorrow’s health today, we want to hear from you. The Senior Data Engineer - Databricks with CoverMyMeds will support and expand the data platforms that process, store, and organize data critical for the Data & Analytics team. We are hiring for two complementary roles within this scope: Ingestion-Focused Engineer – responsible for defining scalable ingestion patterns, establishing standards, and mentoring junior engineers as we expand data intake across domains Pipeline-Focused Engineer (Databricks) – responsible for designing and building robust data pipelines, leveraging Databricks and modern data engineering best practices Both roles play a key part in shaping how data flows into and across our platform and will partner closely to deliver scalable, high-quality data solutions. A background in pharmacy claims data is a strong plus for either role. These roles participate in technical strategy and execution to provide trusted, stable, reliable, responsive, and secure solutions, while proactively informing business partners on platform health and issue resolution. The Senior Data Engineer will work collaboratively with Data Systems Analysts as well as Analytics and Technology partners to solve business problems and deliver scalable solutions. Candidate must reside in the Columbus, OH area to support a hybrid work schedule. At this time, we are not able to offer sponsorship for employment visas for this role. This includes individuals currently on F-1 OPT, STEM OPT, or any other visa status that would require future sponsorship. Candidates must be authorized to work in the United States on a permanent basis without the need for current or future sponsorship.

Requirements

  • Bachelor's degree in Computer Science, Information Systems or related field, or related experience, and typically requires 7+ years of technical and professional expereince related to the responsibilities listed above
  • Strong hands-on experience with cloud data platforms (Databricks strongly preferred), data warehousing, and pipeline development
  • Experience building and maintaining scalable data ingestion frameworks and pipelines
  • Experience working with structured and unstructured data in batch and real-time environments
  • Strong SQL skills and experience applying business logic within data transformations
  • Experience supporting Reporting & Analytics, Real-Time Analytics, Systems Integration, and Data Governance
  • Demonstrated expertise in data modeling and database design
  • Experience with large-scale data platform implementations

Nice To Haves

  • Experience with Databricks (Delta Lake, Spark, workflows, and pipeline orchestration)
  • Background working with pharmacy or healthcare claims data
  • Excellent written and verbal communication skills; clear, timely communication with the ability to influence across teams
  • Ability to find creative solutions to complex problems
  • Strong sense of urgency and ownership for task/project completion
  • Highly collaborative, able to work across business and technical teams to achieve results

Responsibilities

  • Design and develop solutions and commissioning of complex data across systems for internal and external customers
  • Develop data ingestion and integration pipelines from various sources into the data warehouse / lakehouse
  • Define and implement scalable ingestion and pipeline patterns to support both batch and real-time processing
  • Work with databases, files, and unstructured data to identify, transport, and validate data required for synchronization and incremental loads
  • Write code in SQL and/or cloud-based tools such as Databricks (preferred) or Snowflake to cleanse, apply business logic, and standardize data according to business rules
  • Design conceptual and logical data models based on business reporting requirements
  • Partner with business and application teams to understand source system data flows and architect efficient downstream solutions
  • Mentor junior engineers and contribute to engineering standards, best practices, and code quality
  • Role alignment (ingestion vs. pipeline focus) will be determined based on candidate strengths and team needs
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service