About The Position

The Scientific Computing & Data group at the Icahn School of Medicine at Mount Sinai partners with scientists to accelerate scientific discovery. To achieve these aims, we support a cutting-edge high-performance computing (HPC) and data ecosystem along with MD/PhD-level support for researchers. The group is composed of a high-performance computing team, the research clinical data warehouse team and a research data services team. The Scientific Computing and Data group operate the secure enclave for the Kidney Precision Medicine Project (KPMP) clinical data warehouse and data lake for the KPMP. The Kidney Precision Medicine Project (KPMP) is an ambitious, multi-year project funded by the National Institute of Diabetes and Digestive and Kidney Diseases (NIDDK) with the purpose of understanding and finding new ways to treat chronic kidney disease (CKD) and acute kidney injury (AKI). The Lead Clinical Data Warehouse Engineer is a senior technical specialist responsible for leading the ongoing development of the Kidney Precision Medicine Project (KPMP) data warehouse, a centralized multimodal data repository for biomedical research. In this position, the Lead Clinical Data Warehouse Engineer will work with other members of the data warehouse team and KPMP to expand the functionality and integrate new multimodal data sources into the data warehouse. The KPMP data warehouse is built on the Microsoft SQL Server technology stack, with data extracted from KPMP’s clinical systems, and transformed into OHDSI’s Observational Medical Outcomes Partnership (OMOP) Common Data Model. This position is responsible for driving and leading the technology direction of current and future development in an area of expertise which may include application development, systems programming, data management, etc. This individual may function as a team lead and influences strategic technical implementation and provides creative solutions for very complex requirements often leading large project efforts. Project Scope: Large complex programs and strategic initiatives involving multiple sites or up to 40 integrated work tracks.

Requirements

  • Bachelors degree in a technical discipline; Masters degree preferred
  • 12-15 years preferred of related experience, including 8 years of experience with the design, development and maintenance of relational databases and ETL jobs/data pipelines.
  • Expert level ETL development experience with Microsoft SSIS, Visual Studio, SQL Server Data Tools and SQL Agent jobs is required. Prior experience with other ETL tools is a plus.
  • Expert level Microsoft T-SQL development skills are required: stored procedures, user defined functions, sequences, advanced T-SQL features, metadata or parameter driven dynamic SQL, and the use of Linked Servers. Experience with Oracle PL/SQL and/or PostgreSQL, PL/pgSQL is a strong plus.
  • Prior experience working with healthcare data is strongly preferred, which may include electronic health records, medical and pharmacy billing & claims, and/or healthcare cost accounting systems. Experience with Epic Systems Corporation's Clarity & Caboodle reporting databases in strongly preferred.
  • Strong consultative customer service and communications skills (verbal and written) to effectively interact with researchers, clinicians, hospital administrators, and other IT personnel.
  • Knowledge of ETL development best practices is required, including explicit source-to-target mappings; pattern-based and parameter driven development; and rigorous version control using GIT, BitBucket, Microsoft Teams Foundation Server (TFS), or equivalent.
  • Experience with DevOps and Software Development Life Cycle (SDLC) best practices. Experience with Agile methodologies (Scrum, Kanban) and tools (Atlassian, Jira and Confluence) is a strong plus.
  • Strong knowledge of relational and dimensional data modeling principles and best practices.
  • Excellent critical thinking, problem solving, multi-tasking and collaboration skills.
  • Fast learner with the ability to work independently and productively in a fast-paced environment.
  • Experience with database administration tasks in a MS SQL Server environment is a plus: database configuration, performance tuning indexes, partitioning, materialized views, and other methods for improving query performance; user provisioning and access permissions; backups and restorations.
  • Experience working with MS SSRS, MS Power BI, Tableau or similar business intelligence tools.
  • Scripting experience (shell, Python, PERL or similar) .
  • Some experience with HL7 messages and web services interfaces (e.g. REST API).
  • Experience with OMOP.

Nice To Haves

  • Expert knowledge of data warehousing principles: database design/3NF, dimensional modeling (e.g. types of fact tables, slowly changing dimensions), change data capture/bi-temporal versioning, incremental data loads, maintain transparent data lineage, etc.

Responsibilities

  • Design, build, test, maintain and control data pipelines and ETL jobs for integrating data into the KPMP data warehouse. Such data pipelines are implemented in T-SQL stored procedures, SSIS and SQL Agent jobs, and make extensive use of metadata-driven dynamic SQL and Varigence's BIML.
  • Develop standards and procedures for database and data pipeline design, operational management and ongoing maintenance.
  • Design, build, test, maintain, and control other processes for data management, including management and loading of flat files in a variety of formats (e.g. csv, tsv, pipe-delimited, XML, JSON.)
  • Facilitate design sessions, code walkthroughs, peer reviews, and produce technical documentation.
  • Performance tuning of database objects, stored procedures, ETL jobs, and related scripts to optimize both end-user customer queries and data pipeline ETL processing.
  • Monitor scheduled ETL jobs and other processes to ensure expected functioning and uptime of data pipelines. Lead and/or coordinate the troubleshooting and remediation of all ETL job failures in a timely manner. On call support is sometimes required.
  • Deliver world class customer service in all interactions with customers, stakeholders, and other teams within Mount Sinai's Digital & Technology Partners (DTP) organization. Maintain a customer-focused approach to provide solutions that are science/research driven.
  • Maintain patient privacy and the integrity and security of healthcare data in all databases and systems, including compliance with all applicable laws, regulations, and Mount Sinai's policies related to the Institutional Review Board (IRB), patient privacy and IT cybersecurity. These laws include but are not limited to, the "Common Rule" (45 CFR Part 46), HIPAA (45CFR Part 164), 42 CFR Part 2, and New York's Article 27-F.
  • Work effectively and collaboratively with other IT professionals throughout Mount Sinai.
  • As a senior technical lead, independently investigate and stay abreast of new and emerging technologies that Scientific Computing & Data could leverage to provide new capabilities, to boost efficiencies or quality, and/or lower operating costs.
  • May act as a Project Leader.
  • May assist the DBAs with database administration tasks, including user provisioning and access permissions, backups, environment restorations, capacity planning and load monitoring, performance tuning, etc.
  • May assist with the design, build, test and performance tuning of reports and dashboards in SSRS, Microsoft Power BI, Tableau, or similar business intelligence tools.
  • Perform related duties as assigned or requested.
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service