Lead with Purpose. Partner with Impact. We are seeking a seasoned Databricks Data Engineer with expertise in Azure cloud services and the Databricks Lakehouse platform. The role involves designing and optimizing large-scale data pipelines, modernizing cloud-based data ecosystems, and enabling secure, governed data solutions. Strong skills in SQL, Python, PySpark, ETL/ELT frameworks, and experience with Delta Lake, Unity Catalog, and CI/CD automation are essential. What you’ll Do: Design, build, and optimize large-scale data pipelines on the Databricks Lakehouse platform, ensuring reliability, scalability, and governance. Modernize the Azure-based data ecosystem, contributing to cloud architecture, distributed data engineering, data modeling, security, and CI/CD automation. Utilize Apache Airflow and similar tools for orchestration and workflow automation. Work with financial or regulated datasets, applying strong compliance and governance practices. Develop and optimize ETL/ELT pipelines using Python, PySpark, Spark SQL, and Databricks notebooks. Design and optimize Delta Lake data models for reliability, performance, and scalability. Implement and manage Unity Catalog for RBAC, lineage, governance, and secure data sharing. Build reusable frameworks using Databricks Workflows, Repos, and Delta Live Tables. Create scalable ingestion pipelines for APIs, databases, files, streaming sources, and MDM systems. Automate API ingestion and workflows using Python and REST APIs. Support data governance, lineage, cataloging, and metadata initiatives. Enable downstream consumption for BI, data science, and application workloads. Write optimized SQL/T-SQL queries, stored procedures, and curated datasets for reporting. Automate deployments, DevOps workflows, testing pipelines, and workspace configuration.
Stand Out From the Crowd
Upload your resume and get instant feedback on how well it matches this job.
Job Type
Full-time
Career Level
Mid Level
Education Level
No Education Listed
Number of Employees
1,001-5,000 employees