Databricks SME

Scicom Infrastructure ServicesAtlanta, GA
4d

About The Position

Key Responsibilities: Architecture & Platform Design Design enterprise Databricks Lakehouse architectures aligned with the Databricks Well-Architected Framework Define reference architectures for batch, streaming, analytics, and ML workloads Select and standardize cluster, compute, and workspace architectures Design multi-workspace strategies (dev/test/prod, shared vs. isolated) Ensure architectures meet scalability, availability, and performance requirements Well-Architected Framework Alignment Apply Databricks best practices across all pillars, including: Security & Governance (Unity Catalog, IAM, data access controls) Reliability & Resilience (job retries, checkpointing, failure isolation) Performance Efficiency (cluster sizing, autoscaling, caching) Cost Optimization (compute policies, workload separation, monitoring) Operational Excellence (monitoring, automation, CI/CD, runbooks) Implementation & Engineering Lead Databricks workspace, cluster, and Unity Catalog implementations Implement Delta Lake, Delta Live Tables (DLT), and Structured Streaming Build and optimize ETL/ELT pipelines using Spark and SQL Integrate Databricks with cloud services (S3/ADLS/GCS, IAM, Key Vault, networking) Establish CI/CD pipelines for notebooks, jobs, and infrastructure Security, Governance & Compliance Implement Unity Catalog for centralized governance Define data classification, lineage, and audit strategies Enforce least-privilege access and secure networking patterns Support compliance requirements (HIPAA, SOC 2, PCI, GDPR as applicable) Operations & Optimization Monitor platform health, performance, and cost Troubleshoot production issues across jobs, clusters, and data pipelines Perform workload tuning and cost-performance optimization Define SLOs, alerts, and operational metrics Collaboration & Leadership Partner with Data Engineering, Analytics, ML, Platform, and Security teams Translate business requirements into technical architectures Provide architectural guidance and technical mentorship Communicate risks, tradeoffs, and recommendations to leadership

Requirements

  • 7+ years in data engineering, analytics, or platform architecture
  • 3–5+ years hands-on Databricks experience in production environments
  • Proven experience applying the Databricks Well-Architected Framework
  • Experience designing cloud-native lakehouse architectures
  • Experience supporting mission-critical data platforms
  • Databricks Lakehouse Platform
  • Apache Spark (PySpark / Scala / Spark SQL)
  • Delta Lake, Delta Live Tables, Structured Streaming
  • Unity Catalog (governance, lineage, access controls)
  • Cloud platforms: AWS, Azure, or GCP
  • Infrastructure as Code (Terraform strongly preferred)
  • CI/CD tools (GitHub Actions, Azure DevOps, GitLab, etc.)
  • Data formats and protocols (Parquet, JSON, Avro)
  • Databricks Certified Data Engineer Professional
  • Databricks Certified Professional Architect (or equivalent advanced certification)
  • Strong architectural decision-making and documentation skills
  • Excellent communication with technical and non-technical stakeholders
  • Ability to lead design reviews and architecture governance forums
  • Strong troubleshooting and performance-tuning mindset

Nice To Haves

  • MLflow and MLOps architectures
  • Real-time analytics and streaming pipelines
  • Multi-region or cross-account data architectures
  • Consulting or MSP delivery experience
  • AWS Certified Solutions Architect (Associate or Professional)
  • Azure Solutions Architect Expert
  • Google Professional Data Engineer
  • Databricks Machine Learning Professional
  • Snowflake or other cloud data platform certifications

Responsibilities

  • Design enterprise Databricks Lakehouse architectures aligned with the Databricks Well-Architected Framework
  • Define reference architectures for batch, streaming, analytics, and ML workloads
  • Select and standardize cluster, compute, and workspace architectures
  • Design multi-workspace strategies (dev/test/prod, shared vs. isolated)
  • Ensure architectures meet scalability, availability, and performance requirements
  • Apply Databricks best practices across all pillars
  • Lead Databricks workspace, cluster, and Unity Catalog implementations
  • Implement Delta Lake, Delta Live Tables (DLT), and Structured Streaming
  • Build and optimize ETL/ELT pipelines using Spark and SQL
  • Integrate Databricks with cloud services (S3/ADLS/GCS, IAM, Key Vault, networking)
  • Establish CI/CD pipelines for notebooks, jobs, and infrastructure
  • Implement Unity Catalog for centralized governance
  • Define data classification, lineage, and audit strategies
  • Enforce least-privilege access and secure networking patterns
  • Support compliance requirements (HIPAA, SOC 2, PCI, GDPR as applicable)
  • Monitor platform health, performance, and cost
  • Troubleshoot production issues across jobs, clusters, and data pipelines
  • Perform workload tuning and cost-performance optimization
  • Define SLOs, alerts, and operational metrics
  • Partner with Data Engineering, Analytics, ML, Platform, and Security teams
  • Translate business requirements into technical architectures
  • Provide architectural guidance and technical mentorship
  • Communicate risks, tradeoffs, and recommendations to leadership

Stand Out From the Crowd

Upload your resume and get instant feedback on how well it matches this job.

Upload and Match Resume

What This Job Offers

Job Type

Full-time

Career Level

Mid Level

Education Level

No Education Listed

Number of Employees

1-10 employees

© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service