About The Position

A position at White Cap isn’t your ordinary job. You’ll work in an exciting and diverse environment, meet interesting people, and have a variety of career opportunities. The White Cap family is committed to Building Trust on Every Job. We do this by being deeply knowledgeable, fully capable, and always dependable, and our associates are the driving force behind this commitment. Responsible for designing and implementing high-performance data pipelines, APIs, and integrations between transactional systems and our analytics platform. The ideal candidate will have expertise in Postgres OLTP systems, Databricks-based data processing, and modern middleware technologies to enable secure and scalable data exchange.

Requirements

  • Typically requires BS/BA in a related discipline.
  • Generally, 2-5 years of experience in a related field OR MS/MA and generally 2-4 years of experience in a related field.
  • Certification is required in some areas.
  • Demonstrates skill in data analysis techniques by resolving missing/incomplete information and inconsistencies/anomalies in more complex research/data.
  • Nature of work requires increasing independence; receives guidance only on unusual, complex problems or issues.

Nice To Haves

  • Proficiency in Python or Scala, with strong SQL skills.
  • Hands-on experience with Databricks or Spark-based data engineering.
  • Experience integrating APIs, building middleware connectors, and managing event-based data flows.
  • Solid understanding of Postgres or similar OLTP databases.
  • Familiarity with cloud environments (Azure preferred) and containerization (Docker/Kubernetes).
  • Strong problem-solving, performance tuning, and communication skills.
  • Relevant certifications (e.g., Databricks Certified Data Engineer, Azure Data Engineer Associate)
  • Experience working in Agile/Scrum environments.
  • Strong documentation and technical writing skills.

Responsibilities

  • Design, build, and maintain batch and streaming data pipelines using Databricks (PySpark, Delta Live Tables, Unity Catalog).
  • Develop and manage inbound/outbound data feeds via APIs, SFTP, pub/sub, or middleware platforms.
  • Build and optimize data models in Postgres and synchronize with analytical layers.
  • Collaborate with product, architecture, and InfoSec teams to ensure secure and compliant data movement.
  • Implement data quality, observability, and governance standards.
  • Automate deployment and testing with CI/CD tools (e.g., Databricks Asset Bundles, GitHub Actions, or Azure DevOps).
  • Participate in refactoring existing data pipelines to modern, scalable approaches.
  • Aid in retirement of former techniques and communications around new methods
  • Create build vs buy proposals.
  • Implement “greenfield” solutions or integrate 3rd party apps and connectors.
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service