Senior Data & Integration Engineer

Parker ManagementSavannah, GA

About The Position

The Senior Data and Integration Engineer is a senior technical role responsible for architecting, developing, and optimizing enterprise data and integration solutions that enable seamless, reliable data flow across Parker’s Kitchen’s technology ecosystem. This role serves as a technical leader and subject matter expert in ETL/ELT pipeline design, API integration architecture, data warehousing strategy, and cloud-based data platforms. The ideal candidate will drive technical standards, mentor junior engineers, and work closely with business stakeholders, developers, and data teams to deliver scalable, high-quality data infrastructure that supports operational excellence and strategic decision-making.

Requirements

  • Bachelor’s degree in computer science, Information Systems, Data Engineering, or a related field.
  • 5+ years of progressive experience in data engineering, data integration, or related roles, with demonstrated ownership of enterprise-scale solutions.
  • Deep expertise in ETL/ELT tools (Azure Data Factory, Apache Airflow, SSIS, Informatica, or equivalent).
  • Advanced SQL proficiency and strong experience with database technologies (SQL Server, Snowflake, PostgreSQL).
  • Extensive hands-on experience designing and consuming APIs, web services, and integration platforms.
  • Strong experience with cloud data platforms, preferably Azure (Synapse, Data Lake Storage, Functions).
  • Proficiency in scripting languages (Python, PowerShell, or Bash) for automation and data processing.
  • Experience with DevOps practices, CI/CD pipelines, and Infrastructure as Code (Terraform, ARM templates).
  • Proven ability to lead technical initiatives, define standards, and influence architecture decisions.
  • Strong problem-solving, analytical, and communication skills with the ability to translate complex technical concepts for business stakeholders.

Nice To Haves

  • Experience in retail, convenience store, or food service technology environments.
  • Experience with data observability and pipeline monitoring tools.
  • Knowledge of event-driven architecture, message queues (Kafka, Azure Service Bus), and stream processing.
  • Experience with Snowflake administration, optimization, and Snowpark.
  • Relevant certifications (Azure Data Engineer Associate, Snowflake SnowPro, AWS Data Analytics, etc.)

Responsibilities

  • Architect, develop, and maintain complex ETL/ELT pipelines to extract, transform, and load data from multiple sources into data lakes and data warehouses, ensuring scalability and fault tolerance.
  • Define and enforce data transformation standards to ensure data quality, consistency, and performance across all pipelines.
  • Automate data ingestion and transformation processes using tools such as Azure Data Factory (ADF), Apache Airflow, SSIS, or equivalent platforms.
  • Evaluate emerging data technologies and recommend adoption strategies aligned with organizational goals.
  • Design and implement data lake and Lakehouse architectures to support analytics, reporting, and operational workloads.
  • Design, develop, and govern API-based integrations between enterprise applications, cloud platforms, and third-party services, establishing reusable integration patterns.
  • Work with RESTful and SOAP APIs, JSON, XML, webhooks, and event-driven architectures.
  • Implement and enforce authentication and security protocols (OAuth 2.0, JWT, API keys, mTLS) across all integration points.
  • Define integration architecture standards, including error handling, retry logic, rate limiting, and monitoring practices.
  • Lead integration efforts for new vendor platforms and system migrations, serving as the technical point of contact.
  • Design and optimize database structures for data warehousing, reporting, and real-time analytics solutions.
  • Write and optimize complex SQL queries, stored procedures, views, and materialized views to support data extraction and analytics workloads.
  • Work with relational databases (SQL Server, PostgreSQL, Snowflake) and NoSQL solutions, recommending the appropriate technology for each use case.
  • Develop and deploy data solutions in cloud environments (primarily Azure), leveraging cloud storage, compute services, and serverless technologies for scalable data processing.
  • Contribute to Infrastructure as Code (Terraform, ARM templates) and CI/CD pipeline development for data platform deployments.
  • Implement and champion data governance best practices, including metadata management, data lineage tracking, data cataloging, and access controls.
  • Ensure compliance with data privacy regulations (PCI DSS, state privacy laws) and organizational security policies.
  • Define and monitor data quality metrics, SLAs, and alerting thresholds for all production data pipelines.
  • Conduct security reviews of data integrations and recommend hardening measures.
  • Serve as a technical mentor and resource for peers and cross-functional teams on data engineering best practices.
  • Create and maintain technical documentation, architecture diagrams, runbooks, and process workflows.
  • Lead troubleshooting and root cause analysis for data integration failures and performance degradation.
  • Support cross-functional teams in identifying data-related issues and translating business requirements into technical solutions.
  • Participate in architecture review boards and contribute to the organization’s technology roadmap.
  • Drive continuous improvement initiatives to reduce technical debt and improve pipeline reliability and performance.
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service