Senior Data Engineer – Enterprise Data Frameworks (Java Spark Focus)

Citizens Financial GroupPhoenix, AZ
Hybrid

About The Position

The Enterprise Data Frameworks team is seeking a Senior Java focused software engineer who builds and maintains large scale data processing systems using Java, Apache Spark, and Kafka. This role is intended for experienced backend engineers with strong core Java fundamentals who apply traditional software engineering practices to data intensive platforms and distributed systems. The ideal candidate has hands on experience developing production grade Java applications using modern frameworks and IntelliJ based development workflows, paired with practical experience building Spark based processing pipelines and Kafka driven data ingestion services. In this role, you will contribute to the core components of enterprise data frameworks, working closely with senior engineers and architects while remaining deeply involved in design and implementation. This role requires strong Java engineering skills, the ability to debug and optimize complex Spark applications, and comfort operating in regulated environments where stability, data quality, and reliability are critical.

Requirements

  • 7+ years of experience as a software engineer or data engineer with strong emphasis on Java backend development
  • Strong hands on proficiency in Java, including object oriented design, debugging, and performance optimization
  • Experience building Spark applications in Java for large scale data ingestion and transformation
  • Practical experience with Apache Kafka and event driven data architectures
  • Experience developing Java based services using frameworks such as Spring Boot
  • Solid understanding of distributed systems concepts and data processing architectures
  • Familiarity with data lake technologies, columnar storage formats such as Parquet or Iceberg, and metadata driven frameworks
  • Experience using Git based workflows, CI CD pipelines, and modern development practices
  • Exposure to cloud environments, with preference for AWS based data platforms such as S3, EMR, Lambda, or API Gateway
  • Experience working in regulated or enterprise environments with strong data governance requirements

Nice To Haves

  • Experience collaborating with UI or platform teams to integrate backend data services
  • Experience with containerization and orchestration tools such as Docker or Kubernetes
  • Exposure to additional JVM or scripting languages such as Scala or Python in a data context
  • Experience with developer productivity or automation tooling

Responsibilities

  • Design, build, and maintain Java based data processing pipelines using Apache Spark for batch and streaming workloads
  • Develop and support Java based backend services and APIs that orchestrate data workflows and framework components
  • Apply solid object oriented design principles to distributed data processing systems
  • Optimize Spark applications for performance, reliability, and cost efficiency across cloud and on premises environments
  • Collaborate with architects and senior engineers on technical design and framework evolution
  • Partner with frontend and platform teams to integrate backend data services where applicable
  • Translate business and technical requirements into scalable, well structured engineering solutions
  • Participate in code reviews, contribute to shared standards, and mentor junior engineers
  • Ensure solutions align with data governance, security, and change management standards
  • Participate in Agile ceremonies and support continuous improvement of engineering practices
© 2026 Teal Labs, Inc
Privacy PolicyTerms of Service