Senior PySpark Developer - Vice President

CitiTampa, FL
$113,840 - $170,760Onsite

About The Position

We are seeking a highly skilled and experienced Senior PySpark Developer to join our dynamic technology team. This role requires an individual with deep expertise in Python, PySpark, Big Data technologies, and SQL, coupled with a strong ability to work independently and contribute significantly to complex data engineering initiatives. The ideal candidate will have a proven track record in designing, developing, and optimizing scalable data solutions, with experience in ETL processes and a keen interest in leveraging the latest technologies. Domain knowledge in Finance will be a significant advantage, enabling the candidate to contribute to critical financial crime compliance projects. This role offers a challenging and rewarding opportunity for a seasoned developer to make a significant impact on critical data initiatives within Citi's global technology landscape, contributing to robust and compliant financial systems. This job description provides a high-level review of the types of work performed. Other job-related duties may be assigned as required.

Requirements

  • 10+ years of experience in Applications Development, Systems Analysis, or equivalent senior engineering roles.
  • Extensive hands‑on experience delivering enterprise‑scale, database‑driven platforms in a regulated environment.
  • Expert-level proficiency in Python programming, including object-oriented design, data structures, algorithms, and extensive experience with various Python libraries.
  • Deep expertise in developing, optimizing, and deploying PySpark applications for large-scale data processing, ETL, and real-time analytics on distributed systems (e.g., Spark SQL, Spark Streaming, DataFrames).
  • Strong understanding of Apache Spark architecture, Hadoop ecosystem, and experience with distributed computing concepts.
  • Familiarity with big data storage formats (e.g., Parquet, ORC).
  • Solid experience with both relational databases (e.g., Oracle) and NoSQL databases (e.g., MongoDB).
  • Strong SQL writing and optimization skills.
  • Experience in designing, developing, and consuming RESTful APIs using Python frameworks (e.g., Flask, FastAPI, Django REST Framework).
  • Strong understanding and practical experience with CI/CD tools (e.g., Jenkins) and containerization technologies (Docker, Kubernetes).
  • Expert-level proficiency with Git.
  • Experience with unit testing (e.g., Pytest), integration testing, and performance testing frameworks for Python and PySpark applications.
  • Bachelor’s degree/University degree or equivalent experience

Nice To Haves

  • Domain knowledge in Finance will be a significant advantage.
  • Experience with various ETL tools and methodologies is a plus.
  • Exposure to at least one major cloud provider (AWS, Azure, or GCP), specifically with their compute, storage, and data services (e.g., S3, ADLS, EMR, Databricks, Azure Synapse) preferred.
  • Exposure to or direct experience with Artificial Intelligence (AI) and Machine Learning (ML) concepts, frameworks (e.g., TensorFlow, PyTorch), or relevant projects is a significant advantage.
  • Master’s degree preferred.

Responsibilities

  • Design, develop, and implement robust, scalable, and high-performance data pipelines and applications using Python, PySpark, and Big Data technologies.
  • Work autonomously to analyze requirements, propose technical solutions, and deliver high-quality code and data products, ensuring alignment with architectural standards and business objectives.
  • Utilize expertise in various Big Data platforms (e.g., Hadoop, Hive, Kafka, Spark) to process, transform, and manage large datasets efficiently.
  • Write complex SQL queries, stored procedures, and optimize database performance for large-scale data warehousing and analytics solutions.
  • Develop and enhance ETL (Extract, Transform, Load) processes, ensuring data quality, integrity, and timely delivery.
  • Proactively research, evaluate, and integrate new and emerging technologies, frameworks, and tools to improve development processes and solution capabilities.
  • Ensure adherence to coding standards, conduct thorough code reviews, and implement best practices for software development, data governance, and security.
  • Diagnose and resolve complex technical issues related to data pipelines, performance bottlenecks, and system integrations in a fast-paced environment.
  • Collaborate effectively with cross-functional teams including architects, data scientists, business analysts, and QA engineers.
  • Provide technical guidance and mentorship to junior team members.
  • Identify opportunities to use AI tools to speed up development, code reviews, unit testing and deployment.

Benefits

  • medical, dental & vision coverage
  • 401(k)
  • life, accident, and disability insurance
  • wellness programs
  • paid time off packages, including planned time off (vacation), unplanned time off (sick leave), and paid holidays.
© 2026 Teal Labs, Inc
Privacy PolicyTerms of Service