JPMorgan Chase & Co.-posted about 2 months ago
Full-time • Mid Level
Wilmington, DE
5,001-10,000 employees
Credit Intermediation and Related Activities

As an AWS Software Engineer III-Databricks/Python/PySpark at JPMorgan Chase within the Corporate Sector-Global Finance Team, you will be a key member of an agile team, tasked with designing and delivering cutting-edge products that are secure, stable, and scalable. Your role involves implementing essential technology solutions across diverse technical domains to support the firm's business goals effectively.

  • Developing and optimizing data pipelines and workflows to support data integration, transformation, and analysis
  • Implementing best practices for data management, ensuring data quality, security, and compliance
  • Writing secure, high-quality production code following AWS best practices, and deploying efficiently using CI/CD pipelines
  • Creating architecture and design documents for complex applications, ensuring software code meets design constraints
  • Identifying hidden issues and patterns in data to enhance coding practices and system architecture
  • Contributing to software engineering communities, promoting diversity, opportunity, inclusion, and respect within the team
  • Formal training or certification on software engineering concepts and 3+ years applied experience
  • Experience with Spark and SQL
  • Expertise in Lakehouse/Delta Lake architecture, system design, application development, testing, and ensuring operational stability
  • Strong programming skills in Python/PySpark
  • Proficient in orchestration using Airflow
  • In-depth knowledge of Big Data and data warehousing concepts
  • Proficient in SQL/SparkSQL
  • Experience with CI/CD processes
  • Thorough understanding of the Software Development Life Cycle (SDLC)
  • Solid understanding of agile methodologies, including DevOps practices, application resiliency, and security measures
  • Proven expertise in software applications and technical processes within a specialized technical domain
  • Experience in full-stack development with strong proficiency in Python
  • Experience with Databricks, and the AWS cloud ecosystem
  • Familiarity with Snowflake, Terraform and LLM
  • Exposure to cloud technologies such as AWS Glue, S3, SQS/SNS, Lambda etc.
  • Familiarity with Data Observability, Data Quality, Query Optimization & Cost Optimization
  • AWS certifications such as SAA, Associate Developer, Data Analytics Specialty, or Databricks certification
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service