Software Engineer [Multiple Positions Available]

JPMorganChasePlano, TX
5hOnsite

About The Position

DESCRIPTION: Duties: Lead a team of cloud data engineers and architects to deliver scalable, cloud-native data lakes and real-time analytics platforms on AWS, Databricks, and Immuta. Drive performance management by setting goals, conducting evaluations, and mentoring team members to enhance their skills. Establish data engineering best practices, including cloud adoption frameworks and automation strategies. Manage budgeting and cost optimization for cloud data platforms using FinOps strategies, AWS Cost Explorer, and right-sizing techniques. Define observability and reliability goals by implementing SLIs, SLOs, and automation for monitoring and self- healing architectures. Lead vendor and tool evaluations to select scalable, cost-effective technologies. Oversee infrastructure planning and governance, ensuring compliance with security and data privacy regulations through ABAC and fine-grained permissions with Immuta. Collaborate with data science, analytics, security, and DevOps teams to align data solutions with enterprise objectives. Lead the development of operational processes for data ingestion, transformation, governance, and consumption. Define the technical roadmap and data strategy to drive innovation and cost optimization. Architect and optimize cloud-based data pipeline using Databricks, Starburst, Iceberg, and Snowflake. Ensure reliability and performance of data processing workflows with Apache Spark, Spark Streaming, Delta Live Tables, and AWS Kinesis. Lead CI/CD and automation initiatives using Jenkins, GitHub Actions, Terraform, and Databricks Asset Bundles. Provide oversight in ML pipeline automation with MLflow and model lifecycle management. Enhance security and compliance with data protection frameworks and automated governance. Optimize federated query execution with Starburst and Starburst Stargate. QUALIFICATIONS: Minimum education and experience required: Bachelor's degree in Electrical and Electronic Engineering, Computer Science, Computer Engineering, or related field of study plus seven (7) years of experience in the job offered or as Software Engineer, Architect, Site Reliability Engineer, Applications Support, Oracle Apps Systems Engineer, or related occupation. Skills Required: This position requires experience with the following: leading the design and development of data-driven applications using Python, PySpark, Shell scripts, and PL/SQL to ensure scalable deployment on the AWS cloud platform; managing cloud infrastructure with Terraform, adhering to Infrastructure as Code (IaC) principles and optimizing AWS resources; driving solutions leveraging AWS services including EC2, EKS (Kubernetes), ECS, S3, EMR, Lake Formation, Glue Catalog, Glue Crawlers, Lambda, and Step Functions to build high-performing data platforms; designing and implementing modern data lake solutions with technologies including Databricks, Immuta, Iceberg, and Snowflake to ensure scalability, security, and governance; developing attribute-based access control (ABAC) using Immuta for fine-grained, policy-driven access across AWS and on-premises data lakes; integrating Databricks, Starburst, and Immuta with Tableau and Alteryx for secure, governed data consumption; creating real-time streaming solutions using orchestration tools, data ingestion frameworks, Delta Live Tables, and Spark Streaming for scalable analytics; designing monitoring and observability solutions with service level indicator (SLI)-based and service level objective (SLO)- based alerting, telemetry collection, and anomaly detection using CloudWatch, CloudTrail, Grafana, OpenTelemetry, and Splunk; shaping AI and ML workflows with MLflow for feature engineering, model training, validation, and deployment to align with business; automating deployment pipelines using Jenkins, GitHub Actions, Databricks Asset Bundles, and Terraform for efficient application rollouts; and troubleshooting and resolving performance, reliability, and scalability issues, collaborating with engineers to implement best practices for distributed systems. Job Location: 8181 Communications Pkwy, Plano, TX 75024. This position requires up to 10% domestic travel to JPMC facilities.

Requirements

  • Bachelor's degree in Electrical and Electronic Engineering, Computer Science, Computer Engineering, or related field of study plus seven (7) years of experience in the job offered or as Software Engineer, Architect, Site Reliability Engineer, Applications Support, Oracle Apps Systems Engineer, or related occupation.
  • Experience with leading the design and development of data-driven applications using Python, PySpark, Shell scripts, and PL/SQL to ensure scalable deployment on the AWS cloud platform
  • Experience with managing cloud infrastructure with Terraform, adhering to Infrastructure as Code (IaC) principles and optimizing AWS resources
  • Experience with driving solutions leveraging AWS services including EC2, EKS (Kubernetes), ECS, S3, EMR, Lake Formation, Glue Catalog, Glue Crawlers, Lambda, and Step Functions to build high-performing data platforms
  • Experience with designing and implementing modern data lake solutions with technologies including Databricks, Immuta, Iceberg, and Snowflake to ensure scalability, security, and governance
  • Experience with developing attribute-based access control (ABAC) using Immuta for fine-grained, policy-driven access across AWS and on-premises data lakes
  • Experience with integrating Databricks, Starburst, and Immuta with Tableau and Alteryx for secure, governed data consumption
  • Experience with creating real-time streaming solutions using orchestration tools, data ingestion frameworks, Delta Live Tables, and Spark Streaming for scalable analytics
  • Experience with designing monitoring and observability solutions with service level indicator (SLI)-based and service level objective (SLO)- based alerting, telemetry collection, and anomaly detection using CloudWatch, CloudTrail, Grafana, OpenTelemetry, and Splunk
  • Experience with shaping AI and ML workflows with MLflow for feature engineering, model training, validation, and deployment to align with business
  • Experience with automating deployment pipelines using Jenkins, GitHub Actions, Databricks Asset Bundles, and Terraform for efficient application rollouts
  • Experience with troubleshooting and resolving performance, reliability, and scalability issues, collaborating with engineers to implement best practices for distributed systems

Responsibilities

  • Lead a team of cloud data engineers and architects to deliver scalable, cloud-native data lakes and real-time analytics platforms on AWS, Databricks, and Immuta.
  • Drive performance management by setting goals, conducting evaluations, and mentoring team members to enhance their skills.
  • Establish data engineering best practices, including cloud adoption frameworks and automation strategies.
  • Manage budgeting and cost optimization for cloud data platforms using FinOps strategies, AWS Cost Explorer, and right-sizing techniques.
  • Define observability and reliability goals by implementing SLIs, SLOs, and automation for monitoring and self- healing architectures.
  • Lead vendor and tool evaluations to select scalable, cost-effective technologies.
  • Oversee infrastructure planning and governance, ensuring compliance with security and data privacy regulations through ABAC and fine-grained permissions with Immuta.
  • Collaborate with data science, analytics, security, and DevOps teams to align data solutions with enterprise objectives.
  • Lead the development of operational processes for data ingestion, transformation, governance, and consumption.
  • Define the technical roadmap and data strategy to drive innovation and cost optimization.
  • Architect and optimize cloud-based data pipeline using Databricks, Starburst, Iceberg, and Snowflake.
  • Ensure reliability and performance of data processing workflows with Apache Spark, Spark Streaming, Delta Live Tables, and AWS Kinesis.
  • Lead CI/CD and automation initiatives using Jenkins, GitHub Actions, Terraform, and Databricks Asset Bundles.
  • Provide oversight in ML pipeline automation with MLflow and model lifecycle management.
  • Enhance security and compliance with data protection frameworks and automated governance.
  • Optimize federated query execution with Starburst and Starburst Stargate.
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service