ADS AI Services Software Engineer

Bank of AmericaChandler, AZ
Onsite

About The Position

At Bank of America, we are guided by a common purpose to help make financial lives better through the power of every connection. We do this by driving Responsible Growth and delivering for our clients, teammates, communities and shareholders every day. Being a Great Place to Work is core to how we drive Responsible Growth. This includes our commitment to being an inclusive workplace, attracting and developing exceptional talent, supporting our teammates’ physical, emotional, and financial wellness, recognizing and rewarding performance, and how we make an impact in the communities we serve. Bank of America is committed to an in-office culture with specific requirements for office-based attendance and which allows for an appropriate level of flexibility for our teammates and businesses based on role-specific considerations. At Bank of America, you can build a successful career with opportunities to learn, grow, and make an impact. Join us! Position Summary: This role is responsible for designing, developing, and operating container‑based application platforms that support Generative AI and Large Language Model (LLM) workloads at enterprise scale. The engineer will partner closely with application developers, data scientists, and platform teams to ensure AI workloads are deployed securely, efficiently, and reliably across Kubernetes‑based environments. The successful candidate will focus on building and managing GPU‑accelerated containerized services, enabling scalable inference platforms, and supporting production‑grade AI frameworks. This role operates within a large‑scale enterprise environment and contributes to Agile delivery, DevOps automation, and continuous platform improvement.

Requirements

  • 8+ years in a technology environment with 5+ years of experience with container tools
  • Strong hands‑on experience with Kubernetes, including OpenShift, and container tools such as Docker and Podman.
  • Deep understanding of container orchestration concepts, including scheduling, networking, storage, configuration, and secrets management.
  • Experience operating container platforms supporting GPU‑accelerated workloads.
  • Proficiency in Python for developing and operationalizing AI‑driven applications.
  • Hands‑on experience with Large Language Models (LLMs) and inference‑focused frameworks, including: vLLM NVIDIA Triton Inference Server NVIDIA NeMo framework
  • Understanding of AI workload patterns, including real‑time and batch inference, scaling strategies, and high‑throughput serving.
  • Experience working in large‑scale enterprise environments with strong requirements for security, reliability, and compliance.
  • Familiarity with CI/CD pipelines and DevOps practices for containerized applications.
  • Experience contributing within Agile frameworks (Scrum, Kanban, or SAFe).
  • Working knowledge of infrastructure‑as‑code and automated deployment approaches.
  • Strong problem‑solving skills and ability to troubleshoot complex platform issues.
  • Clear, concise communication with technical and non‑technical stakeholders.
  • Ability to work effectively across engineering, infrastructure, security, and data science teams.

Nice To Haves

  • Experience operating container platforms in regulated or highly secure environments.
  • Exposure to observability tools (logging, metrics, tracing) for distributed and AI‑driven systems.
  • Experience supporting multi‑tenant platforms or shared AI inference services at scale.

Responsibilities

  • Designing, deploying, and maintaining containerized applications on Kubernetes and OpenShift platforms.
  • Supporting Generative AI inference environments, including model packaging, deployment, scaling, and performance optimization.
  • Enabling GPU‑based workloads and ensuring efficient resource utilization and isolation.
  • Collaborating with cross‑functional teams to deliver secure, resilient, and production‑ready AI platforms.
  • Contributing to CI/CD pipelines, infrastructure automation, and operational best practices.
  • Participating in Agile ceremonies and supporting iterative, high‑quality software delivery.

Stand Out From the Crowd

Upload your resume and get instant feedback on how well it matches this job.

Upload and Match Resume

What This Job Offers

Job Type

Full-time

Career Level

Mid Level

Education Level

No Education Listed

Number of Employees

5,001-10,000 employees

© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service