Software Engineering Intern - LLM Systems & Applied AI Engineering

A10 NetworksSan Jose, CA
32d$50 - $65Onsite

About The Position

We're looking for a Software Engineering Intern passionate about building systems that make Large Language Models (LLMs) work in the real world. You'll design and implement backend components, APIs, and pipelines that connect models to applications, enabling fine-tuning, inference, evaluation, and monitoring at scale You'll join a fast-moving team at the intersection of software engineering and AI, where every service you build helps make intelligent systems faster, safer, and more reliable. Our environment blends Go, Python, and C++ (bonus) for high-performance backend development, paired with tools and frameworks for LLM inference and experimentation This is a 12-week, full-time, on-site internship at our San Jose, California office. You'll work on high-impact projects that directly contribute to our mission, applying your technical skills to shape the future of responsible AI

Requirements

  • Currently enrolled in a Bachelor's, Master's, or PhD program in Computer Engineering or a related field in the U.S. for the full duration of the internship
  • Graduation expected between December 2026 - June 2027
  • Available for 12 weeks between May-August 2026 or June-September 2026

Nice To Haves

  • Proficiency in Go and/or Python
  • Strong understanding of software engineering fundamentals and API design
  • Exposure to LLM frameworks such as Hugging Face, vLLM, or OpenAI API
  • Interest in model fine-tuning, evaluation pipelines, and prompt optimization
  • Familiarity with inference performance concepts such as token throughput, latency, and caching
  • Curiosity about how backend systems bring AI models to life from request to response
  • Self-driven mindset and eagerness to learn across AI and systems boundaries

Responsibilities

  • Design and implement backend components to support LLM inference and evaluation
  • Integrate model endpoints into APIs and services powering real-world applications
  • Contribute to tooling for fine-tuning, model orchestration, and inference optimization
  • Build monitoring and analytics layers to track model responses, latency, and reliability
  • Collaborate with ML engineers on serving pipelines, prompt evaluation, and guardrail logic
  • Prototype and ship features that connect models to production-grade systems

Benefits

  • Hands-on experience with LLM model integration, fine-tuning, and inference systems
  • Deep understanding of how AI models are deployed, scaled, and evaluated in production
  • Mentorship from engineers building real-world AI infrastructure and safety systems
  • A collaborative, fast-paced environment where you can experiment, learn, and grow quickly

Stand Out From the Crowd

Upload your resume and get instant feedback on how well it matches this job.

Upload and Match Resume

What This Job Offers

Job Type

Full-time

Career Level

Intern

Industry

Professional, Scientific, and Technical Services

Number of Employees

251-500 employees

© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service