Senior Machine Learning Engineer, LLM Compressor and Quantization

Red HatBoston, MA
66d$170,770 - $281,770

About The Position

At Red Hat we believe the future of AI is open and we are on a mission to bring the power of open-source LLMs and vLLM to every enterprise. Red Hat Inference team accelerates AI for the enterprise and brings operational simplicity to GenAI deployments. As leading developers, maintainers of the vLLM project, and inventors of state-of-the-art techniques for model quantization and sparsification, our team provides a stable platform for enterprises to build, optimize, and scale LLM deployments. As a Machine Learning Engineer focused on model optimization algorithms, you will work closely with our product and research teams to develop SOTA deep learning software. You will collaborate with our technical and research teams to develop training and deployment pipelines, implement model compression algorithms, and productize deep learning research. If you are someone who wants to contribute to solving challenging technical problems at the forefront of deep learning in the open source way, this is the role for you. Join us in shaping the future of AI.

Requirements

  • Strong understanding of machine learning and deep learning fundamentals with experience in one or more of Computer Vision, NLP, and reinforcement learning
  • Strong Python and PyTorch experience
  • Experience with tensor math libraries such as PyTorch and NumPy
  • Strong programming skills with proven experience implementing Python based machine learning solutions
  • Ability to develop and implement research ideas and algorithms
  • Experience with mathematical software, especially linear algebra
  • Strong communications skills with both technical and non-technical team members
  • BS, or MS in computer science or computer engineering or a related field. A PhD in a ML related domain is considered a plus

Responsibilities

  • Contribute to the design, development, and testing of various inference optimization algorithms in the vLLM LLM-compressor project
  • Create and manage training and deployment pipelines
  • Benchmark and evaluate different quantization approaches to determine the best performance for specific hardware and models
  • Participate in technical design discussions and provide innovative solutions to complex problems
  • Give thoughtful and prompt code reviews
  • Mentor and guide other engineers and foster a culture of continuous learning and innovation
  • Stay up-to-date with the latest advancements in quantization research

Benefits

  • Comprehensive medical, dental, and vision coverage
  • Flexible Spending Account - healthcare and dependent care
  • Health Savings Account - high deductible medical plan
  • Retirement 401(k) with employer match
  • Paid time off and holidays
  • Paid parental leave plans for all new parents
  • Leave benefits including disability, paid family medical leave, and paid military leave
  • Additional benefits including employee stock purchase plan, family planning reimbursement, tuition reimbursement, transportation expense account, employee assistance program, and more!

Stand Out From the Crowd

Upload your resume and get instant feedback on how well it matches this job.

Upload and Match Resume

What This Job Offers

Job Type

Full-time

Career Level

Mid Level

Industry

Professional, Scientific, and Technical Services

Education Level

Master's degree

Number of Employees

5,001-10,000 employees

© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service