Machine Learning Researcher -LLM Agents & Efficient Deep Learning

Huawei Technologies Canada Co., Ltd.Montreal, QC

About The Position

Huawei Canada has an immediate 12-month opening for a Machine Learning Researcher. The Noah's Ark lab, founded in 2012, is a prominent research organization focused on advancing artificial intelligence and related fields. The lab's mission is to enhance state-of-the-art research and integrate innovations into the company's products and services, including LLMs, RL, NLP, computer vision, AI theory, and Autonomous driving. This role will focus on multi-agent systems, function call models, and building scalable SLM pipelines for training, evaluation, and deployment. The researcher will develop methods for efficient training and inference, explore advances in parameter-efficient fine-tuning, memory and long-context modeling, and adapt GPU-based LLMs for CPU-based SLMs. The position involves prototyping new ideas, validating them through experimentation, optimizing inference performance, and integrating reasoning models with tool-use frameworks.

Requirements

  • MSc or PhD in Computer Science, Electrical Engineering, or related field
  • Strong publication record in top AI conferences such as NeurIPS, ICML, AAAI, ICLR, etc.
  • Strong background in deep learning and optimization
  • Experience with transformer architectures, LLM, SLM
  • Proficiency in Pytorch
  • Solid understanding of Linear algebra and probability
  • Solid understanding of Optimization algorithms (SGD, Adam, etc.)
  • Experience implementing and scaling ML systems in production or research settings.

Responsibilities

  • Build scalable SLM pipelines for training, evaluation, and deployment.
  • Develop methods for efficient training and inference, including Quantization (e.g., low-precision formats, INT8/FP8); Pruning and sparsity; Low-rank and tensor decompositions.
  • Explore advances in Parameter-efficient fine-tuning such as LoRA; Memory and long-context modeling; Adapt GPU-based LLMs for CPU-based SLMs.
  • Prototype new ideas and validate them through rigorous experimentation.
  • Optimize inference performance (latency, throughput, KV-cache efficiency).
  • Integrate reasoning models with tool-use frameworks (e.g., function calling, APIs).
© 2026 Teal Labs, Inc
Privacy PolicyTerms of Service