About The Position

webAI is pioneering the future of artificial intelligence by establishing the first distributed AI infrastructure dedicated to personalized AI. We recognize the evolving demands of a data-driven society for scalability and flexibility, and we firmly believe that the future of AI lies in distributed processing at the edge, bringing computation closer to the source of data generation. Our mission is to build a future where a company's valuable data and intellectual property remain entirely private, enabling the deployment of large-scale AI models directly on standard consumer hardware without compromising the information embedded within those models. We are developing an end-to-end platform that is secure, scalable, and fully under the control of our users, empowering enterprises with AI that understands their unique business. We are a team driven by truth, ownership, tenacity, and humility, and we seek individuals who resonate with these core values and are passionate about shaping the next generation of AI. We’re looking for an On-Device Machine Learning Engineer to bring modern ML capabilities directly onto consumer hardware, specifically fast, private, and reliable. You’ll own the design, optimization, and lifecycle of models running locally (e.g., iPhone/iPad/Mac-class devices), with a sharp focus on latency, battery, thermal behavior, and real-world UX. This role sits at the intersection of ML systems, product engineering, and performance tuning, and will help power local RAG, memory, and personalized experiences without relying on the network.

Requirements

  • Strong experience shipping ML features into production, ideally including mobile / edge / consumer devices.
  • Hands-on proficiency with Core ML and/or MLX, and the practical realities of running models locally.
  • Solid understanding of quantization and optimization techniques for inference (accuracy/perf tradeoffs, calibration, benchmarking).
  • Experience building or operating retrieval systems (embedding generation, vector search/indexing, caching strategies)—especially under resource constraints.
  • Fluency in performance engineering: profiling, latency breakdowns, memory analysis, and tuning on real devices.
  • Strong software engineering fundamentals: maintainable code, testing, CI, and debugging across complex systems.

Nice To Haves

  • Experience with on-device LLMs, multimodal models, or real-time interactive ML features.
  • Familiarity with Metal / GPU compute, or performance tuning of ML workloads on Apple platforms.
  • Experience designing privacy-preserving personalization and memory (local-first data handling, encryption, retention policies).
  • Experience building developer tooling for model packaging, benchmarking, and release management.
  • Prior work on offline-first architectures, edge inference, or battery/thermal-aware scheduling.

Responsibilities

  • On-device model optimization and deployment
  • Convert, optimize, and deploy models to run efficiently on-device using Core ML and/or MLX.
  • Implement quantization strategies (e.g., 8-bit / 4-bit where applicable), compression, pruning, distillation, and other techniques to meet performance targets.
  • Profile and improve model execution across compute backends (CPU/GPU/Neural Engine where relevant), and reduce memory footprint.
  • Local RAG + memory systems
  • Build and optimize local retrieval pipelines (embeddings, indexing, caching, ranking) that work offline and under tight resource constraints.
  • Implement local memory systems (short/long-term) with careful attention to privacy, durability, and performance.
  • Collaborate with product/design to translate “memory” behavior into concrete technical architectures and measurable quality targets.
  • Model lifecycle on consumer hardware
  • Own the on-device model lifecycle: packaging, versioning, updates, rollback strategies, on-device A/B testing approaches, telemetry, and quality monitoring.
  • Build robust evaluation and regression suites that reflect real device constraints and user workflows.
  • Ensure models degrade gracefully (low-power mode, thermals, backgrounding, OS interruptions).
  • Performance, reliability, and user experience
  • Treat battery, thermal, and latency as first-class product requirements: instrument, benchmark, and optimize continuously.
  • Design inference pipelines and scheduling strategies that respect app responsiveness, animations, and UI smoothness.
  • Partner with platform engineers to integrate ML into production apps with clean APIs and stable runtime behavior.

Benefits

  • Competitive salary and performance-based incentives.
  • Comprehensive health, dental, and vision benefits package.
  • 401k Match (US-based only)
  • $200/mos Health and Wellness Stipend
  • $400/year Continuing Education Credit
  • $500/year Function Health subscription (US-based only)
  • Free parking, for in-office employees
  • Unlimited Approved PTO
  • Parental Leave for Eligible Employees
  • Supplemental Life Insurance
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service