Perception Engineer III - R&D

Autonomous SolutionsMendon, UT
2d

About The Position

At ASI , we are revolutionizing industries with state-of-the-art autonomous robotics solutions. Within the fields of agriculture, construction, landscaping, and logistics, we deliver technologies that enhance safety, productivity, and efficiency. With our core values of Simplicity , Safety , Transparency , Humility , Attention to Detail and Growth guiding everything we do, we're shaping the future of automation in dynamic markets. As a Perception Engineer III in R&D , you will develop, evaluate, and deploy advanced sensing and perception algorithms that enable autonomous vehicles to perceive and navigate complex environments. This research focused role emphasizes real time C++ algorithm development, sensor integration, field deployment, and close collaboration with project managers and cross functional engineering teams to drive capability into customer sites and experimental platforms.

Requirements

  • Bachelor's degree in Computer Science, Robotics, Electrical Engineering, or a related technical field.
  • Five or more years of experience developing perception, computer vision, or sensor processing systems, preferably in robotics or automotive domains.
  • Expert proficiency in C++, preferably C++11 or newer, for real time perception development.
  • Experience developing multithreaded embedded or Linux applications and familiarity with real time processing constraints.
  • Experience processing sensor data including LiDAR point clouds, camera images, and radar signals.
  • Proficiency with a scripting language such as Python or MATLAB for prototyping and analysis.
  • Thorough understanding of linear algebra, coordinate transforms, least squares solutions, and pose estimation.

Nice To Haves

  • Hands on experience with ROS or ROS2, and experience with DDS or other publish-subscribe middleware is preferred.
  • Experience deploying perception algorithms to embedded targets and GPUs, including profiling and optimization.

Responsibilities

  • Evaluate and deploy LiDAR, radar, cameras, ultrasonic, and other exteroceptive sensors for autonomous vehicle operation.
  • Develop real time perception algorithms in C++ (C++11 or newer) to interface with sensors, process data, and enable intelligent decision making.
  • Implement multithreaded data processing pipelines and ensure low latency performance in Linux and embedded environments.
  • Integrate sensor drivers, synchronization, and data acquisition systems across modalities for reliable timestamps and fused inputs.
  • Process point clouds, images, and radar returns for detection, classification, segmentation, and tracking.
  • Develop sensor fusion and object tracking modules that combine LiDAR, radar, and camera data for robust environment understanding.
  • Optimize perception pipelines for latency, throughput, memory usage, and robustness under field conditions and variable sensor noise.
  • Build tools and test harnesses for simulation, hardware in the loop, and field validation of perception systems.
  • Work with project managers to plan, deploy, and configure perception systems at customer sites and experimental testbeds.
  • Collaborate with robotics, controls, planning, and software teams to ensure perception outputs meet downstream requirements.
  • Profile, debug, and troubleshoot perception code on embedded targets and GPUs to find and fix performance or correctness issues.
  • Document algorithms, interfaces, calibration procedures, and deployment steps to support reproducibility and handoff.
  • Validate perception performance using datasets, simulation scenarios, and real world testing to measure accuracy, latency, and failure modes.
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service