Autonomy Software Engineer - Mapping and Localization

Full-timeSunnyvale, CAEngineering
Posted July 28, 2025

About UCR

UCR (Under Control Robotics) builds multipurpose robots to support human workers in the world's toughest jobs—turning dangerous work from a necessity into a choice. Our work demands reliability, robustness, and readiness for the unexpected—on time, every time. We're assembling a mission-driven team focused on delivering real impact in heavy industry, from construction and mining to energy. If you're driven to build rugged, reliable products that solve real-world problems, we'd love to talk.

Position Overview

At UCR, building robots is a team sport. As a Robotics Autonomy Engineer, you'll take ownership and lead the development of autonomy systems that power our multipurpose robots across diverse and unstructured environments. You'll design, implement, and optimize cutting-edge localization, mapping, navigation, and SLAM systems—including advanced techniques such as 3D Gaussian Splatting—that enable our robots to perceive, understand, and act in the real world with confidence.

Responsibilities

  • Develop and maintain real-time mapping, localization, and navigation software for mobility robotic systems
  • Build scalable SLAM pipelines using a mix of sensors, including LiDAR, vision, and IMU
  • Implement 3D scene representations using cutting-edge techniques such as 3D Gaussian Splatting, NeRFs, and other neural or volumetric methods
  • Integrate localization and mapping modules with motion planning and control systems
  • Deploy robust autonomy stacks to on-board compute platforms and validate them in both simulation and real-world testing
  • Analyze and tune performance of perception and SLAM systems in challenging environments
  • Collaborate with mechanical, electrical, and software engineers to develop co-designed autonomy solutions
  • Write clean, modular, production-quality code with thorough documentation and testing
  • Operate and support robots during field testing and customer deployment

Requirements

  • 4+ years of experience working in robotics, autonomy, or a closely related field
  • Strong foundation in SLAM, probabilistic localization, 3D reconstruction, and navigation algorithms
  • Deep experience with C++ and Python, especially in real-time robotics or embedded systems
  • Experience building and deploying autonomy stacks using frameworks such as ROS or ROS2
  • Proven ability to develop algorithms for sensor fusion and state estimation (e.g., EKF, UKF, particle filters)
  • Hands-on experience with real robot systems—ground, legged, or aerial platforms
  • Familiarity with 3D mapping techniques including voxel grids, mesh reconstruction, and Gaussian Splatting
  • Demonstrated rapid growth and technical ownership on complex autonomy projects
  • Ability to prioritize and execute tasks in a fast-paced, dynamic environment
  • Excellent communication and collaboration skills across disciplines

Nice to Have

  • Experience with GPU-accelerated vision or perception pipelines (CUDA, TensorRT)
  • Exposure to deep learning-based SLAM, view synthesis, or scene understanding techniques
  • Experience with multirobot SLAM, loop closure, or graph optimization frameworks
  • Contributions to open-source robotics or perception libraries
  • Comfort debugging hardware/software integration in field settings
  • Experience with autonomy in unstructured or GPS-denied environments
  • Strong understanding of simulation frameworks (e.g., Gazebo, Isaac Sim, Unity Robotics)

To apply, submit your resume here or email people@ucr.bot. To increase your chances of being selected for an interview, we encourage you to include a public portfolio of your most representative work featuring your individual contributions and public demonstrations of autonomy or SLAM systems.

Apply for this Position