Sponsored

Sr. Software Engineer, ML Edge Inference Engineer

Serve Robotics North York

Job Description

Job Description

Job Description

At Serve Robotics, we’re reimagining how things move in cities. Our personable sidewalk robot is our vision for the future. It’s designed to take deliveries away from congested streets, make deliveries available to more people, and benefit local businesses.

The Serve fleet has been delighting merchants, customers, and pedestrians along the way in Los Angeles, Miami, Dallas, Atlanta and Chicago while doing commercial deliveries. We’re looking for talented individuals who will grow robotic deliveries from surprising novelty to efficient ubiquity.

Who We Are

We are tech industry veterans in software, hardware, and design who are pooling our skills to build the future we want to live in. We are solving real-world problems leveraging robotics, machine learning and computer vision, among other disciplines, with a mindful eye towards the end-to-end user experience. Our team is agile, diverse, and driven. We believe that the best way to solve complicated dynamic problems is collaboratively and respectfully.

We are seeking a highly skilled Sr. Software Engineer, ML Edge Inference Engineer Role to join our robotics team. This technical role bridges the gap between ML research and real-time deployment, enabling advanced ML models to run efficiently on edge hardware such as NVIDIA Jetson platforms. You will work closely with ML researchers, embedded systems engineers, and robotics software teams to ensure that state-of-the-art models can be deployed with optimal performance on robotic platforms.

Responsibilities
  • Own the full lifecycle of ML model deployment on robots—from handoff by the ML team to full system integration.

  • Convert, optimize, and integrate trained models (e.g., PyTorch/ONNX/TensorRT) for Jetson platforms using NVIDIA tools.

  • Develop and optimize CUDA kernels and pipelines for low-latency, high-throughput model inference.

  • Profile and benchmark existing ML workloads using tools like Nsight, nvprof, and TensorRT profiler.

  • Identify and remove compute and memory bottlenecks for real-time inference.

  • Design and implement strategies for quantization, pruning, and other model compression techniques suited for edge inference.

  • Ensure models are robust to the resource constraints of real-time, low-power robotic systems.

  • Manage memory layout, concurrency, and scheduling for optimized GPU and CPU usage on Jetson devices.

  • Build benchmarking pipelines for continuous performance evaluation on hardware-in-the-loop systems.

  • Collaborate with QA and systems teams to validate model behavior in field scenarios.

  • Work closely with ML researchers to influence model architectures for edge deployability and provide technical guidance on the feasibility of real-time ML models in the robotics stack.

Qualifications
  • Bachelor’s degree in Computer Science, Robotics, Electrical Engineering, or equivalent field.

  • 5+ years experience in deploying ML models on embedded or edge platforms (preferably robotics).

  • 3+ years of experience with CUDA, TensorRT, and other NVIDIA acceleration tools.

  • Proficient in Python and C++, especially for performance-sensitive systems.

  • Experience with NVIDIA Jetson (e.g., Xavier, Orin) and edge inference tools.

  • Familiarity with model conversion workflows (e.g., PyTorch → ONNX → TensorRT).

What Makes You Standout
  • Master’s degree in Computer Science, Robotics, Electrical Engineering, or equivalent field.

  • Experience with real-time robotics systems (e.g., ROS2, middleware, safety-critical constraints and linux embedded systems).

  • Knowledge of performance tuning under thermal, power, and memory constraints on embedded devices.

  • Experience with model quantization (e.g., INT8), sparsity, and latency-aware model design.

  • Contributions to open-source ML or CUDA projects is a plus.

* Please note: The base salary range listed in this job description reflects compensation for candidates based in the San Francisco Bay Area. While we prefer candidates located in the Bay Area, we are also open to qualified talent working remotely across the United States. Base salary range (U.S. – all locations): $180,000 – $205,000

Compensation Range: $190K - $240K

How to Apply

Ready to start your career as a Sr. Software Engineer, ML Edge Inference Engineer at Serve Robotics?

  1. Click the "Apply Now" button below.
  2. Review the safety warning in the modal.
  3. You will be redirected to the employer's official portal to complete your application.
  4. Ensure your resume and cover letter are tailored to the job description using our AI tools.

Frequently Asked Questions

Who is hiring?

This role is with Serve Robotics in North York.

Is this a remote position?

This appears to be an on-site role in North York.

What is the hiring process?

After you click "Apply Now", you will be redirected to the employer's official site to submit your resume. You can typically expect to hear back within 1-2 weeks if shortlisted.

How can I improve my application?

Tailor your resume to the specific job description. You can use our free Resume Analyzer to see how well you match the requirements.

What skills are needed?

Refer to the "Job Description" section above for a detailed list of required and preferred qualifications.

Sponsored

Safety & Disclaimer

External Application

You are leaving Hiringgg.com to apply on the employer's website.

Safety Tip: Never provide bank details, credit card info, or pay any fees to apply for a job.