Position Overview:
We are looking for a highly skilled engineer to design and optimize the GPU/AI infrastructure behind our Perception & Planning stack, covering object detection, segmentation, depth estimation, and trajectory planning.
This role is technical: you will push the limits of GPU efficiency, distributed training, and real-time inference, turning state-of-the‑art research into production‑ready systems.
Responsibilities
- Architect and optimize large-scale training pipelines with advanced techniques (FSDP/ZeRO-DP, tensor/pipeline parallelism, activation checkpointing, CPU/NVMe offloading, FlashAttention, mixed precision/bfloat16, comm/comp overlap).
- Profile end‑to‑end pipelines (data → GPU kernels → inference) and eliminate bottlenecks using tools such as torch.profiler, Nsight Systems, Nsight Compute, TensorBoard Profiler, and low‑level debuggers (perf, NVTX/NCCL tracing).
- Implement performance‑critical components in CUDA/C++ (custom kernels, TensorRT plugins, efficient memory layouts).
- Tune GPU utilization, memory hierarchy (HBM, L2, shared), and communication efficiency (PCIe/NVLink/NCCL) to maximize throughput and minimize latency.
- Drive model conversion and deployment workflows (ONNX/TensorRT, mixed precision, quantization) with strict real‑time FPS requirements.
- Lead distributed training scaling and orchestration (multi‑node DDP/FSDP, NCCL tuning, experiment automation).
- Build reliability and observability into systems with low‑overhead logging, metrics, and health monitoring.
- Maintain benchmarks, profiling reports, and best‑practice documentation to guide the team.
Qualifications
- Master’s or Ph.D. in Computer Science, Electrical/Computer Engineering, or related technical discipline.
- Strong foundation in ML/CV with proven experience in GPU/AI infrastructure and performance optimization.
- Expert‑level coding in C++ and Python; ability to implement, debug, and optimize CUDA kernels.
- Hands‑on experience with GPU profiling and tuning, with a track record of improving throughput, utilization, and memory efficiency.
- Familiarity with ONNX, TensorRT, NCCL, and other performance‑oriented frameworks and libraries.
- Demonstrated success deploying real‑time inference systems on GPUs/edge devices.
- Strong problem‑solving, debugging, and performance‑analysis skills; thrives in low‑level, high‑performance system challenges.