Kaushik Singh
EECS Department
University of California, Berkeley
Technical Report No. UCB/EECS-2025-110
May 16, 2025
http://www2.eecs.berkeley.edu/Pubs/TechRpts/2025/EECS-2025-110.pdf
Autonomous racing presents a uniquely constrained yet demanding testbed for perception systems: cars compete at high speeds on fixed circuits with known boundaries, but must reliably detect and track only their opponents under stringent real-time constraints. This thesis addresses the challenge of robust multi-modal perception for autonomous racecars by developing, analyzing, and experimentally validating modular fusion architectures that leverages LiDAR, radar, and camera sensors. We begin by formulating the problem of opponent detection - estimating the two-dimensional position, orientation, and velocity of other vehicles - under assumptions of a separate localization system and a predefined track. After surveying classical and end-to-end learning approaches, we motivate a classical “early-stage” fusion pipeline based on perspective projection and extrinsic calibration, alongside a “late-stage” fusion design that independently processes each modality before combining outputs via an Extended Kalman Filter. Preliminary experiments - benchmarked against transponder-derived ground truth - evaluate positional accuracy and computational load for both fusion methods. Results demonstrate promising indications that our late-stage fusion method achieves superior robustness to misclassification and miscalibration, while maintaining real-time performance on the racecar’s onboard compute.
Advisor: S. Shankar Sastry
";
?>
BibTeX citation:
@mastersthesis{Singh:EECS-2025-110, Author = {Singh, Kaushik}, Title = {Robust Multimodal Perception Stack for High-Speed Autonomous Racecars}, School = {EECS Department, University of California, Berkeley}, Year = {2025}, Month = {May}, URL = {http://www2.eecs.berkeley.edu/Pubs/TechRpts/2025/EECS-2025-110.html}, Number = {UCB/EECS-2025-110}, Abstract = {Autonomous racing presents a uniquely constrained yet demanding testbed for perception systems: cars compete at high speeds on fixed circuits with known boundaries, but must reliably detect and track only their opponents under stringent real-time constraints. This thesis addresses the challenge of robust multi-modal perception for autonomous racecars by developing, analyzing, and experimentally validating modular fusion architectures that leverages LiDAR, radar, and camera sensors. We begin by formulating the problem of opponent detection - estimating the two-dimensional position, orientation, and velocity of other vehicles - under assumptions of a separate localization system and a predefined track. After surveying classical and end-to-end learning approaches, we motivate a classical “early-stage” fusion pipeline based on perspective projection and extrinsic calibration, alongside a “late-stage” fusion design that independently processes each modality before combining outputs via an Extended Kalman Filter. Preliminary experiments - benchmarked against transponder-derived ground truth - evaluate positional accuracy and computational load for both fusion methods. Results demonstrate promising indications that our late-stage fusion method achieves superior robustness to misclassification and miscalibration, while maintaining real-time performance on the racecar’s onboard compute.} }
EndNote citation:
%0 Thesis %A Singh, Kaushik %T Robust Multimodal Perception Stack for High-Speed Autonomous Racecars %I EECS Department, University of California, Berkeley %D 2025 %8 May 16 %@ UCB/EECS-2025-110 %U http://www2.eecs.berkeley.edu/Pubs/TechRpts/2025/EECS-2025-110.html %F Singh:EECS-2025-110