ISSN 1000-1239 CN 11-1777/TP

• 网络技术 •

基于多模态融合的自动驾驶感知及计算

1. (中国科学技术大学计算机科学与技术学院 合肥 230027) (yanyongz@ustc.edu.cn)
• 出版日期: 2020-09-01
• 基金资助:
科技创新2030—“新一代人工智能”重大项目(2018AAA0100500)

Multi-Modality Fusion Perception and Computing in Autonomous Driving

Zhang Yanyong, Zhang Sha, Zhang Yu, Ji Jianmin, Duan Yifan, Huang Yitong, Peng Jie, Zhang Yuxiang

1. (School of Computer Science and Technology, University of Science and Technology of China, Hefei 230027)
• Online: 2020-09-01
• Supported by:
This work was supported by the National Major Program for Technological Innovation 2030—New Generation Artificial Intelligence (2018AAA0100500).

Abstract: The goal of autonomous driving is to provide a safe, comfortable and efficient driving environment for people. In order to have wide-spread deployment of autonomous driving systems, we need to process the sensory data from multiple streams in a timely and accurate fashion. The challenges that arise are thus two-fold: leveraging the multiple sensors that are available on autonomous vehicles to boost the perception accuracy; jointly optimizing perception models and the underlying computing models to meet the real-time requirements. To address these challenges, this paper surveys the latest research on sensing and edge computing for autonomous driving and presents our own autonomous driving system, Sonic. Specifically, we propose a multi-modality perception model, ImageFusion, that combines the lidar data and camera data for 3D object detection, and a computational optimization framework, MPInfer.