Sensor Fusion for 3D Object Detection for Autonomous Vehicles

Sensor Fusion for 3D Object Detection for Autonomous Vehicles

Author: Yahya Massoud

Publisher:

Published: 2021

Total Pages:

ISBN-13:

DOWNLOAD EBOOK

Thanks to the major advancements in hardware and computational power, sensor technology, and artificial intelligence, the race for fully autonomous driving systems is heating up. With a countless number of challenging conditions and driving scenarios, researchers are tackling the most challenging problems in driverless cars. One of the most critical components is the perception module, which enables an autonomous vehicle to "see" and "understand" its surrounding environment. Given that modern vehicles can have large number of sensors and available data streams, this thesis presents a deep learning-based framework that leverages multimodal data - i.e. sensor fusion, to perform the task of 3D object detection and localization. We provide an extensive review of the advancements of deep learning-based methods in computer vision, specifically in 2D and 3D object detection tasks. We also study the progress of the literature in both single-sensor and multi-sensor data fusion techniques. Furthermore, we present an in-depth explanation of our proposed approach that performs sensor fusion using input streams from LiDAR and Camera sensors, aiming to simultaneously perform 2D, 3D, and Bird's Eye View detection. Our experiments highlight the importance of learnable data fusion mechanisms and multi-task learning, the impact of different CNN design decisions, speed-accuracy tradeoffs, and ways to deal with overfitting in multi-sensor data fusion frameworks.


3D Object Detection and Tracking for Autonomous Vehicles

3D Object Detection and Tracking for Autonomous Vehicles

Author: Su Pang

Publisher:

Published: 2022

Total Pages: 0

ISBN-13:

DOWNLOAD EBOOK

Autonomous driving systems require accurate 3D object detection and tracking to achieve reliable path planning and navigation. For object detection, there have been significant advances in neural networks for single-modality approaches. However, it has been surprisingly difficult to train networks to use multiple modalities in a way that demonstrates gain over single-modality networks. In this dissertation, we first propose three networks for Camera-LiDAR and Camera-Radar fusion. For Camera-LiDAR fusion, CLOCs (Camera-LiDAR Object Candidates fusion) and Fast-CLOCs are presented. CLOCs fusion provides a multi-modal fusion framework that significantly improves the performance of single-modality detectors. CLOCs operates on the combined output candidates before Non-Maximum Suppression (NMS) of any 2D and any 3D detector, and is trained to leverage their geometric and semantic consistencies to produce more accurate 3D detection results. Fast-CLOCs can run in near real-time with less computational requirements compared to CLOCs. Fast-CLOCs eliminates the separate heavy 2D detector, and instead uses a 3D detector-cued 2D image detector (3D-Q-2D) to reduce memory and computation. For Camera-Radar fusion, we propose TransCAR, a Transformer-based Camera-And-Radar fusion solution for 3D object detection. The cross-attention layer within the transformer decoder can adaptively learn the soft-association between the radar features and vision queries instead of hard-association based on sensor calibration only. Then, we propose to solve the 3D multiple object tracking (MOT) problem for autonomous driving applications using a random finite set-based (RFS) Multiple Measurement Models filter (RFS-M3). In particular, we propose multiple measurement models for a Poisson multi-Bernoulli mixture (PMBM) filter in support of different application scenarios. Our RFS-M3 filter can naturally model these uncertainties accurately and elegantly. We combine learning-based detections with our RFS-M3 tracker by incorporating the detection confidence score into the PMBM prediction and update step. We have evaluated our CLOCs, Fast-CLOCs and TransCAR fusion-based 3D detector and RFS-M3 3D tracker using challenging datasets including KITTI, nuScenes, Argoverse and Waymo that are released by academia and industry leaders. Superior experimental results demonstrated the effectiveness of the proposed approaches.


Multi-sensor Fusion for 3D Object Detection

Multi-sensor Fusion for 3D Object Detection

Author: Darshan Ramesh Bhanushali

Publisher:

Published: 2020

Total Pages: 58

ISBN-13:

DOWNLOAD EBOOK

"Sensing and modelling of the surrounding environment is crucial for solving many of the problems in intelligent machines like self-driving cars, autonomous robots, and augmented reality displays. Performance, reliability and safety of the autonomous agents rely heavily on the way the environment is modelled. Two-dimensional models are inadequate to capture the three-dimensional nature of real-world scenes. Three-dimensional models are necessary to achieve the standards required by the autonomy stack for intelligent agents to work alongside humans. Data driven deep learning methodologies for three-dimensional scene modelling has evolved greatly in the past few years because of the availability of huge amounts of data from variety of sensors in the form of well-designed datasets. 3D object detection and localization are two of the key requirements for tasks such as obstacle avoidance, agent-to-agent interaction, and path planning. Most methodologies for object detection work on a single sensor data like camera or LiDAR. Camera sensors provide feature rich scene data and LiDAR provides us 3D geometrical information. Advanced object detection and localization can be achieved by leveraging the information from both camera and LiDAR sensors. In order to effectively quantify the uncertainty of each sensor channel, an appropriate fusion strategy is needed to fuse the independently encoded point clouds from LiDAR with the RGB images from standard vision cameras. In this work, we introduce a fusion strategy and develop a multimodal pipeline which utilizes existing state-of-the-art deep learning based data encoders to produce robust 3D object detection and localization in real-time. The performance of the proposed fusion model is evaluated on the popular KITTI 3D benchmark dataset."--Abstract.


Robust Environmental Perception and Reliability Control for Intelligent Vehicles

Robust Environmental Perception and Reliability Control for Intelligent Vehicles

Author: Huihui Pan

Publisher: Springer Nature

Published: 2023-11-25

Total Pages: 308

ISBN-13: 9819977908

DOWNLOAD EBOOK

This book presents the most recent state-of-the-art algorithms on robust environmental perception and reliability control for intelligent vehicle systems. By integrating object detection, semantic segmentation, trajectory prediction, multi-object tracking, multi-sensor fusion, and reliability control in a systematic way, this book is aimed at guaranteeing that intelligent vehicles can run safely in complex road traffic scenes. Adopts the multi-sensor data fusion-based neural networks to environmental perception fault tolerance algorithms, solving the problem of perception reliability when some sensors fail by using data redundancy. Presents the camera-based monocular approach to implement the robust perception tasks, which introduces sequential feature association and depth hint augmentation, and introduces seven adaptive methods. Proposes efficient and robust semantic segmentation of traffic scenes through real-time deep dual-resolution networks and representation separation of vision transformers. Focuses on trajectory prediction and proposes phased and progressive trajectory prediction methods that is more consistent with human psychological characteristics, which is able to take both social interactions and personal intentions into account. Puts forward methods based on conditional random field and multi-task segmentation learning to solve the robust multi-object tracking problem for environment perception in autonomous vehicle scenarios. Presents the novel reliability control strategies of intelligent vehicles to optimize the dynamic tracking performance and investigates the completely unknown autonomous vehicle tracking issues with actuator faults.


Autonomous driving algorithms and Its IC Design

Autonomous driving algorithms and Its IC Design

Author: Jianfeng Ren

Publisher: Springer Nature

Published: 2023-08-09

Total Pages: 306

ISBN-13: 9819928974

DOWNLOAD EBOOK

With the rapid development of artificial intelligence and the emergence of various new sensors, autonomous driving has grown in popularity in recent years. The implementation of autonomous driving requires new sources of sensory data, such as cameras, radars, and lidars, and the algorithm processing requires a high degree of parallel computing. In this regard, traditional CPUs have insufficient computing power, while DSPs are good at image processing but lack sufficient performance for deep learning. Although GPUs are good at training, they are too “power-hungry,” which can affect vehicle performance. Therefore, this book looks to the future, arguing that custom ASICs are bound to become mainstream. With the goal of ICs design for autonomous driving, this book discusses the theory and engineering practice of designing future-oriented autonomous driving SoC chips. The content is divided into thirteen chapters, the first chapter mainly introduces readers to the current challenges and research directions in autonomous driving. Chapters 2–6 focus on algorithm design for perception and planning control. Chapters 7–10 address the optimization of deep learning models and the design of deep learning chips, while Chapters 11-12 cover automatic driving software architecture design. Chapter 13 discusses the 5G application on autonomous drving. This book is suitable for all undergraduates, graduate students, and engineering technicians who are interested in autonomous driving.


Real-time 3D Object Detection for Autonomous Driving

Real-time 3D Object Detection for Autonomous Driving

Author: Melissa Mozifian

Publisher:

Published: 2018

Total Pages: 72

ISBN-13:

DOWNLOAD EBOOK

This thesis focuses on advancing the state-of-the-art 3D object detection and localization in autonomous driving. An autonomous vehicle requires operating within a very unpredictable and dynamic environment. Hence a robust perception system is essential. This work proposes a novel architecture, AVOD, an \textbf{A}ggregate \textbf{V}iew \textbf{O}bject \textbf{D}etection architecture for autonomous driving capable of generating accurate 3D bounding boxes on road scenes. AVOD uses LIDAR point clouds and RGB images to generate features that are shared by two subnetworks: a region proposal network (RPN) and a second stage detector network. The proposed RPN uses a novel architecture capable of performing multimodal feature fusion on high resolution feature maps to generate reliable 3D object proposals for multiple object classes in road scenes. Using these proposals, the second stage detection network performs accurate oriented 3D bounding box regression and category classification to predict the extents, orientation, and classification of objects in 3D space. AVOD is differentiated from the state-of-the-art by using a high resolution feature extractor coupled with a multimodal fusion RPN architecture, and is therefore able to produce accurate region proposals for small classes in road scenes. AVOD also employs explicit orientation vector regression to resolve the ambiguous orientation estimate inferred from a bounding box. Experiments on the challenging KITTI dataset show the superiority of AVOD over the state-of-the-art detectors on the 3D localization, orientation estimation, and category classification tasks. Finally, AVOD is shown to run in real time and with a low memory overhead. The robustness of AVOD is also visually demonstrated when deployed on our autonomous vehicle operating under low lighting conditions such as night time as well as in snowy scenes. Furthermore, AVOD-SSD is proposed as a 3D Single Stage Detector. This work demonstrates how a single stage detector can achieve similar accuracy as that of a two-stage detector. An analysis of speed and accuracy trade-offs between AVOD and AVOD-SSD are presented.


Autonomous Driving Perception

Autonomous Driving Perception

Author: Rui Fan

Publisher: Springer Nature

Published: 2023-10-06

Total Pages: 391

ISBN-13: 981994287X

DOWNLOAD EBOOK

Discover the captivating world of computer vision and deep learning for autonomous driving with our comprehensive and in-depth guide. Immerse yourself in an in-depth exploration of cutting-edge topics, carefully crafted to engage tertiary students and ignite the curiosity of researchers and professionals in the field. From fundamental principles to practical applications, this comprehensive guide offers a gentle introduction, expert evaluations of state-of-the-art methods, and inspiring research directions. With a broad range of topics covered, it is also an invaluable resource for university programs offering computer vision and deep learning courses. This book provides clear and simplified algorithm descriptions, making it easy for beginners to understand the complex concepts. We also include carefully selected problems and examples to help reinforce your learning. Don't miss out on this essential guide to computer vision and deep learning for autonomous driving.


Point Cloud Processing for Environmental Analysis in Autonomous Driving using Deep Learning

Point Cloud Processing for Environmental Analysis in Autonomous Driving using Deep Learning

Author: Martin Simon

Publisher: BoD – Books on Demand

Published: 2023-01-01

Total Pages: 194

ISBN-13: 3863602722

DOWNLOAD EBOOK

Autonomous self-driving cars need a very precise perception system of their environment, working for every conceivable scenario. Therefore, different kinds of sensor types, such as lidar scanners, are in use. This thesis contributes highly efficient algorithms for 3D object recognition to the scientific community. It provides a Deep Neural Network with specific layers and a novel loss to safely localize and estimate the orientation of objects from point clouds originating from lidar sensors. First, a single-shot 3D object detector is developed that outputs dense predictions in only one forward pass. Next, this detector is refined by fusing complementary semantic features from cameras and joint probabilistic tracking to stabilize predictions and filter outliers. The last part presents an evaluation of data from automotive-grade lidar scanners. A Generative Adversarial Network is also being developed as an alternative for target-specific artificial data generation.


Autonomous Vehicles and Systems

Autonomous Vehicles and Systems

Author: Ishwar K. Sethi

Publisher: CRC Press

Published: 2024-02-06

Total Pages: 464

ISBN-13: 1003810675

DOWNLOAD EBOOK

This book captures multidisciplinary research encompassing various facets of autonomous vehicle systems (AVS) research and developments. The AVS field is rapidly moving towards realization with numerous advances continually reported. The contributions to this field come from widely varying branches of knowledge, making it a truly multidisciplinary area of research and development. The topics covered in the book include: AI and deep learning for AVS Autonomous steering through deep neural networks Adversarial attacks and defenses on autonomous vehicles Gesture recognition for vehicle control Multi-sensor fusion in autonomous vehicles Teleoperation technologies for AVS Simulation and game theoretic decision making for AVS Path following control system design for AVS Hybrid cloud and edge solutions for AVS Ethics of AVS