Refine Your Search

Search Results

Viewing 1 to 3 of 3
Technical Paper

Drivable Area Detection and Vehicle Localization Based on Multi-Sensor Information

2020-04-14
2020-01-1027
Multi-sensor information fusion framework is the eyes for unmanned driving and Advanced Driver Assistance System (ADAS) to perceive the surrounding environment. In addition to the perception of the surrounding environment, real-time vehicle localization is also the key and difficult point of unmanned driving technology. The disappearance of high-precision GPS signal suddenly and defect of the lane line will bring much more difficult and dangerous for vehicle localization when the vehicle is on unmanned driving. In this paper, a road boundary feature extraction algorithm is proposed based on multi-sensor information fusion of automotive radar and vision to realize the auxiliary localization of vehicles. Firstly, we designed a 79GHz (78-81GHz) Ultra-Wide Band (UWB) millimeter-wave radar, which can obtain the point cloud information of road boundary features such as guardrail or green belt and so on.
Technical Paper

Lane Marking Detection for Highway Scenes based on Solid-state LiDARs

2021-12-15
2021-01-7008
Lane marking detection plays a crucial role in Autonomous Driving Systems or Advanced Driving Assistance System. Vision based lane marking detection technology has been well discussed and put into practical application. LiDAR is more stable for challenging environment compared to cameras, and with the development of LiDAR technology, price and lifetime are no longer an issue. We propose a lane marking detection algorithm based on solid-state LiDARs. First a series of data pre-processing operations were done for the solid-state LiDARs with small field of view, and the needed ground points are extracted by the RANSAC method. Then, based on the OTSU method, we propose an approach for extracting lane marking points using intensity information.
Technical Paper

Vehicle Detection Based on Deep Neural Network Combined with Radar Attention Mechanism

2020-12-29
2020-01-5171
In the autonomous driving perception task, the accuracy of target detection is an essential evaluation, especially for small targets. In this work, we propose a multi-sensor fusion neural network that combines radar and image data to improve the confidence level of the camera when detecting targets and the accuracy of the prediction box regression. The fusion network is based on the basic structure of single-shot multi-box detection (SSD). Inspired by the attention mechanism in image processing, our work incorporates the a priori knowledge of radar detection in the convolutional block attention module (CBAM), which forms a new attention mechanism module called radar convolutional block attention module (RCBAM). We add the RCBAM into the SSD target detection network to build a deep neural network fusing millimeter-wave radar and camera.
X