Developing an On-Road Object Detection System Using Monovision and Radar Fusion
release_3x36h324jzeb3bgmm4zrvzf55e
by
Ya-Wen Hsu, Yi-horng Lai, Kai-Quan Zhong, Tang-Kai Yin, Jau-Woei Perng
Abstract
In this study, a millimeter-wave (MMW) radar and an onboard camera are used to develop a sensor fusion algorithm for a forward collision warning system. This study proposed integrating an MMW radar and camera to compensate for the deficiencies caused by relying on a single sensor and to improve frontal object detection rates. Density-based spatial clustering of applications with noise and particle filter algorithms are used in the radar-based object detection system to remove non-object noise and track the target object. Meanwhile, the two-stage vision recognition system can detect and recognize the objects in front of a vehicle. The detected objects include pedestrians, motorcycles, and cars. The spatial alignment uses a radial basis function neural network to learn the conversion relationship between the distance information of the MMW radar and the coordinate information in the image. Then a neural network is utilized for object matching. The sensor with a higher confidence index is selected as the system output. Finally, three kinds of scenario conditions (daytime, nighttime, and rainy-day) were designed to test the performance of the proposed method. The detection rates and the false alarm rates of proposed system were approximately 90.5% and 0.6%, respectively.
In application/xml+jats
format
Archived Files and Locations
application/pdf 10.3 MB
file_3vtdclpaafgmfdtaszoc5uhebm
|
res.mdpi.com (web) web.archive.org (webarchive) |
application/pdf 10.4 MB
file_zo2tegbzdvgmvf3nr635izqlym
|
Open Access Publication
In DOAJ
In ISSN ROAD
In Keepers Registry
ISSN-L:
1996-1073
access all versions, variants, and formats of this works (eg, pre-prints)
Crossref Metadata (via API)
Worldcat
SHERPA/RoMEO (journal policies)
wikidata.org
CORE.ac.uk
Semantic Scholar
Google Scholar