Using LiDAR as Camera for End-to-End Driving

dc.contributor.advisorMatiisen, Tambet, juhendaja
dc.contributor.authorvan Gent, Jan Aare
dc.contributor.otherTartu Ülikool. Loodus- ja täppisteaduste valdkondet
dc.contributor.otherTartu Ülikool. Arvutiteaduse instituutet
dc.date.accessioned2023-09-21T12:40:45Z
dc.date.available2023-09-21T12:40:45Z
dc.date.issued2021
dc.description.abstractResearch on autonomous driving has seen a growing surge in popularity in the last decade. One of the more interesting avenues of autonomous driving, known as end-toend (E2E) driving, involves training a neural network to predict control signals directly from input sensors. Usually, the main sensor used for E2E driving is a regular frontfacing camera. Cameras are the preferred sensor since they can perceive the road and traffic the way humans see it. Additionally, to make self-driving affordable, the sensor set should be relatively simple and cost-effective, which simple front-facing dashcams excel at. However, Light Detection And Ranging (LiDAR) instruments give accurate distance estimations and can be more robust to weather and lighting conditions than regular cameras. In this thesis, the feasibility of using LiDAR as a camera for E2E driving is evaluated. Specifically, the sensor examined is the Ouster OS1-128 LiDAR instrument, which can output measurements as a 360-degree raster image with range, intensity and ambience channels. A convolutional neural network (CNN) was trained to predict steering angles from LiDAR images. In addition, multiple experiments, including varying the data and the network architecture, were performed. The trained models were evaluated with both offline with open-loop metrics and online with closed-loop metrics. The evaluation results confirm that using LiDAR measurements as a raster image (instead of point cloud) allows to make use of the well-tested CNN networks for E2E driving. This means that Ouster OS1-128 lidar can be used as a drop-in replacement for the camera in E2E driving solutions, with potential improvements due to range sensing, less sensitivity to weather and lighting conditions and novel data augmentation opportunities.et
dc.identifier.urihttps://hdl.handle.net/10062/92343
dc.language.isoenget
dc.publisherTartu Ülikoolet
dc.rightsopenAccesset
dc.rightsAttribution-NonCommercial-NoDerivatives 4.0 International*
dc.rights.urihttp://creativecommons.org/licenses/by-nc-nd/4.0/*
dc.subjectComputer Visionet
dc.subjectMachine Learninget
dc.subjectDeep Learninget
dc.subject.othermagistritöödet
dc.subject.otherinformaatikaet
dc.subject.otherinfotehnoloogiaet
dc.subject.otherinformaticset
dc.subject.otherinfotechnologyet
dc.titleUsing LiDAR as Camera for End-to-End Drivinget
dc.typeThesiset

Failid

Originaal pakett

Nüüd näidatakse 1 - 1 1
Laen...
Pisipilt
Nimi:
vanGent_ComputerScience_2021.pdf
Suurus:
8.11 MB
Formaat:
Adobe Portable Document Format
Kirjeldus:

Litsentsi pakett

Nüüd näidatakse 1 - 1 1
Laen...
Pisipilt
Nimi:
license.txt
Suurus:
1.71 KB
Formaat:
Item-specific license agreed upon to submission
Kirjeldus: