Abstract

Perception is a fundamental component of any autonomous driving system. Semantic segmentation is the perception task of assigning semantic class labels to sensor inputs. While autonomous driving systems are currently equipped with a suite of sensors, much focus in the literature has been on semantic segmentation of camera images only. Research in the fusion of different sensor modalities for semantic segmentation has not been investigated as much. Deep learning models based on transformer architectures have proven successful in many tasks in computer vision and natural language processing. This work explores the use of deep learning transformers to fuse information from LiDAR and camera sensors to improve the segmentation of LiDAR point clouds. It also addresses the question of which fusion level in this deep learning framework provides better performance. This was done following an empirical approach in which different fusion models were designed and evaluated against each other using SemanticKITTI dataset.

School

School of Sciences and Engineering

Department

Computer Science & Engineering Department

Degree Name

MS in Computer Science

Graduation Date

Winter 1-31-2022

Submission Date

1-24-2022

First Advisor

Mohamed Moustafa

Committee Member 1

Amr Goneid

Committee Member 2

Mahmoud Khalil

Extent

46p.

Document Type

Master's Thesis

Institutional Review Board (IRB) Approval

Not necessary for this item

Share

COinS