• norsk
    • English
  • norsk 
    • norsk
    • English
  • Logg inn
Vis innførsel 
  •   Hjem
  • Norges miljø- og biovitenskapelige universitet
  • Faculty of Science and Technology (RealTek)
  • Master's theses (RealTek)
  • Vis innførsel
  •   Hjem
  • Norges miljø- og biovitenskapelige universitet
  • Faculty of Science and Technology (RealTek)
  • Master's theses (RealTek)
  • Vis innførsel
JavaScript is disabled for your browser. Some features of this site may not work without it.

Deep Learning based Models for Traffic Participant and Object Classification

Sivathas, Sathuriyan
Master thesis
Thumbnail
Åpne
no.nmbu:wiseflow:7110333:59110598.pdf (28.48Mb)
Permanent lenke
https://hdl.handle.net/11250/3147971
Utgivelsesdato
2024
Metadata
Vis full innførsel
Samlinger
  • Master's theses (RealTek) [2009]
Sammendrag
Abstract

Research on deep learning models is constantly advancing, and has emerged as an important field to study. Deep learning models have a vital role in the development of self-driving cars, making it essential to thoroughly understand their performance. This thesis focuses on the performance of selected deep learning models on various datasets related to traffic objects and participants. The deep learning models evaluated in this thesis include VGG-16, ResNet-50, WideResNet-50-2, EfficientNetB0 and Vision Transformer, all explored within the context of transfer learning. These models have been further trained on three datasets based on the images from the Audi Autonomous Driving Dataset (A2D2), specifically prepared for classifying traffic participants and objects. The first dataset comprises of normal and augmented images, and is referred to as NAI dataset. The second dataset, referred to as the NANI dataset, comprises of normal, augmented and noisy mixed-class images. The third dataset consists of normal, augmented and synthetic images generated from a Deep Convolutional Generative Adversarial Network (DCGAN), and is referred to as NASI dataset. All deep learning models were trained on the NAI, and NANI datasets, while VGG-16 and Vision Transformer are the only models to be trained on the NASI dataset. Through these datasets, the impact of normal and augmented images, the impact of noisy mixed-class images and the impact of images generated by a DCGANs are evaluated. The VGG-16 model outperformed all others, and achieved consistent performance across all three datasets. The WideResNet-50-2 model also performed well on the two datasets it was evaluated on, but did not match the performance of the VGG-16 model. The Vision Transformer model also showed promising results across all datasets, particularly in terms of consistency and stability. These results indicate that deep learning models can perform effectively and consistently across diverse datasets involving traffic participant and objects, whether the images are normal, augmented, noisy, or synthetic.
 
 
 
Utgiver
Norwegian University of Life Sciences

Kontakt oss | Gi tilbakemelding

Personvernerklæring
DSpace software copyright © 2002-2019  DuraSpace

Levert av  Unit
 

 

Bla i

Hele arkivetDelarkiv og samlingerUtgivelsesdatoForfattereTitlerEmneordDokumenttyperTidsskrifterDenne samlingenUtgivelsesdatoForfattereTitlerEmneordDokumenttyperTidsskrifter

Min side

Logg inn

Statistikk

Besøksstatistikk

Kontakt oss | Gi tilbakemelding

Personvernerklæring
DSpace software copyright © 2002-2019  DuraSpace

Levert av  Unit