FINE-TUNING DEEP LEARNING MODELS FOR PEDESTRIAN DETECTION

Detalhes bibliográficos
Autor(a) principal: Amisse, Caisse
Data de Publicação: 2022
Outros Autores: Jijón-Palma, Mario Ernesto, Centeno, Jorge Antonio Silva
Tipo de documento: Artigo
Idioma: por
Título da fonte: Boletim de Ciências Geodésicas
Texto Completo: https://revistas.ufpr.br/bcg/article/view/82504
Resumo: Object detection in high resolution images is a new challenge that the remote sensing community is facing thanks to introduction of unmanned aerial vehicles and monitoring cameras. One of the interests is to detect and trace persons in the images. Different from general objects, pedestrians can have different poses and are undergoing constant morphological changes while moving, this task needs an intelligent solution. Fine-tuning has woken up great interest among researchers due to its relevance for retraining convolutional networks for many and interesting applications. For object classification, detection, and segmentation fine-tuned models have shown state-of-the-art performance. In the present work, we evaluate the performance of fine-tuned models with a variation of training data by comparing Faster Region-based Convolutional Neural Network (Faster R-CNN) Inception v2, Single Shot MultiBox Detector (SSD) Inception v2, and SSD Mobilenet v2. To achieve the goal, the effect of varying training data on performance metrics such as accuracy, precision, F1-score, and recall are taken into account. After testing the detectors, it was identified that the precision and recall are more sensitive on the variation of the amount of training data. Under five variation of the amount of training data, we observe that the proportion of 60%-80% consistently achieve highly comparable performance, whereas in all variation of training data Faster R-CNN Inception v2 outperforms SSD Inception v2 and SSD Mobilenet v2 in evaluated metrics, but the SSD converges relatively quickly during the training phase. Overall, partitioning 80% of total data for fine-tuning trained models produces efficient detectors even with only 700 data samples.
id UFPR-2_1ed36d95eebc49c2974aaa616099d89a
oai_identifier_str oai:revistas.ufpr.br:article/82504
network_acronym_str UFPR-2
network_name_str Boletim de Ciências Geodésicas
repository_id_str
spelling FINE-TUNING DEEP LEARNING MODELS FOR PEDESTRIAN DETECTIONFINE-TUNING DEEP LEARNING MODELS FOR PEDESTRIAN DETECTIONGeociências, Ciências da Terrafine-tuning; pedestrian detection; training data; deep learning models.Object detection in high resolution images is a new challenge that the remote sensing community is facing thanks to introduction of unmanned aerial vehicles and monitoring cameras. One of the interests is to detect and trace persons in the images. Different from general objects, pedestrians can have different poses and are undergoing constant morphological changes while moving, this task needs an intelligent solution. Fine-tuning has woken up great interest among researchers due to its relevance for retraining convolutional networks for many and interesting applications. For object classification, detection, and segmentation fine-tuned models have shown state-of-the-art performance. In the present work, we evaluate the performance of fine-tuned models with a variation of training data by comparing Faster Region-based Convolutional Neural Network (Faster R-CNN) Inception v2, Single Shot MultiBox Detector (SSD) Inception v2, and SSD Mobilenet v2. To achieve the goal, the effect of varying training data on performance metrics such as accuracy, precision, F1-score, and recall are taken into account. After testing the detectors, it was identified that the precision and recall are more sensitive on the variation of the amount of training data. Under five variation of the amount of training data, we observe that the proportion of 60%-80% consistently achieve highly comparable performance, whereas in all variation of training data Faster R-CNN Inception v2 outperforms SSD Inception v2 and SSD Mobilenet v2 in evaluated metrics, but the SSD converges relatively quickly during the training phase. Overall, partitioning 80% of total data for fine-tuning trained models produces efficient detectors even with only 700 data samples.Object detection in high resolution images is a new challenge that the remote sensing community is facing thanks to introduction of unmanned aerial vehicles and monitoring cameras. One of the interests is to detect and trace persons in the images. Different from general objects, pedestrians can have different poses and are undergoing constant morphological changes while moving, this task needs an intelligent solution. Fine-tuning has woken up great interest among researchers due to its relevance for retraining convolutional networks for many and interesting applications. For object classification, detection, and segmentation fine-tuned models have shown state-of-the-art performance. In the present work, we evaluate the performance of fine-tuned models with a variation of training data by comparing Faster Region-based Convolutional Neural Network (Faster R-CNN) Inception v2, Single Shot MultiBox Detector (SSD) Inception v2, and SSD Mobilenet v2. To achieve the goal, the effect of varying training data on performance metrics such as accuracy, precision, F1-score, and recall are taken into account. After testing the detectors, it was identified that the precision and recall are more sensitive on the variation of the amount of training data. Under five variation of the amount of training data, we observe that the proportion of 60%-80% consistently achieve highly comparable performance, whereas in all variation of training data Faster R-CNN Inception v2 outperforms SSD Inception v2 and SSD Mobilenet v2 in evaluated metrics, but the SSD converges relatively quickly during the training phase. Overall, partitioning 80% of total data for fine-tuning trained models produces efficient detectors even with only 700 data samples.Boletim de Ciências GeodésicasBulletin of Geodetic SciencesAmisse, CaisseJijón-Palma, Mario ErnestoCenteno, Jorge Antonio Silva2022-07-06info:eu-repo/semantics/articleinfo:eu-repo/semantics/publishedVersionapplication/pdfhttps://revistas.ufpr.br/bcg/article/view/82504Boletim de Ciências Geodésicas; Vol 27, No 2 (2021)Bulletin of Geodetic Sciences; Vol 27, No 2 (2021)1982-21701413-4853reponame:Boletim de Ciências Geodésicasinstname:Universidade Federal do Paraná (UFPR)instacron:UFPRporhttps://revistas.ufpr.br/bcg/article/view/82504/44495Copyright (c) 2021 Caisse Amisse, Mario Ernesto Jijón-Palma, Jorge Antonio Silva Centenohttp://creativecommons.org/licenses/by-nc/4.0info:eu-repo/semantics/openAccess2022-07-06T03:05:42Zoai:revistas.ufpr.br:article/82504Revistahttps://revistas.ufpr.br/bcgPUBhttps://revistas.ufpr.br/bcg/oaiqdalmolin@ufpr.br|| danielsantos@ufpr.br||qdalmolin@ufpr.br|| danielsantos@ufpr.br1982-21701413-4853opendoar:2022-07-06T03:05:42Boletim de Ciências Geodésicas - Universidade Federal do Paraná (UFPR)false
dc.title.none.fl_str_mv FINE-TUNING DEEP LEARNING MODELS FOR PEDESTRIAN DETECTION
FINE-TUNING DEEP LEARNING MODELS FOR PEDESTRIAN DETECTION
title FINE-TUNING DEEP LEARNING MODELS FOR PEDESTRIAN DETECTION
spellingShingle FINE-TUNING DEEP LEARNING MODELS FOR PEDESTRIAN DETECTION
Amisse, Caisse
Geociências, Ciências da Terra
fine-tuning; pedestrian detection; training data; deep learning models.
title_short FINE-TUNING DEEP LEARNING MODELS FOR PEDESTRIAN DETECTION
title_full FINE-TUNING DEEP LEARNING MODELS FOR PEDESTRIAN DETECTION
title_fullStr FINE-TUNING DEEP LEARNING MODELS FOR PEDESTRIAN DETECTION
title_full_unstemmed FINE-TUNING DEEP LEARNING MODELS FOR PEDESTRIAN DETECTION
title_sort FINE-TUNING DEEP LEARNING MODELS FOR PEDESTRIAN DETECTION
author Amisse, Caisse
author_facet Amisse, Caisse
Jijón-Palma, Mario Ernesto
Centeno, Jorge Antonio Silva
author_role author
author2 Jijón-Palma, Mario Ernesto
Centeno, Jorge Antonio Silva
author2_role author
author
dc.contributor.none.fl_str_mv

dc.contributor.author.fl_str_mv Amisse, Caisse
Jijón-Palma, Mario Ernesto
Centeno, Jorge Antonio Silva
dc.subject.none.fl_str_mv

dc.subject.por.fl_str_mv Geociências, Ciências da Terra
fine-tuning; pedestrian detection; training data; deep learning models.
topic Geociências, Ciências da Terra
fine-tuning; pedestrian detection; training data; deep learning models.
description Object detection in high resolution images is a new challenge that the remote sensing community is facing thanks to introduction of unmanned aerial vehicles and monitoring cameras. One of the interests is to detect and trace persons in the images. Different from general objects, pedestrians can have different poses and are undergoing constant morphological changes while moving, this task needs an intelligent solution. Fine-tuning has woken up great interest among researchers due to its relevance for retraining convolutional networks for many and interesting applications. For object classification, detection, and segmentation fine-tuned models have shown state-of-the-art performance. In the present work, we evaluate the performance of fine-tuned models with a variation of training data by comparing Faster Region-based Convolutional Neural Network (Faster R-CNN) Inception v2, Single Shot MultiBox Detector (SSD) Inception v2, and SSD Mobilenet v2. To achieve the goal, the effect of varying training data on performance metrics such as accuracy, precision, F1-score, and recall are taken into account. After testing the detectors, it was identified that the precision and recall are more sensitive on the variation of the amount of training data. Under five variation of the amount of training data, we observe that the proportion of 60%-80% consistently achieve highly comparable performance, whereas in all variation of training data Faster R-CNN Inception v2 outperforms SSD Inception v2 and SSD Mobilenet v2 in evaluated metrics, but the SSD converges relatively quickly during the training phase. Overall, partitioning 80% of total data for fine-tuning trained models produces efficient detectors even with only 700 data samples.
publishDate 2022
dc.date.none.fl_str_mv 2022-07-06
dc.type.none.fl_str_mv

dc.type.driver.fl_str_mv info:eu-repo/semantics/article
info:eu-repo/semantics/publishedVersion
format article
status_str publishedVersion
dc.identifier.uri.fl_str_mv https://revistas.ufpr.br/bcg/article/view/82504
url https://revistas.ufpr.br/bcg/article/view/82504
dc.language.iso.fl_str_mv por
language por
dc.relation.none.fl_str_mv https://revistas.ufpr.br/bcg/article/view/82504/44495
dc.rights.driver.fl_str_mv Copyright (c) 2021 Caisse Amisse, Mario Ernesto Jijón-Palma, Jorge Antonio Silva Centeno
http://creativecommons.org/licenses/by-nc/4.0
info:eu-repo/semantics/openAccess
rights_invalid_str_mv Copyright (c) 2021 Caisse Amisse, Mario Ernesto Jijón-Palma, Jorge Antonio Silva Centeno
http://creativecommons.org/licenses/by-nc/4.0
eu_rights_str_mv openAccess
dc.format.none.fl_str_mv application/pdf
dc.publisher.none.fl_str_mv Boletim de Ciências Geodésicas
Bulletin of Geodetic Sciences
publisher.none.fl_str_mv Boletim de Ciências Geodésicas
Bulletin of Geodetic Sciences
dc.source.none.fl_str_mv Boletim de Ciências Geodésicas; Vol 27, No 2 (2021)
Bulletin of Geodetic Sciences; Vol 27, No 2 (2021)
1982-2170
1413-4853
reponame:Boletim de Ciências Geodésicas
instname:Universidade Federal do Paraná (UFPR)
instacron:UFPR
instname_str Universidade Federal do Paraná (UFPR)
instacron_str UFPR
institution UFPR
reponame_str Boletim de Ciências Geodésicas
collection Boletim de Ciências Geodésicas
repository.name.fl_str_mv Boletim de Ciências Geodésicas - Universidade Federal do Paraná (UFPR)
repository.mail.fl_str_mv qdalmolin@ufpr.br|| danielsantos@ufpr.br||qdalmolin@ufpr.br|| danielsantos@ufpr.br
_version_ 1799771720029044736