FINE-TUNING DEEP LEARNING MODELS FOR PEDESTRIAN DETECTION
Autor(a) principal: | |
---|---|
Data de Publicação: | 2022 |
Outros Autores: | , |
Tipo de documento: | Artigo |
Idioma: | por |
Título da fonte: | Boletim de Ciências Geodésicas |
Texto Completo: | https://revistas.ufpr.br/bcg/article/view/82504 |
Resumo: | Object detection in high resolution images is a new challenge that the remote sensing community is facing thanks to introduction of unmanned aerial vehicles and monitoring cameras. One of the interests is to detect and trace persons in the images. Different from general objects, pedestrians can have different poses and are undergoing constant morphological changes while moving, this task needs an intelligent solution. Fine-tuning has woken up great interest among researchers due to its relevance for retraining convolutional networks for many and interesting applications. For object classification, detection, and segmentation fine-tuned models have shown state-of-the-art performance. In the present work, we evaluate the performance of fine-tuned models with a variation of training data by comparing Faster Region-based Convolutional Neural Network (Faster R-CNN) Inception v2, Single Shot MultiBox Detector (SSD) Inception v2, and SSD Mobilenet v2. To achieve the goal, the effect of varying training data on performance metrics such as accuracy, precision, F1-score, and recall are taken into account. After testing the detectors, it was identified that the precision and recall are more sensitive on the variation of the amount of training data. Under five variation of the amount of training data, we observe that the proportion of 60%-80% consistently achieve highly comparable performance, whereas in all variation of training data Faster R-CNN Inception v2 outperforms SSD Inception v2 and SSD Mobilenet v2 in evaluated metrics, but the SSD converges relatively quickly during the training phase. Overall, partitioning 80% of total data for fine-tuning trained models produces efficient detectors even with only 700 data samples. |
id |
UFPR-2_1ed36d95eebc49c2974aaa616099d89a |
---|---|
oai_identifier_str |
oai:revistas.ufpr.br:article/82504 |
network_acronym_str |
UFPR-2 |
network_name_str |
Boletim de Ciências Geodésicas |
repository_id_str |
|
spelling |
FINE-TUNING DEEP LEARNING MODELS FOR PEDESTRIAN DETECTIONFINE-TUNING DEEP LEARNING MODELS FOR PEDESTRIAN DETECTIONGeociências, Ciências da Terrafine-tuning; pedestrian detection; training data; deep learning models.Object detection in high resolution images is a new challenge that the remote sensing community is facing thanks to introduction of unmanned aerial vehicles and monitoring cameras. One of the interests is to detect and trace persons in the images. Different from general objects, pedestrians can have different poses and are undergoing constant morphological changes while moving, this task needs an intelligent solution. Fine-tuning has woken up great interest among researchers due to its relevance for retraining convolutional networks for many and interesting applications. For object classification, detection, and segmentation fine-tuned models have shown state-of-the-art performance. In the present work, we evaluate the performance of fine-tuned models with a variation of training data by comparing Faster Region-based Convolutional Neural Network (Faster R-CNN) Inception v2, Single Shot MultiBox Detector (SSD) Inception v2, and SSD Mobilenet v2. To achieve the goal, the effect of varying training data on performance metrics such as accuracy, precision, F1-score, and recall are taken into account. After testing the detectors, it was identified that the precision and recall are more sensitive on the variation of the amount of training data. Under five variation of the amount of training data, we observe that the proportion of 60%-80% consistently achieve highly comparable performance, whereas in all variation of training data Faster R-CNN Inception v2 outperforms SSD Inception v2 and SSD Mobilenet v2 in evaluated metrics, but the SSD converges relatively quickly during the training phase. Overall, partitioning 80% of total data for fine-tuning trained models produces efficient detectors even with only 700 data samples.Object detection in high resolution images is a new challenge that the remote sensing community is facing thanks to introduction of unmanned aerial vehicles and monitoring cameras. One of the interests is to detect and trace persons in the images. Different from general objects, pedestrians can have different poses and are undergoing constant morphological changes while moving, this task needs an intelligent solution. Fine-tuning has woken up great interest among researchers due to its relevance for retraining convolutional networks for many and interesting applications. For object classification, detection, and segmentation fine-tuned models have shown state-of-the-art performance. In the present work, we evaluate the performance of fine-tuned models with a variation of training data by comparing Faster Region-based Convolutional Neural Network (Faster R-CNN) Inception v2, Single Shot MultiBox Detector (SSD) Inception v2, and SSD Mobilenet v2. To achieve the goal, the effect of varying training data on performance metrics such as accuracy, precision, F1-score, and recall are taken into account. After testing the detectors, it was identified that the precision and recall are more sensitive on the variation of the amount of training data. Under five variation of the amount of training data, we observe that the proportion of 60%-80% consistently achieve highly comparable performance, whereas in all variation of training data Faster R-CNN Inception v2 outperforms SSD Inception v2 and SSD Mobilenet v2 in evaluated metrics, but the SSD converges relatively quickly during the training phase. Overall, partitioning 80% of total data for fine-tuning trained models produces efficient detectors even with only 700 data samples.Boletim de Ciências GeodésicasBulletin of Geodetic SciencesAmisse, CaisseJijón-Palma, Mario ErnestoCenteno, Jorge Antonio Silva2022-07-06info:eu-repo/semantics/articleinfo:eu-repo/semantics/publishedVersionapplication/pdfhttps://revistas.ufpr.br/bcg/article/view/82504Boletim de Ciências Geodésicas; Vol 27, No 2 (2021)Bulletin of Geodetic Sciences; Vol 27, No 2 (2021)1982-21701413-4853reponame:Boletim de Ciências Geodésicasinstname:Universidade Federal do Paraná (UFPR)instacron:UFPRporhttps://revistas.ufpr.br/bcg/article/view/82504/44495Copyright (c) 2021 Caisse Amisse, Mario Ernesto Jijón-Palma, Jorge Antonio Silva Centenohttp://creativecommons.org/licenses/by-nc/4.0info:eu-repo/semantics/openAccess2022-07-06T03:05:42Zoai:revistas.ufpr.br:article/82504Revistahttps://revistas.ufpr.br/bcgPUBhttps://revistas.ufpr.br/bcg/oaiqdalmolin@ufpr.br|| danielsantos@ufpr.br||qdalmolin@ufpr.br|| danielsantos@ufpr.br1982-21701413-4853opendoar:2022-07-06T03:05:42Boletim de Ciências Geodésicas - Universidade Federal do Paraná (UFPR)false |
dc.title.none.fl_str_mv |
FINE-TUNING DEEP LEARNING MODELS FOR PEDESTRIAN DETECTION FINE-TUNING DEEP LEARNING MODELS FOR PEDESTRIAN DETECTION |
title |
FINE-TUNING DEEP LEARNING MODELS FOR PEDESTRIAN DETECTION |
spellingShingle |
FINE-TUNING DEEP LEARNING MODELS FOR PEDESTRIAN DETECTION Amisse, Caisse Geociências, Ciências da Terra fine-tuning; pedestrian detection; training data; deep learning models. |
title_short |
FINE-TUNING DEEP LEARNING MODELS FOR PEDESTRIAN DETECTION |
title_full |
FINE-TUNING DEEP LEARNING MODELS FOR PEDESTRIAN DETECTION |
title_fullStr |
FINE-TUNING DEEP LEARNING MODELS FOR PEDESTRIAN DETECTION |
title_full_unstemmed |
FINE-TUNING DEEP LEARNING MODELS FOR PEDESTRIAN DETECTION |
title_sort |
FINE-TUNING DEEP LEARNING MODELS FOR PEDESTRIAN DETECTION |
author |
Amisse, Caisse |
author_facet |
Amisse, Caisse Jijón-Palma, Mario Ernesto Centeno, Jorge Antonio Silva |
author_role |
author |
author2 |
Jijón-Palma, Mario Ernesto Centeno, Jorge Antonio Silva |
author2_role |
author author |
dc.contributor.none.fl_str_mv |
|
dc.contributor.author.fl_str_mv |
Amisse, Caisse Jijón-Palma, Mario Ernesto Centeno, Jorge Antonio Silva |
dc.subject.none.fl_str_mv |
|
dc.subject.por.fl_str_mv |
Geociências, Ciências da Terra fine-tuning; pedestrian detection; training data; deep learning models. |
topic |
Geociências, Ciências da Terra fine-tuning; pedestrian detection; training data; deep learning models. |
description |
Object detection in high resolution images is a new challenge that the remote sensing community is facing thanks to introduction of unmanned aerial vehicles and monitoring cameras. One of the interests is to detect and trace persons in the images. Different from general objects, pedestrians can have different poses and are undergoing constant morphological changes while moving, this task needs an intelligent solution. Fine-tuning has woken up great interest among researchers due to its relevance for retraining convolutional networks for many and interesting applications. For object classification, detection, and segmentation fine-tuned models have shown state-of-the-art performance. In the present work, we evaluate the performance of fine-tuned models with a variation of training data by comparing Faster Region-based Convolutional Neural Network (Faster R-CNN) Inception v2, Single Shot MultiBox Detector (SSD) Inception v2, and SSD Mobilenet v2. To achieve the goal, the effect of varying training data on performance metrics such as accuracy, precision, F1-score, and recall are taken into account. After testing the detectors, it was identified that the precision and recall are more sensitive on the variation of the amount of training data. Under five variation of the amount of training data, we observe that the proportion of 60%-80% consistently achieve highly comparable performance, whereas in all variation of training data Faster R-CNN Inception v2 outperforms SSD Inception v2 and SSD Mobilenet v2 in evaluated metrics, but the SSD converges relatively quickly during the training phase. Overall, partitioning 80% of total data for fine-tuning trained models produces efficient detectors even with only 700 data samples. |
publishDate |
2022 |
dc.date.none.fl_str_mv |
2022-07-06 |
dc.type.none.fl_str_mv |
|
dc.type.driver.fl_str_mv |
info:eu-repo/semantics/article info:eu-repo/semantics/publishedVersion |
format |
article |
status_str |
publishedVersion |
dc.identifier.uri.fl_str_mv |
https://revistas.ufpr.br/bcg/article/view/82504 |
url |
https://revistas.ufpr.br/bcg/article/view/82504 |
dc.language.iso.fl_str_mv |
por |
language |
por |
dc.relation.none.fl_str_mv |
https://revistas.ufpr.br/bcg/article/view/82504/44495 |
dc.rights.driver.fl_str_mv |
Copyright (c) 2021 Caisse Amisse, Mario Ernesto Jijón-Palma, Jorge Antonio Silva Centeno http://creativecommons.org/licenses/by-nc/4.0 info:eu-repo/semantics/openAccess |
rights_invalid_str_mv |
Copyright (c) 2021 Caisse Amisse, Mario Ernesto Jijón-Palma, Jorge Antonio Silva Centeno http://creativecommons.org/licenses/by-nc/4.0 |
eu_rights_str_mv |
openAccess |
dc.format.none.fl_str_mv |
application/pdf |
dc.publisher.none.fl_str_mv |
Boletim de Ciências Geodésicas Bulletin of Geodetic Sciences |
publisher.none.fl_str_mv |
Boletim de Ciências Geodésicas Bulletin of Geodetic Sciences |
dc.source.none.fl_str_mv |
Boletim de Ciências Geodésicas; Vol 27, No 2 (2021) Bulletin of Geodetic Sciences; Vol 27, No 2 (2021) 1982-2170 1413-4853 reponame:Boletim de Ciências Geodésicas instname:Universidade Federal do Paraná (UFPR) instacron:UFPR |
instname_str |
Universidade Federal do Paraná (UFPR) |
instacron_str |
UFPR |
institution |
UFPR |
reponame_str |
Boletim de Ciências Geodésicas |
collection |
Boletim de Ciências Geodésicas |
repository.name.fl_str_mv |
Boletim de Ciências Geodésicas - Universidade Federal do Paraná (UFPR) |
repository.mail.fl_str_mv |
qdalmolin@ufpr.br|| danielsantos@ufpr.br||qdalmolin@ufpr.br|| danielsantos@ufpr.br |
_version_ |
1799771720029044736 |