Counting and locating high-density objects using convolutional neural network

Detalhes bibliográficos
Autor(a) principal: Mauro dos Santos de Arruda
Data de Publicação: 2022
Tipo de documento: Tese
Idioma: por
Título da fonte: Repositório Institucional da UFMS
Texto Completo: https://repositorio.ufms.br/handle/123456789/5101
Resumo: Counting and locating objects are essential in different types of applications, as they allow performance improvements in the execution of manual tasks. Deep learning methods are becoming more prominent in this type of application because they can perform good object characterizations. However, challenges such as overlapping, occlusion, scale variations and high density of objects hinder the method’s performance, making this problem remains open. Such methods usually use bounding box annotations, which hinder their performance in high-density scenes with adjacent objects. To overcome these limitations, advancing the state-of-the-art, we propose a method for counting and locating objects using confidence maps. The first application allows for the definition of a method based on convolutional neural networks that receive a Multispectral image and detect objects from peaks on the confidence map. In a second application, we insert global and local context information with the Pyramid Pooling Module, to detect different scale objects. In addition we improve the successive refinement of the confidence map with multiple sigma values in the Multi-Sigma Stage phase. In the third application of the method, we propose a band selection module to work with hyperspectral images. In the fourth application, we evaluated the proposed method on high-density objects RGB images and compared it with state-of-the-art methods: YOLO, Faster R-CNN and RetinaNet. Finally, we expanded the method by proposing a two-branched architecture enabling the exchange of information between them. This improvement allows the method to simultaneously detect plants and plantation-rows in different datasets. The results described in this thesis show that the use of convolutional neural networks and confidence maps for counting and locating objects allows high performance. The contributions of this work should support significant advances in the areas of object detection and deep learning.
id UFMS_66947743013faf4ccd3e359a9648750d
oai_identifier_str oai:repositorio.ufms.br:123456789/5101
network_acronym_str UFMS
network_name_str Repositório Institucional da UFMS
repository_id_str 2124
spelling 2022-09-20T16:22:40Z2022-09-20T16:22:40Z2022https://repositorio.ufms.br/handle/123456789/5101Counting and locating objects are essential in different types of applications, as they allow performance improvements in the execution of manual tasks. Deep learning methods are becoming more prominent in this type of application because they can perform good object characterizations. However, challenges such as overlapping, occlusion, scale variations and high density of objects hinder the method’s performance, making this problem remains open. Such methods usually use bounding box annotations, which hinder their performance in high-density scenes with adjacent objects. To overcome these limitations, advancing the state-of-the-art, we propose a method for counting and locating objects using confidence maps. The first application allows for the definition of a method based on convolutional neural networks that receive a Multispectral image and detect objects from peaks on the confidence map. In a second application, we insert global and local context information with the Pyramid Pooling Module, to detect different scale objects. In addition we improve the successive refinement of the confidence map with multiple sigma values in the Multi-Sigma Stage phase. In the third application of the method, we propose a band selection module to work with hyperspectral images. In the fourth application, we evaluated the proposed method on high-density objects RGB images and compared it with state-of-the-art methods: YOLO, Faster R-CNN and RetinaNet. Finally, we expanded the method by proposing a two-branched architecture enabling the exchange of information between them. This improvement allows the method to simultaneously detect plants and plantation-rows in different datasets. The results described in this thesis show that the use of convolutional neural networks and confidence maps for counting and locating objects allows high performance. The contributions of this work should support significant advances in the areas of object detection and deep learning.Contagem e detecção automática de objetos são essenciais em diferentes tipos de aplicações pois permitem melhorias desempenhos na execução das tarefas manuais. Métodos de aprendizado profundo estão se destacando cada vez mais nesse tipo de aplicação pois conseguem realizar boas caracterizações dos objetos. Entretanto, desafios como a sobreposição, oclusão, diferentes de escalas e alta densidade de objetos atrapalham o desempenho desses métodos, fazendo com que esse problema permaneça aberto. Tais métodos normalmente usam anotações por caixas delimitadoras, o que prejudica seu desempenho em cenas de alta densidade com adjacência de objetos. Para superar tais limitações, avançando o estado da arte, nós propomos um método de contagem e detecção de objetos usando mapas de confiança. A primeira aplicação permitiu definir um método baseado em redes neurais convolucionais que recebem como entrada uma imagem multiespectral e detecta os objetos a partir de picos no mapa de confiança. Em uma segunda aplicação, nós inserimos informações de contexto global e local através do módulo PPM, para a detecção de objetos em diferentes escalas. Além disso, melhoramos o refinamento sucessivo do mapa de confiança com múltiplos valores de sigma na fase MSS. Na terceira aplicação do método, nós propomos um módulo de seleção de bandas para trabalhar com imagens hiperespectrais. Em uma quarta aplicação, nós avaliamos o método proposto em imagens RGB de alta densidade de objetos e comparamos com métodos do estado da arte: YOLO, Faster R-CNN e RetinaNet. Por último, expandimos o método propondo uma arquitetura de duas ramificações permitindo a troca de informações entre eles. Essa melhoria permite que o método detecte simultaneamente plantas e linhas de plantio em diferentes conjuntos de dados. Os resultados descritos nesta tese mostram que a utilização de redes neurais convolucionais e mapas de confiança para a detecção e contagem de objetos permite alto desempenho. As contribuições descritas aqui, devem suportar avanços significativos nas áreas de detecção de objetos e aprendizado profundo.Fundação Universidade Federal de Mato Grosso do SulUFMSBrasilDeep learning, object counting, line detectionCounting and locating high-density objects using convolutional neural networkinfo:eu-repo/semantics/publishedVersioninfo:eu-repo/semantics/doctoralThesisWesley Nunes GoncalvesMauro dos Santos de Arrudainfo:eu-repo/semantics/openAccessporreponame:Repositório Institucional da UFMSinstname:Universidade Federal de Mato Grosso do Sul (UFMS)instacron:UFMSORIGINALTESE_MAURO_UFMS.pdfTESE_MAURO_UFMS.pdfapplication/pdf12399196https://repositorio.ufms.br/bitstream/123456789/5101/-1/TESE_MAURO_UFMS.pdf11c6366be59bc7f8357ac4f2ee0ce549MD5-1123456789/51012022-09-20 12:22:43.044oai:repositorio.ufms.br:123456789/5101Repositório InstitucionalPUBhttps://repositorio.ufms.br/oai/requestri.prograd@ufms.bropendoar:21242022-09-20T16:22:43Repositório Institucional da UFMS - Universidade Federal de Mato Grosso do Sul (UFMS)false
dc.title.pt_BR.fl_str_mv Counting and locating high-density objects using convolutional neural network
title Counting and locating high-density objects using convolutional neural network
spellingShingle Counting and locating high-density objects using convolutional neural network
Mauro dos Santos de Arruda
Deep learning, object counting, line detection
title_short Counting and locating high-density objects using convolutional neural network
title_full Counting and locating high-density objects using convolutional neural network
title_fullStr Counting and locating high-density objects using convolutional neural network
title_full_unstemmed Counting and locating high-density objects using convolutional neural network
title_sort Counting and locating high-density objects using convolutional neural network
author Mauro dos Santos de Arruda
author_facet Mauro dos Santos de Arruda
author_role author
dc.contributor.advisor1.fl_str_mv Wesley Nunes Goncalves
dc.contributor.author.fl_str_mv Mauro dos Santos de Arruda
contributor_str_mv Wesley Nunes Goncalves
dc.subject.por.fl_str_mv Deep learning, object counting, line detection
topic Deep learning, object counting, line detection
description Counting and locating objects are essential in different types of applications, as they allow performance improvements in the execution of manual tasks. Deep learning methods are becoming more prominent in this type of application because they can perform good object characterizations. However, challenges such as overlapping, occlusion, scale variations and high density of objects hinder the method’s performance, making this problem remains open. Such methods usually use bounding box annotations, which hinder their performance in high-density scenes with adjacent objects. To overcome these limitations, advancing the state-of-the-art, we propose a method for counting and locating objects using confidence maps. The first application allows for the definition of a method based on convolutional neural networks that receive a Multispectral image and detect objects from peaks on the confidence map. In a second application, we insert global and local context information with the Pyramid Pooling Module, to detect different scale objects. In addition we improve the successive refinement of the confidence map with multiple sigma values in the Multi-Sigma Stage phase. In the third application of the method, we propose a band selection module to work with hyperspectral images. In the fourth application, we evaluated the proposed method on high-density objects RGB images and compared it with state-of-the-art methods: YOLO, Faster R-CNN and RetinaNet. Finally, we expanded the method by proposing a two-branched architecture enabling the exchange of information between them. This improvement allows the method to simultaneously detect plants and plantation-rows in different datasets. The results described in this thesis show that the use of convolutional neural networks and confidence maps for counting and locating objects allows high performance. The contributions of this work should support significant advances in the areas of object detection and deep learning.
publishDate 2022
dc.date.accessioned.fl_str_mv 2022-09-20T16:22:40Z
dc.date.available.fl_str_mv 2022-09-20T16:22:40Z
dc.date.issued.fl_str_mv 2022
dc.type.status.fl_str_mv info:eu-repo/semantics/publishedVersion
dc.type.driver.fl_str_mv info:eu-repo/semantics/doctoralThesis
format doctoralThesis
status_str publishedVersion
dc.identifier.uri.fl_str_mv https://repositorio.ufms.br/handle/123456789/5101
url https://repositorio.ufms.br/handle/123456789/5101
dc.language.iso.fl_str_mv por
language por
dc.rights.driver.fl_str_mv info:eu-repo/semantics/openAccess
eu_rights_str_mv openAccess
dc.publisher.none.fl_str_mv Fundação Universidade Federal de Mato Grosso do Sul
dc.publisher.initials.fl_str_mv UFMS
dc.publisher.country.fl_str_mv Brasil
publisher.none.fl_str_mv Fundação Universidade Federal de Mato Grosso do Sul
dc.source.none.fl_str_mv reponame:Repositório Institucional da UFMS
instname:Universidade Federal de Mato Grosso do Sul (UFMS)
instacron:UFMS
instname_str Universidade Federal de Mato Grosso do Sul (UFMS)
instacron_str UFMS
institution UFMS
reponame_str Repositório Institucional da UFMS
collection Repositório Institucional da UFMS
bitstream.url.fl_str_mv https://repositorio.ufms.br/bitstream/123456789/5101/-1/TESE_MAURO_UFMS.pdf
bitstream.checksum.fl_str_mv 11c6366be59bc7f8357ac4f2ee0ce549
bitstream.checksumAlgorithm.fl_str_mv MD5
repository.name.fl_str_mv Repositório Institucional da UFMS - Universidade Federal de Mato Grosso do Sul (UFMS)
repository.mail.fl_str_mv ri.prograd@ufms.br
_version_ 1807552831187582976