A Layer-Wise Information Reinforcement Approach to Improve Learning in Deep Belief Networks

Detalhes bibliográficos
Autor(a) principal: Roder, Mateus [UNESP]
Data de Publicação: 2020
Outros Autores: Passos, Leandro A. [UNESP], Ribeiro, Luiz Carlos Felix [UNESP], Pereira, Clayton [UNESP], Papa, João Paulo [UNESP]
Tipo de documento: Artigo de conferência
Idioma: eng
Título da fonte: Repositório Institucional da UNESP
Texto Completo: http://dx.doi.org/10.1007/978-3-030-61401-0_22
http://hdl.handle.net/11449/208175
Resumo: With the advent of deep learning, the number of works proposing new methods or improving existent ones has grown exponentially in the last years. In this scenario, “very deep” models were emerging, once they were expected to extract more intrinsic and abstract features while supporting a better performance. However, such models suffer from the gradient vanishing problem, i.e., backpropagation values become too close to zero in their shallower layers, ultimately causing learning to stagnate. Such an issue was overcome in the context of convolution neural networks by creating “shortcut connections” between layers, in a so-called deep residual learning framework. Nonetheless, a very popular deep learning technique called Deep Belief Network still suffers from gradient vanishing when dealing with discriminative tasks. Therefore, this paper proposes the Residual Deep Belief Network, which considers the information reinforcement layer-by-layer to improve the feature extraction and knowledge retaining, that support better discriminative performance. Experiments conducted over three public datasets demonstrate its robustness concerning the task of binary image classification.
id UNSP_a223a87af65eb0be5aaa3a00477eae04
oai_identifier_str oai:repositorio.unesp.br:11449/208175
network_acronym_str UNSP
network_name_str Repositório Institucional da UNESP
repository_id_str 2946
spelling A Layer-Wise Information Reinforcement Approach to Improve Learning in Deep Belief NetworksDeep Belief NetworksResidual networksRestricted Boltzmann MachinesWith the advent of deep learning, the number of works proposing new methods or improving existent ones has grown exponentially in the last years. In this scenario, “very deep” models were emerging, once they were expected to extract more intrinsic and abstract features while supporting a better performance. However, such models suffer from the gradient vanishing problem, i.e., backpropagation values become too close to zero in their shallower layers, ultimately causing learning to stagnate. Such an issue was overcome in the context of convolution neural networks by creating “shortcut connections” between layers, in a so-called deep residual learning framework. Nonetheless, a very popular deep learning technique called Deep Belief Network still suffers from gradient vanishing when dealing with discriminative tasks. Therefore, this paper proposes the Residual Deep Belief Network, which considers the information reinforcement layer-by-layer to improve the feature extraction and knowledge retaining, that support better discriminative performance. Experiments conducted over three public datasets demonstrate its robustness concerning the task of binary image classification.São Paulo State University - UNESPSão Paulo State University - UNESPUniversidade Estadual Paulista (Unesp)Roder, Mateus [UNESP]Passos, Leandro A. [UNESP]Ribeiro, Luiz Carlos Felix [UNESP]Pereira, Clayton [UNESP]Papa, João Paulo [UNESP]2021-06-25T11:07:37Z2021-06-25T11:07:37Z2020-01-01info:eu-repo/semantics/publishedVersioninfo:eu-repo/semantics/conferenceObject231-241http://dx.doi.org/10.1007/978-3-030-61401-0_22Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), v. 12415 LNAI, p. 231-241.1611-33490302-9743http://hdl.handle.net/11449/20817510.1007/978-3-030-61401-0_222-s2.0-85096530062Scopusreponame:Repositório Institucional da UNESPinstname:Universidade Estadual Paulista (UNESP)instacron:UNESPengLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)info:eu-repo/semantics/openAccess2024-04-23T16:11:19Zoai:repositorio.unesp.br:11449/208175Repositório InstitucionalPUBhttp://repositorio.unesp.br/oai/requestopendoar:29462024-08-05T15:52:57.871879Repositório Institucional da UNESP - Universidade Estadual Paulista (UNESP)false
dc.title.none.fl_str_mv A Layer-Wise Information Reinforcement Approach to Improve Learning in Deep Belief Networks
title A Layer-Wise Information Reinforcement Approach to Improve Learning in Deep Belief Networks
spellingShingle A Layer-Wise Information Reinforcement Approach to Improve Learning in Deep Belief Networks
Roder, Mateus [UNESP]
Deep Belief Networks
Residual networks
Restricted Boltzmann Machines
title_short A Layer-Wise Information Reinforcement Approach to Improve Learning in Deep Belief Networks
title_full A Layer-Wise Information Reinforcement Approach to Improve Learning in Deep Belief Networks
title_fullStr A Layer-Wise Information Reinforcement Approach to Improve Learning in Deep Belief Networks
title_full_unstemmed A Layer-Wise Information Reinforcement Approach to Improve Learning in Deep Belief Networks
title_sort A Layer-Wise Information Reinforcement Approach to Improve Learning in Deep Belief Networks
author Roder, Mateus [UNESP]
author_facet Roder, Mateus [UNESP]
Passos, Leandro A. [UNESP]
Ribeiro, Luiz Carlos Felix [UNESP]
Pereira, Clayton [UNESP]
Papa, João Paulo [UNESP]
author_role author
author2 Passos, Leandro A. [UNESP]
Ribeiro, Luiz Carlos Felix [UNESP]
Pereira, Clayton [UNESP]
Papa, João Paulo [UNESP]
author2_role author
author
author
author
dc.contributor.none.fl_str_mv Universidade Estadual Paulista (Unesp)
dc.contributor.author.fl_str_mv Roder, Mateus [UNESP]
Passos, Leandro A. [UNESP]
Ribeiro, Luiz Carlos Felix [UNESP]
Pereira, Clayton [UNESP]
Papa, João Paulo [UNESP]
dc.subject.por.fl_str_mv Deep Belief Networks
Residual networks
Restricted Boltzmann Machines
topic Deep Belief Networks
Residual networks
Restricted Boltzmann Machines
description With the advent of deep learning, the number of works proposing new methods or improving existent ones has grown exponentially in the last years. In this scenario, “very deep” models were emerging, once they were expected to extract more intrinsic and abstract features while supporting a better performance. However, such models suffer from the gradient vanishing problem, i.e., backpropagation values become too close to zero in their shallower layers, ultimately causing learning to stagnate. Such an issue was overcome in the context of convolution neural networks by creating “shortcut connections” between layers, in a so-called deep residual learning framework. Nonetheless, a very popular deep learning technique called Deep Belief Network still suffers from gradient vanishing when dealing with discriminative tasks. Therefore, this paper proposes the Residual Deep Belief Network, which considers the information reinforcement layer-by-layer to improve the feature extraction and knowledge retaining, that support better discriminative performance. Experiments conducted over three public datasets demonstrate its robustness concerning the task of binary image classification.
publishDate 2020
dc.date.none.fl_str_mv 2020-01-01
2021-06-25T11:07:37Z
2021-06-25T11:07:37Z
dc.type.status.fl_str_mv info:eu-repo/semantics/publishedVersion
dc.type.driver.fl_str_mv info:eu-repo/semantics/conferenceObject
format conferenceObject
status_str publishedVersion
dc.identifier.uri.fl_str_mv http://dx.doi.org/10.1007/978-3-030-61401-0_22
Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), v. 12415 LNAI, p. 231-241.
1611-3349
0302-9743
http://hdl.handle.net/11449/208175
10.1007/978-3-030-61401-0_22
2-s2.0-85096530062
url http://dx.doi.org/10.1007/978-3-030-61401-0_22
http://hdl.handle.net/11449/208175
identifier_str_mv Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), v. 12415 LNAI, p. 231-241.
1611-3349
0302-9743
10.1007/978-3-030-61401-0_22
2-s2.0-85096530062
dc.language.iso.fl_str_mv eng
language eng
dc.relation.none.fl_str_mv Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
dc.rights.driver.fl_str_mv info:eu-repo/semantics/openAccess
eu_rights_str_mv openAccess
dc.format.none.fl_str_mv 231-241
dc.source.none.fl_str_mv Scopus
reponame:Repositório Institucional da UNESP
instname:Universidade Estadual Paulista (UNESP)
instacron:UNESP
instname_str Universidade Estadual Paulista (UNESP)
instacron_str UNESP
institution UNESP
reponame_str Repositório Institucional da UNESP
collection Repositório Institucional da UNESP
repository.name.fl_str_mv Repositório Institucional da UNESP - Universidade Estadual Paulista (UNESP)
repository.mail.fl_str_mv
_version_ 1808128578176417792