A Layer-Wise Information Reinforcement Approach to Improve Learning in Deep Belief Networks
Autor(a) principal: | |
---|---|
Data de Publicação: | 2020 |
Outros Autores: | , , , |
Tipo de documento: | Artigo de conferência |
Idioma: | eng |
Título da fonte: | Repositório Institucional da UNESP |
Texto Completo: | http://dx.doi.org/10.1007/978-3-030-61401-0_22 http://hdl.handle.net/11449/208175 |
Resumo: | With the advent of deep learning, the number of works proposing new methods or improving existent ones has grown exponentially in the last years. In this scenario, “very deep” models were emerging, once they were expected to extract more intrinsic and abstract features while supporting a better performance. However, such models suffer from the gradient vanishing problem, i.e., backpropagation values become too close to zero in their shallower layers, ultimately causing learning to stagnate. Such an issue was overcome in the context of convolution neural networks by creating “shortcut connections” between layers, in a so-called deep residual learning framework. Nonetheless, a very popular deep learning technique called Deep Belief Network still suffers from gradient vanishing when dealing with discriminative tasks. Therefore, this paper proposes the Residual Deep Belief Network, which considers the information reinforcement layer-by-layer to improve the feature extraction and knowledge retaining, that support better discriminative performance. Experiments conducted over three public datasets demonstrate its robustness concerning the task of binary image classification. |
id |
UNSP_a223a87af65eb0be5aaa3a00477eae04 |
---|---|
oai_identifier_str |
oai:repositorio.unesp.br:11449/208175 |
network_acronym_str |
UNSP |
network_name_str |
Repositório Institucional da UNESP |
repository_id_str |
2946 |
spelling |
A Layer-Wise Information Reinforcement Approach to Improve Learning in Deep Belief NetworksDeep Belief NetworksResidual networksRestricted Boltzmann MachinesWith the advent of deep learning, the number of works proposing new methods or improving existent ones has grown exponentially in the last years. In this scenario, “very deep” models were emerging, once they were expected to extract more intrinsic and abstract features while supporting a better performance. However, such models suffer from the gradient vanishing problem, i.e., backpropagation values become too close to zero in their shallower layers, ultimately causing learning to stagnate. Such an issue was overcome in the context of convolution neural networks by creating “shortcut connections” between layers, in a so-called deep residual learning framework. Nonetheless, a very popular deep learning technique called Deep Belief Network still suffers from gradient vanishing when dealing with discriminative tasks. Therefore, this paper proposes the Residual Deep Belief Network, which considers the information reinforcement layer-by-layer to improve the feature extraction and knowledge retaining, that support better discriminative performance. Experiments conducted over three public datasets demonstrate its robustness concerning the task of binary image classification.São Paulo State University - UNESPSão Paulo State University - UNESPUniversidade Estadual Paulista (Unesp)Roder, Mateus [UNESP]Passos, Leandro A. [UNESP]Ribeiro, Luiz Carlos Felix [UNESP]Pereira, Clayton [UNESP]Papa, João Paulo [UNESP]2021-06-25T11:07:37Z2021-06-25T11:07:37Z2020-01-01info:eu-repo/semantics/publishedVersioninfo:eu-repo/semantics/conferenceObject231-241http://dx.doi.org/10.1007/978-3-030-61401-0_22Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), v. 12415 LNAI, p. 231-241.1611-33490302-9743http://hdl.handle.net/11449/20817510.1007/978-3-030-61401-0_222-s2.0-85096530062Scopusreponame:Repositório Institucional da UNESPinstname:Universidade Estadual Paulista (UNESP)instacron:UNESPengLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)info:eu-repo/semantics/openAccess2024-04-23T16:11:19Zoai:repositorio.unesp.br:11449/208175Repositório InstitucionalPUBhttp://repositorio.unesp.br/oai/requestopendoar:29462024-08-05T15:52:57.871879Repositório Institucional da UNESP - Universidade Estadual Paulista (UNESP)false |
dc.title.none.fl_str_mv |
A Layer-Wise Information Reinforcement Approach to Improve Learning in Deep Belief Networks |
title |
A Layer-Wise Information Reinforcement Approach to Improve Learning in Deep Belief Networks |
spellingShingle |
A Layer-Wise Information Reinforcement Approach to Improve Learning in Deep Belief Networks Roder, Mateus [UNESP] Deep Belief Networks Residual networks Restricted Boltzmann Machines |
title_short |
A Layer-Wise Information Reinforcement Approach to Improve Learning in Deep Belief Networks |
title_full |
A Layer-Wise Information Reinforcement Approach to Improve Learning in Deep Belief Networks |
title_fullStr |
A Layer-Wise Information Reinforcement Approach to Improve Learning in Deep Belief Networks |
title_full_unstemmed |
A Layer-Wise Information Reinforcement Approach to Improve Learning in Deep Belief Networks |
title_sort |
A Layer-Wise Information Reinforcement Approach to Improve Learning in Deep Belief Networks |
author |
Roder, Mateus [UNESP] |
author_facet |
Roder, Mateus [UNESP] Passos, Leandro A. [UNESP] Ribeiro, Luiz Carlos Felix [UNESP] Pereira, Clayton [UNESP] Papa, João Paulo [UNESP] |
author_role |
author |
author2 |
Passos, Leandro A. [UNESP] Ribeiro, Luiz Carlos Felix [UNESP] Pereira, Clayton [UNESP] Papa, João Paulo [UNESP] |
author2_role |
author author author author |
dc.contributor.none.fl_str_mv |
Universidade Estadual Paulista (Unesp) |
dc.contributor.author.fl_str_mv |
Roder, Mateus [UNESP] Passos, Leandro A. [UNESP] Ribeiro, Luiz Carlos Felix [UNESP] Pereira, Clayton [UNESP] Papa, João Paulo [UNESP] |
dc.subject.por.fl_str_mv |
Deep Belief Networks Residual networks Restricted Boltzmann Machines |
topic |
Deep Belief Networks Residual networks Restricted Boltzmann Machines |
description |
With the advent of deep learning, the number of works proposing new methods or improving existent ones has grown exponentially in the last years. In this scenario, “very deep” models were emerging, once they were expected to extract more intrinsic and abstract features while supporting a better performance. However, such models suffer from the gradient vanishing problem, i.e., backpropagation values become too close to zero in their shallower layers, ultimately causing learning to stagnate. Such an issue was overcome in the context of convolution neural networks by creating “shortcut connections” between layers, in a so-called deep residual learning framework. Nonetheless, a very popular deep learning technique called Deep Belief Network still suffers from gradient vanishing when dealing with discriminative tasks. Therefore, this paper proposes the Residual Deep Belief Network, which considers the information reinforcement layer-by-layer to improve the feature extraction and knowledge retaining, that support better discriminative performance. Experiments conducted over three public datasets demonstrate its robustness concerning the task of binary image classification. |
publishDate |
2020 |
dc.date.none.fl_str_mv |
2020-01-01 2021-06-25T11:07:37Z 2021-06-25T11:07:37Z |
dc.type.status.fl_str_mv |
info:eu-repo/semantics/publishedVersion |
dc.type.driver.fl_str_mv |
info:eu-repo/semantics/conferenceObject |
format |
conferenceObject |
status_str |
publishedVersion |
dc.identifier.uri.fl_str_mv |
http://dx.doi.org/10.1007/978-3-030-61401-0_22 Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), v. 12415 LNAI, p. 231-241. 1611-3349 0302-9743 http://hdl.handle.net/11449/208175 10.1007/978-3-030-61401-0_22 2-s2.0-85096530062 |
url |
http://dx.doi.org/10.1007/978-3-030-61401-0_22 http://hdl.handle.net/11449/208175 |
identifier_str_mv |
Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), v. 12415 LNAI, p. 231-241. 1611-3349 0302-9743 10.1007/978-3-030-61401-0_22 2-s2.0-85096530062 |
dc.language.iso.fl_str_mv |
eng |
language |
eng |
dc.relation.none.fl_str_mv |
Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) |
dc.rights.driver.fl_str_mv |
info:eu-repo/semantics/openAccess |
eu_rights_str_mv |
openAccess |
dc.format.none.fl_str_mv |
231-241 |
dc.source.none.fl_str_mv |
Scopus reponame:Repositório Institucional da UNESP instname:Universidade Estadual Paulista (UNESP) instacron:UNESP |
instname_str |
Universidade Estadual Paulista (UNESP) |
instacron_str |
UNESP |
institution |
UNESP |
reponame_str |
Repositório Institucional da UNESP |
collection |
Repositório Institucional da UNESP |
repository.name.fl_str_mv |
Repositório Institucional da UNESP - Universidade Estadual Paulista (UNESP) |
repository.mail.fl_str_mv |
|
_version_ |
1808128578176417792 |