Rank-based self-training for graph convolutional networks
Autor(a) principal: | |
---|---|
Data de Publicação: | 2021 |
Outros Autores: | |
Tipo de documento: | Artigo |
Idioma: | eng |
Título da fonte: | Repositório Institucional da UNESP |
Texto Completo: | http://dx.doi.org/10.1016/j.ipm.2020.102443 http://hdl.handle.net/11449/206925 |
Resumo: | Graph Convolutional Networks (GCNs) have been established as a fundamental approach for representation learning on graphs, based on convolution operations on non-Euclidean domain, defined by graph-structured data. GCNs and variants have achieved state-of-the-art results on classification tasks, especially in semi-supervised learning scenarios. A central challenge in semi-supervised classification consists in how to exploit the maximum of useful information encoded in the unlabeled data. In this paper, we address this issue through a novel self-training approach for improving the accuracy of GCNs on semi-supervised classification tasks. A margin score is used through a rank-based model to identify the most confident sample predictions. Such predictions are exploited as an expanded labeled set in a second-stage training step. Our model is suitable for different GCN models. Moreover, we also propose a rank aggregation of labeled sets obtained by different GCN models. The experimental evaluation considers four GCN variations and traditional benchmarks extensively used in the literature. Significant accuracy gains were achieved for all evaluated models, reaching results comparable or superior to the state-of-the-art. The best results were achieved for rank aggregation self-training on combinations of the four GCN models. |
id |
UNSP_fcf0ec9b286839fe5753654d94391263 |
---|---|
oai_identifier_str |
oai:repositorio.unesp.br:11449/206925 |
network_acronym_str |
UNSP |
network_name_str |
Repositório Institucional da UNESP |
repository_id_str |
2946 |
spelling |
Rank-based self-training for graph convolutional networksGraph convolutional networksRank modelSelf-trainingSemi-supervised learningGraph Convolutional Networks (GCNs) have been established as a fundamental approach for representation learning on graphs, based on convolution operations on non-Euclidean domain, defined by graph-structured data. GCNs and variants have achieved state-of-the-art results on classification tasks, especially in semi-supervised learning scenarios. A central challenge in semi-supervised classification consists in how to exploit the maximum of useful information encoded in the unlabeled data. In this paper, we address this issue through a novel self-training approach for improving the accuracy of GCNs on semi-supervised classification tasks. A margin score is used through a rank-based model to identify the most confident sample predictions. Such predictions are exploited as an expanded labeled set in a second-stage training step. Our model is suitable for different GCN models. Moreover, we also propose a rank aggregation of labeled sets obtained by different GCN models. The experimental evaluation considers four GCN variations and traditional benchmarks extensively used in the literature. Significant accuracy gains were achieved for all evaluated models, reaching results comparable or superior to the state-of-the-art. The best results were achieved for rank aggregation self-training on combinations of the four GCN models.Microsoft ResearchFundação de Amparo à Pesquisa do Estado de São Paulo (FAPESP)Conselho Nacional de Desenvolvimento Científico e Tecnológico (CNPq)National Science FoundationDepartment of Statistics Applied Mathematics and Computing (DEMAC) São Paulo State University (UNESP)Department of Computer and Information Sciences Temple UniversityDepartment of Statistics Applied Mathematics and Computing (DEMAC) São Paulo State University (UNESP)FAPESP: #2017/25908-6FAPESP: #2018/15597-6CNPq: #308194/2017-9National Science Foundation: IIS-1814745Universidade Estadual Paulista (Unesp)Temple UniversityPedronette, Daniel Carlos Guimarães [UNESP]Latecki, Longin Jan2021-06-25T10:46:04Z2021-06-25T10:46:04Z2021-03-01info:eu-repo/semantics/publishedVersioninfo:eu-repo/semantics/articlehttp://dx.doi.org/10.1016/j.ipm.2020.102443Information Processing and Management, v. 58, n. 2, 2021.0306-4573http://hdl.handle.net/11449/20692510.1016/j.ipm.2020.1024432-s2.0-85097135780Scopusreponame:Repositório Institucional da UNESPinstname:Universidade Estadual Paulista (UNESP)instacron:UNESPengInformation Processing and Managementinfo:eu-repo/semantics/openAccess2021-10-23T15:41:33Zoai:repositorio.unesp.br:11449/206925Repositório InstitucionalPUBhttp://repositorio.unesp.br/oai/requestopendoar:29462024-08-05T17:06:36.332714Repositório Institucional da UNESP - Universidade Estadual Paulista (UNESP)false |
dc.title.none.fl_str_mv |
Rank-based self-training for graph convolutional networks |
title |
Rank-based self-training for graph convolutional networks |
spellingShingle |
Rank-based self-training for graph convolutional networks Pedronette, Daniel Carlos Guimarães [UNESP] Graph convolutional networks Rank model Self-training Semi-supervised learning |
title_short |
Rank-based self-training for graph convolutional networks |
title_full |
Rank-based self-training for graph convolutional networks |
title_fullStr |
Rank-based self-training for graph convolutional networks |
title_full_unstemmed |
Rank-based self-training for graph convolutional networks |
title_sort |
Rank-based self-training for graph convolutional networks |
author |
Pedronette, Daniel Carlos Guimarães [UNESP] |
author_facet |
Pedronette, Daniel Carlos Guimarães [UNESP] Latecki, Longin Jan |
author_role |
author |
author2 |
Latecki, Longin Jan |
author2_role |
author |
dc.contributor.none.fl_str_mv |
Universidade Estadual Paulista (Unesp) Temple University |
dc.contributor.author.fl_str_mv |
Pedronette, Daniel Carlos Guimarães [UNESP] Latecki, Longin Jan |
dc.subject.por.fl_str_mv |
Graph convolutional networks Rank model Self-training Semi-supervised learning |
topic |
Graph convolutional networks Rank model Self-training Semi-supervised learning |
description |
Graph Convolutional Networks (GCNs) have been established as a fundamental approach for representation learning on graphs, based on convolution operations on non-Euclidean domain, defined by graph-structured data. GCNs and variants have achieved state-of-the-art results on classification tasks, especially in semi-supervised learning scenarios. A central challenge in semi-supervised classification consists in how to exploit the maximum of useful information encoded in the unlabeled data. In this paper, we address this issue through a novel self-training approach for improving the accuracy of GCNs on semi-supervised classification tasks. A margin score is used through a rank-based model to identify the most confident sample predictions. Such predictions are exploited as an expanded labeled set in a second-stage training step. Our model is suitable for different GCN models. Moreover, we also propose a rank aggregation of labeled sets obtained by different GCN models. The experimental evaluation considers four GCN variations and traditional benchmarks extensively used in the literature. Significant accuracy gains were achieved for all evaluated models, reaching results comparable or superior to the state-of-the-art. The best results were achieved for rank aggregation self-training on combinations of the four GCN models. |
publishDate |
2021 |
dc.date.none.fl_str_mv |
2021-06-25T10:46:04Z 2021-06-25T10:46:04Z 2021-03-01 |
dc.type.status.fl_str_mv |
info:eu-repo/semantics/publishedVersion |
dc.type.driver.fl_str_mv |
info:eu-repo/semantics/article |
format |
article |
status_str |
publishedVersion |
dc.identifier.uri.fl_str_mv |
http://dx.doi.org/10.1016/j.ipm.2020.102443 Information Processing and Management, v. 58, n. 2, 2021. 0306-4573 http://hdl.handle.net/11449/206925 10.1016/j.ipm.2020.102443 2-s2.0-85097135780 |
url |
http://dx.doi.org/10.1016/j.ipm.2020.102443 http://hdl.handle.net/11449/206925 |
identifier_str_mv |
Information Processing and Management, v. 58, n. 2, 2021. 0306-4573 10.1016/j.ipm.2020.102443 2-s2.0-85097135780 |
dc.language.iso.fl_str_mv |
eng |
language |
eng |
dc.relation.none.fl_str_mv |
Information Processing and Management |
dc.rights.driver.fl_str_mv |
info:eu-repo/semantics/openAccess |
eu_rights_str_mv |
openAccess |
dc.source.none.fl_str_mv |
Scopus reponame:Repositório Institucional da UNESP instname:Universidade Estadual Paulista (UNESP) instacron:UNESP |
instname_str |
Universidade Estadual Paulista (UNESP) |
instacron_str |
UNESP |
institution |
UNESP |
reponame_str |
Repositório Institucional da UNESP |
collection |
Repositório Institucional da UNESP |
repository.name.fl_str_mv |
Repositório Institucional da UNESP - Universidade Estadual Paulista (UNESP) |
repository.mail.fl_str_mv |
|
_version_ |
1808128756612595712 |