Rank-based self-training for graph convolutional networks

Detalhes bibliográficos
Autor(a) principal: Pedronette, Daniel Carlos Guimarães [UNESP]
Data de Publicação: 2021
Outros Autores: Latecki, Longin Jan
Tipo de documento: Artigo
Idioma: eng
Título da fonte: Repositório Institucional da UNESP
Texto Completo: http://dx.doi.org/10.1016/j.ipm.2020.102443
http://hdl.handle.net/11449/206925
Resumo: Graph Convolutional Networks (GCNs) have been established as a fundamental approach for representation learning on graphs, based on convolution operations on non-Euclidean domain, defined by graph-structured data. GCNs and variants have achieved state-of-the-art results on classification tasks, especially in semi-supervised learning scenarios. A central challenge in semi-supervised classification consists in how to exploit the maximum of useful information encoded in the unlabeled data. In this paper, we address this issue through a novel self-training approach for improving the accuracy of GCNs on semi-supervised classification tasks. A margin score is used through a rank-based model to identify the most confident sample predictions. Such predictions are exploited as an expanded labeled set in a second-stage training step. Our model is suitable for different GCN models. Moreover, we also propose a rank aggregation of labeled sets obtained by different GCN models. The experimental evaluation considers four GCN variations and traditional benchmarks extensively used in the literature. Significant accuracy gains were achieved for all evaluated models, reaching results comparable or superior to the state-of-the-art. The best results were achieved for rank aggregation self-training on combinations of the four GCN models.
id UNSP_fcf0ec9b286839fe5753654d94391263
oai_identifier_str oai:repositorio.unesp.br:11449/206925
network_acronym_str UNSP
network_name_str Repositório Institucional da UNESP
repository_id_str 2946
spelling Rank-based self-training for graph convolutional networksGraph convolutional networksRank modelSelf-trainingSemi-supervised learningGraph Convolutional Networks (GCNs) have been established as a fundamental approach for representation learning on graphs, based on convolution operations on non-Euclidean domain, defined by graph-structured data. GCNs and variants have achieved state-of-the-art results on classification tasks, especially in semi-supervised learning scenarios. A central challenge in semi-supervised classification consists in how to exploit the maximum of useful information encoded in the unlabeled data. In this paper, we address this issue through a novel self-training approach for improving the accuracy of GCNs on semi-supervised classification tasks. A margin score is used through a rank-based model to identify the most confident sample predictions. Such predictions are exploited as an expanded labeled set in a second-stage training step. Our model is suitable for different GCN models. Moreover, we also propose a rank aggregation of labeled sets obtained by different GCN models. The experimental evaluation considers four GCN variations and traditional benchmarks extensively used in the literature. Significant accuracy gains were achieved for all evaluated models, reaching results comparable or superior to the state-of-the-art. The best results were achieved for rank aggregation self-training on combinations of the four GCN models.Microsoft ResearchFundação de Amparo à Pesquisa do Estado de São Paulo (FAPESP)Conselho Nacional de Desenvolvimento Científico e Tecnológico (CNPq)National Science FoundationDepartment of Statistics Applied Mathematics and Computing (DEMAC) São Paulo State University (UNESP)Department of Computer and Information Sciences Temple UniversityDepartment of Statistics Applied Mathematics and Computing (DEMAC) São Paulo State University (UNESP)FAPESP: #2017/25908-6FAPESP: #2018/15597-6CNPq: #308194/2017-9National Science Foundation: IIS-1814745Universidade Estadual Paulista (Unesp)Temple UniversityPedronette, Daniel Carlos Guimarães [UNESP]Latecki, Longin Jan2021-06-25T10:46:04Z2021-06-25T10:46:04Z2021-03-01info:eu-repo/semantics/publishedVersioninfo:eu-repo/semantics/articlehttp://dx.doi.org/10.1016/j.ipm.2020.102443Information Processing and Management, v. 58, n. 2, 2021.0306-4573http://hdl.handle.net/11449/20692510.1016/j.ipm.2020.1024432-s2.0-85097135780Scopusreponame:Repositório Institucional da UNESPinstname:Universidade Estadual Paulista (UNESP)instacron:UNESPengInformation Processing and Managementinfo:eu-repo/semantics/openAccess2021-10-23T15:41:33Zoai:repositorio.unesp.br:11449/206925Repositório InstitucionalPUBhttp://repositorio.unesp.br/oai/requestopendoar:29462024-08-05T17:06:36.332714Repositório Institucional da UNESP - Universidade Estadual Paulista (UNESP)false
dc.title.none.fl_str_mv Rank-based self-training for graph convolutional networks
title Rank-based self-training for graph convolutional networks
spellingShingle Rank-based self-training for graph convolutional networks
Pedronette, Daniel Carlos Guimarães [UNESP]
Graph convolutional networks
Rank model
Self-training
Semi-supervised learning
title_short Rank-based self-training for graph convolutional networks
title_full Rank-based self-training for graph convolutional networks
title_fullStr Rank-based self-training for graph convolutional networks
title_full_unstemmed Rank-based self-training for graph convolutional networks
title_sort Rank-based self-training for graph convolutional networks
author Pedronette, Daniel Carlos Guimarães [UNESP]
author_facet Pedronette, Daniel Carlos Guimarães [UNESP]
Latecki, Longin Jan
author_role author
author2 Latecki, Longin Jan
author2_role author
dc.contributor.none.fl_str_mv Universidade Estadual Paulista (Unesp)
Temple University
dc.contributor.author.fl_str_mv Pedronette, Daniel Carlos Guimarães [UNESP]
Latecki, Longin Jan
dc.subject.por.fl_str_mv Graph convolutional networks
Rank model
Self-training
Semi-supervised learning
topic Graph convolutional networks
Rank model
Self-training
Semi-supervised learning
description Graph Convolutional Networks (GCNs) have been established as a fundamental approach for representation learning on graphs, based on convolution operations on non-Euclidean domain, defined by graph-structured data. GCNs and variants have achieved state-of-the-art results on classification tasks, especially in semi-supervised learning scenarios. A central challenge in semi-supervised classification consists in how to exploit the maximum of useful information encoded in the unlabeled data. In this paper, we address this issue through a novel self-training approach for improving the accuracy of GCNs on semi-supervised classification tasks. A margin score is used through a rank-based model to identify the most confident sample predictions. Such predictions are exploited as an expanded labeled set in a second-stage training step. Our model is suitable for different GCN models. Moreover, we also propose a rank aggregation of labeled sets obtained by different GCN models. The experimental evaluation considers four GCN variations and traditional benchmarks extensively used in the literature. Significant accuracy gains were achieved for all evaluated models, reaching results comparable or superior to the state-of-the-art. The best results were achieved for rank aggregation self-training on combinations of the four GCN models.
publishDate 2021
dc.date.none.fl_str_mv 2021-06-25T10:46:04Z
2021-06-25T10:46:04Z
2021-03-01
dc.type.status.fl_str_mv info:eu-repo/semantics/publishedVersion
dc.type.driver.fl_str_mv info:eu-repo/semantics/article
format article
status_str publishedVersion
dc.identifier.uri.fl_str_mv http://dx.doi.org/10.1016/j.ipm.2020.102443
Information Processing and Management, v. 58, n. 2, 2021.
0306-4573
http://hdl.handle.net/11449/206925
10.1016/j.ipm.2020.102443
2-s2.0-85097135780
url http://dx.doi.org/10.1016/j.ipm.2020.102443
http://hdl.handle.net/11449/206925
identifier_str_mv Information Processing and Management, v. 58, n. 2, 2021.
0306-4573
10.1016/j.ipm.2020.102443
2-s2.0-85097135780
dc.language.iso.fl_str_mv eng
language eng
dc.relation.none.fl_str_mv Information Processing and Management
dc.rights.driver.fl_str_mv info:eu-repo/semantics/openAccess
eu_rights_str_mv openAccess
dc.source.none.fl_str_mv Scopus
reponame:Repositório Institucional da UNESP
instname:Universidade Estadual Paulista (UNESP)
instacron:UNESP
instname_str Universidade Estadual Paulista (UNESP)
instacron_str UNESP
institution UNESP
reponame_str Repositório Institucional da UNESP
collection Repositório Institucional da UNESP
repository.name.fl_str_mv Repositório Institucional da UNESP - Universidade Estadual Paulista (UNESP)
repository.mail.fl_str_mv
_version_ 1808128756612595712