Boosting-GNN: Boosting Algorithm for Graph Networks on Imbalanced Node Classification
The graph neural network (GNN) has been widely used for graph data representation. However, the existing researches only consider the ideal balanced dataset, and the imbalanced dataset is rarely considered. Traditional methods such as resampling, reweighting, and synthetic samples that deal with imb...
Guardado en:
Autores principales: | , , , , , |
---|---|
Formato: | article |
Lenguaje: | EN |
Publicado: |
Frontiers Media S.A.
2021
|
Materias: | |
Acceso en línea: | https://doaj.org/article/3ff80d444be64e5e98263f0fcbf7e9e3 |
Etiquetas: |
Agregar Etiqueta
Sin Etiquetas, Sea el primero en etiquetar este registro!
|
id |
oai:doaj.org-article:3ff80d444be64e5e98263f0fcbf7e9e3 |
---|---|
record_format |
dspace |
spelling |
oai:doaj.org-article:3ff80d444be64e5e98263f0fcbf7e9e32021-11-30T22:28:55ZBoosting-GNN: Boosting Algorithm for Graph Networks on Imbalanced Node Classification1662-521810.3389/fnbot.2021.775688https://doaj.org/article/3ff80d444be64e5e98263f0fcbf7e9e32021-11-01T00:00:00Zhttps://www.frontiersin.org/articles/10.3389/fnbot.2021.775688/fullhttps://doaj.org/toc/1662-5218The graph neural network (GNN) has been widely used for graph data representation. However, the existing researches only consider the ideal balanced dataset, and the imbalanced dataset is rarely considered. Traditional methods such as resampling, reweighting, and synthetic samples that deal with imbalanced datasets are no longer applicable in GNN. This study proposes an ensemble model called Boosting-GNN, which uses GNNs as the base classifiers during boosting. In Boosting-GNN, higher weights are set for the training samples that are not correctly classified by the previous classifiers, thus achieving higher classification accuracy and better reliability. Besides, transfer learning is used to reduce computational cost and increase fitting ability. Experimental results indicate that the proposed Boosting-GNN model achieves better performance than graph convolutional network (GCN), GraphSAGE, graph attention network (GAT), simplifying graph convolutional networks (SGC), multi-scale graph convolution networks (N-GCN), and most advanced reweighting and resampling methods on synthetic imbalanced datasets, with an average performance improvement of 4.5%.Shuhao ShiKai QiaoShuai YangLinyuan WangJian ChenBin YanFrontiers Media S.A.articlegraph neural networkimbalanced datasetsensemble learningadaboostnode classificationNeurosciences. Biological psychiatry. NeuropsychiatryRC321-571ENFrontiers in Neurorobotics, Vol 15 (2021) |
institution |
DOAJ |
collection |
DOAJ |
language |
EN |
topic |
graph neural network imbalanced datasets ensemble learning adaboost node classification Neurosciences. Biological psychiatry. Neuropsychiatry RC321-571 |
spellingShingle |
graph neural network imbalanced datasets ensemble learning adaboost node classification Neurosciences. Biological psychiatry. Neuropsychiatry RC321-571 Shuhao Shi Kai Qiao Shuai Yang Linyuan Wang Jian Chen Bin Yan Boosting-GNN: Boosting Algorithm for Graph Networks on Imbalanced Node Classification |
description |
The graph neural network (GNN) has been widely used for graph data representation. However, the existing researches only consider the ideal balanced dataset, and the imbalanced dataset is rarely considered. Traditional methods such as resampling, reweighting, and synthetic samples that deal with imbalanced datasets are no longer applicable in GNN. This study proposes an ensemble model called Boosting-GNN, which uses GNNs as the base classifiers during boosting. In Boosting-GNN, higher weights are set for the training samples that are not correctly classified by the previous classifiers, thus achieving higher classification accuracy and better reliability. Besides, transfer learning is used to reduce computational cost and increase fitting ability. Experimental results indicate that the proposed Boosting-GNN model achieves better performance than graph convolutional network (GCN), GraphSAGE, graph attention network (GAT), simplifying graph convolutional networks (SGC), multi-scale graph convolution networks (N-GCN), and most advanced reweighting and resampling methods on synthetic imbalanced datasets, with an average performance improvement of 4.5%. |
format |
article |
author |
Shuhao Shi Kai Qiao Shuai Yang Linyuan Wang Jian Chen Bin Yan |
author_facet |
Shuhao Shi Kai Qiao Shuai Yang Linyuan Wang Jian Chen Bin Yan |
author_sort |
Shuhao Shi |
title |
Boosting-GNN: Boosting Algorithm for Graph Networks on Imbalanced Node Classification |
title_short |
Boosting-GNN: Boosting Algorithm for Graph Networks on Imbalanced Node Classification |
title_full |
Boosting-GNN: Boosting Algorithm for Graph Networks on Imbalanced Node Classification |
title_fullStr |
Boosting-GNN: Boosting Algorithm for Graph Networks on Imbalanced Node Classification |
title_full_unstemmed |
Boosting-GNN: Boosting Algorithm for Graph Networks on Imbalanced Node Classification |
title_sort |
boosting-gnn: boosting algorithm for graph networks on imbalanced node classification |
publisher |
Frontiers Media S.A. |
publishDate |
2021 |
url |
https://doaj.org/article/3ff80d444be64e5e98263f0fcbf7e9e3 |
work_keys_str_mv |
AT shuhaoshi boostinggnnboostingalgorithmforgraphnetworksonimbalancednodeclassification AT kaiqiao boostinggnnboostingalgorithmforgraphnetworksonimbalancednodeclassification AT shuaiyang boostinggnnboostingalgorithmforgraphnetworksonimbalancednodeclassification AT linyuanwang boostinggnnboostingalgorithmforgraphnetworksonimbalancednodeclassification AT jianchen boostinggnnboostingalgorithmforgraphnetworksonimbalancednodeclassification AT binyan boostinggnnboostingalgorithmforgraphnetworksonimbalancednodeclassification |
_version_ |
1718406196820443136 |