United We Stand, Divided We Fall: Networks to Graph (N2G) Abstraction for Robust Graph Classification Under Graph Label Corruption

Zhiwei Zhen, Yuzhou Chen, Murat Kantarcioglu, Kangkook Jee, Yulia Gel
Proceedings of the Second Learning on Graphs Conference, PMLR 231:30:1-30:19, 2024.

Abstract

Nowadays, graph neural networks (GNN) are the primary machinery to tackle (semi)-supervised graph classification tasks. The aim here is to predict classes for unlabeled graphs, given a collection of graphs with known labels. However, in many real-world applications the available information on graph classes may be distorted either due to incorrect labelling process (e.g., as in biochemistry and bioinformatics) or may be subject to targeted attacks (e.g., as in network-based customer attrition analytics). Over the past few years, the increasing number of studies has indicated that GNNs are prone both to noisy node and noisy graph labels, and while this problem has received noticeable attention for node classification tasks, vulnerability of GNNs for graph classification with perturbed graph labels still remains in its nascence. We hypothesise that this challenge can be addressed by the universal principle {\it United We Stand, Divided We Fall}. In particular, most GNNs view each graph as a standalone entity and, as a result, are limited in their abilities to account for complex interdependencies among the graphs. Inspired by the recent studies on molecular graph learning, we propose a new robust knowledge representation called {\it Networks to Graph} (N2G). The key N2G idea is to construct a new abstraction where each graph in the collection is now represented by a node, while an edge then reflects some sort of similarity among the graphs. As a result, the graph classification task can be then naturally reformulated as a node classification problem. We show that the proposed N2G representation approach does not only improve classification performance both in binary and multi-class scenarios but also substantially enhances robustness against noisy labels in the training data, leading to relative robustness gains up to 11.7\% on social network benchmarks and up to 25.8\% on bioinformatics graph benchmarks under 10\% of graph label corruption rate.

Cite this Paper


BibTeX
@InProceedings{pmlr-v231-zhen24a, title = {United We Stand, Divided We Fall: Networks to Graph (N2G) Abstraction for Robust Graph Classification Under Graph Label Corruption}, author = {Zhen, Zhiwei and Chen, Yuzhou and Kantarcioglu, Murat and Jee, Kangkook and Gel, Yulia}, booktitle = {Proceedings of the Second Learning on Graphs Conference}, pages = {30:1--30:19}, year = {2024}, editor = {Villar, Soledad and Chamberlain, Benjamin}, volume = {231}, series = {Proceedings of Machine Learning Research}, month = {27--30 Nov}, publisher = {PMLR}, pdf = {https://proceedings.mlr.press/v231/zhen24a/zhen24a.pdf}, url = {https://proceedings.mlr.press/v231/zhen24a.html}, abstract = {Nowadays, graph neural networks (GNN) are the primary machinery to tackle (semi)-supervised graph classification tasks. The aim here is to predict classes for unlabeled graphs, given a collection of graphs with known labels. However, in many real-world applications the available information on graph classes may be distorted either due to incorrect labelling process (e.g., as in biochemistry and bioinformatics) or may be subject to targeted attacks (e.g., as in network-based customer attrition analytics). Over the past few years, the increasing number of studies has indicated that GNNs are prone both to noisy node and noisy graph labels, and while this problem has received noticeable attention for node classification tasks, vulnerability of GNNs for graph classification with perturbed graph labels still remains in its nascence. We hypothesise that this challenge can be addressed by the universal principle {\it United We Stand, Divided We Fall}. In particular, most GNNs view each graph as a standalone entity and, as a result, are limited in their abilities to account for complex interdependencies among the graphs. Inspired by the recent studies on molecular graph learning, we propose a new robust knowledge representation called {\it Networks to Graph} (N2G). The key N2G idea is to construct a new abstraction where each graph in the collection is now represented by a node, while an edge then reflects some sort of similarity among the graphs. As a result, the graph classification task can be then naturally reformulated as a node classification problem. We show that the proposed N2G representation approach does not only improve classification performance both in binary and multi-class scenarios but also substantially enhances robustness against noisy labels in the training data, leading to relative robustness gains up to 11.7\% on social network benchmarks and up to 25.8\% on bioinformatics graph benchmarks under 10\% of graph label corruption rate.} }
Endnote
%0 Conference Paper %T United We Stand, Divided We Fall: Networks to Graph (N2G) Abstraction for Robust Graph Classification Under Graph Label Corruption %A Zhiwei Zhen %A Yuzhou Chen %A Murat Kantarcioglu %A Kangkook Jee %A Yulia Gel %B Proceedings of the Second Learning on Graphs Conference %C Proceedings of Machine Learning Research %D 2024 %E Soledad Villar %E Benjamin Chamberlain %F pmlr-v231-zhen24a %I PMLR %P 30:1--30:19 %U https://proceedings.mlr.press/v231/zhen24a.html %V 231 %X Nowadays, graph neural networks (GNN) are the primary machinery to tackle (semi)-supervised graph classification tasks. The aim here is to predict classes for unlabeled graphs, given a collection of graphs with known labels. However, in many real-world applications the available information on graph classes may be distorted either due to incorrect labelling process (e.g., as in biochemistry and bioinformatics) or may be subject to targeted attacks (e.g., as in network-based customer attrition analytics). Over the past few years, the increasing number of studies has indicated that GNNs are prone both to noisy node and noisy graph labels, and while this problem has received noticeable attention for node classification tasks, vulnerability of GNNs for graph classification with perturbed graph labels still remains in its nascence. We hypothesise that this challenge can be addressed by the universal principle {\it United We Stand, Divided We Fall}. In particular, most GNNs view each graph as a standalone entity and, as a result, are limited in their abilities to account for complex interdependencies among the graphs. Inspired by the recent studies on molecular graph learning, we propose a new robust knowledge representation called {\it Networks to Graph} (N2G). The key N2G idea is to construct a new abstraction where each graph in the collection is now represented by a node, while an edge then reflects some sort of similarity among the graphs. As a result, the graph classification task can be then naturally reformulated as a node classification problem. We show that the proposed N2G representation approach does not only improve classification performance both in binary and multi-class scenarios but also substantially enhances robustness against noisy labels in the training data, leading to relative robustness gains up to 11.7\% on social network benchmarks and up to 25.8\% on bioinformatics graph benchmarks under 10\% of graph label corruption rate.
APA
Zhen, Z., Chen, Y., Kantarcioglu, M., Jee, K. & Gel, Y.. (2024). United We Stand, Divided We Fall: Networks to Graph (N2G) Abstraction for Robust Graph Classification Under Graph Label Corruption. Proceedings of the Second Learning on Graphs Conference, in Proceedings of Machine Learning Research 231:30:1-30:19 Available from https://proceedings.mlr.press/v231/zhen24a.html.

Related Material