Loading Now

Summary of Rethinking the Effectiveness Of Graph Classification Datasets in Benchmarks For Assessing Gnns, by Zhengdao Li et al.


Rethinking the Effectiveness of Graph Classification Datasets in Benchmarks for Assessing GNNs

by Zhengdao Li, Yong Cao, Kefan Shuai, Yiming Miao, Kai Hwang

First submitted to arxiv on: 6 Jul 2024

Categories

  • Main: Machine Learning (cs.LG)
  • Secondary: None

     Abstract of paper      PDF of paper


GrooveSquid.com Paper Summaries

GrooveSquid.com’s goal is to make artificial intelligence research accessible by summarizing AI papers in simpler terms. Each summary below covers the same AI paper, written at different levels of difficulty. The medium difficulty and low difficulty versions are original summaries written by GrooveSquid.com, while the high difficulty version is the paper’s original abstract. Feel free to learn from the version that suits you best!

Summary difficulty Written by Summary
High Paper authors High Difficulty Summary
Read the original abstract here
Medium GrooveSquid.com (original content) Medium Difficulty Summary
The proposed empirical protocol based on a fair benchmarking framework investigates the performance discrepancy between simple methods and GNNs. To quantify the effectiveness of datasets in distinguishing advancements of GNNs, a novel metric is introduced considering both dataset complexity and model performance. This study thoroughly examines and provides an explicit definition for dataset effectiveness in graph learning, aligning with existing studies and intuitive assumptions across 16 real-world datasets.
Low GrooveSquid.com (original content) Low Difficulty Summary
This paper looks at how well graph classification benchmarks work. It seems that simple methods like MLPs can perform just as well as more complex Graph Neural Networks (GNNs). The question is: do these benchmarks actually show what GNNs are good for? To answer this, the authors propose a new way to test and compare different models using fair benchmarking rules. They also introduce a new metric to measure how well a dataset helps or hinders the performance of different models. By testing on 16 real-world datasets, they found that their metric agrees with what other studies have shown and makes sense intuitively.

Keywords

» Artificial intelligence  » Classification