Are we measuring oversmoothing in graph neural networks correctly?
Kaicheng Zhang,
Piero Deidda,
Desmond J. Higham,
Francesco Tudisco,
International Conference on Learning Representations (ICLR),
(2026)
Abstract
Oversmoothing is a fundamental challenge in graph neural networks (GNNs): as the number of layers increases, node embeddings become increasingly similar, and model performance drops sharply. Traditionally, oversmoothing has been quantified using metrics that measure the similarity of neighbouring node features, such as the Dirichlet energy. We argue that these metrics have critical limitations and fail to reliably capture oversmoothing in realistic scenarios. For instance, they provide meaningful insights only for very deep networks, while typical GNNs show a performance drop already with as few as 10 layers. As an alternative, we propose measuring oversmoothing by examining the numerical or effective rank of the feature representations. We provide extensive numerical evaluation across diverse graph architectures and datasets to show that rank-based metrics consistently capture oversmoothing, whereas energy-based metrics often fail. Notably, we reveal that drops in the rank align closely with performance degradation, even in scenarios where energy metrics remain unchanged. Along with the experimental evaluation, we provide theoretical support for this approach, clarifying why Dirichlet-like measures may fail to capture performance drop and proving that the numerical rank of feature representations collapses to one for a broad family of GNN architectures.
Please cite this paper as:
@inproceedings{zhang2026measuring,
title={Are we measuring oversmoothing in graph neural networks correctly?},
author={Zhang, Kaicheng and Deidda, Piero and Higham, Desmond J. and Tudisco, Francesco},
booktitle={International Conference on Learning Representations (ICLR)},
year={2026}
}
Links:
arxiv
doi
Keywords:
networks
neural networks
deep learning
graph neural networks
oversmoothing
low-rank