Research

Paper

TESTING March 10, 2026

Transductive Generalization via Optimal Transport and Its Application to Graph Node Classification

Authors

MoonJeong Park, Seungbeom Lee, Kyungmin Kim, Jaeseung Heo, Seunghyuk Cho, Shouheng Li, Sangdon Park, Dongwoo Kim

Abstract

Many existing transductive bounds rely on classical complexity measures that are computationally intractable and often misaligned with empirical behavior. In this work, we establish new representation-based generalization bounds in a distribution-free transductive setting, where learned representations are dependent, and test features are accessible during training. We derive global and class-wise bounds via optimal transport, expressed in terms of Wasserstein distances between encoded feature distributions. We demonstrate that our bounds are efficiently computable and strongly correlate with empirical generalization in graph node classification, improving upon classical complexity measures. Additionally, our analysis reveals how the GNN aggregation process transforms the representation distributions, inducing a trade-off between intra-class concentration and inter-class separation. This yields depth-dependent characterizations that capture the non-monotonic relationship between depth and generalization error observed in practice. The code is available at https://github.com/ml-postech/Transductive-OT-Gen-Bound.

Metadata

arXiv ID: 2603.09257
Provider: ARXIV
Primary Category: cs.LG
Published: 2026-03-10
Fetched: 2026-03-11 06:02

Related papers

Raw Data (Debug)
{
  "raw_xml": "<entry>\n    <id>http://arxiv.org/abs/2603.09257v1</id>\n    <title>Transductive Generalization via Optimal Transport and Its Application to Graph Node Classification</title>\n    <updated>2026-03-10T06:43:18Z</updated>\n    <link href='https://arxiv.org/abs/2603.09257v1' rel='alternate' type='text/html'/>\n    <link href='https://arxiv.org/pdf/2603.09257v1' rel='related' title='pdf' type='application/pdf'/>\n    <summary>Many existing transductive bounds rely on classical complexity measures that are computationally intractable and often misaligned with empirical behavior. In this work, we establish new representation-based generalization bounds in a distribution-free transductive setting, where learned representations are dependent, and test features are accessible during training. We derive global and class-wise bounds via optimal transport, expressed in terms of Wasserstein distances between encoded feature distributions. We demonstrate that our bounds are efficiently computable and strongly correlate with empirical generalization in graph node classification, improving upon classical complexity measures. Additionally, our analysis reveals how the GNN aggregation process transforms the representation distributions, inducing a trade-off between intra-class concentration and inter-class separation. This yields depth-dependent characterizations that capture the non-monotonic relationship between depth and generalization error observed in practice. The code is available at https://github.com/ml-postech/Transductive-OT-Gen-Bound.</summary>\n    <category scheme='http://arxiv.org/schemas/atom' term='cs.LG'/>\n    <category scheme='http://arxiv.org/schemas/atom' term='stat.ML'/>\n    <published>2026-03-10T06:43:18Z</published>\n    <arxiv:primary_category term='cs.LG'/>\n    <author>\n      <name>MoonJeong Park</name>\n    </author>\n    <author>\n      <name>Seungbeom Lee</name>\n    </author>\n    <author>\n      <name>Kyungmin Kim</name>\n    </author>\n    <author>\n      <name>Jaeseung Heo</name>\n    </author>\n    <author>\n      <name>Seunghyuk Cho</name>\n    </author>\n    <author>\n      <name>Shouheng Li</name>\n    </author>\n    <author>\n      <name>Sangdon Park</name>\n    </author>\n    <author>\n      <name>Dongwoo Kim</name>\n    </author>\n  </entry>"
}