Abstract
Contextual Stochastic Block Models (cSBMs) are simple generative graph models commonly used to benchmark the performance of Graph Neural Networks (GNNs). In this thesis, I prove how inducing symmetries on general, nonlinear GNNs improves their cross entropy on cSBMs, and I derive the limiting accuracy of linear Graph Convolutional Networks (GCNs) on cSBMs. I use these results to derive the functional form of cross-entropy optimal GCNs on cSBMs and compare them to accuracy optimal GCNs.
Degree
MS
College and Department
Computational, Mathematical, and Physical Sciences; Mathematics
Rights
https://lib.byu.edu/about/copyright/
BYU ScholarsArchive Citation
Garrity, Trevor, "GNN Convergence and Accuracy Analysis on Contextual Stochastic Block Models" (2025). Theses and Dissertations. 10806.
https://scholarsarchive.byu.edu/etd/10806
Date Submitted
2025-04-23
Document Type
Thesis
Handle
http://hdl.lib.byu.edu/1877/etd13675
Keywords
graph neural networks, cSBMs, cross entropy, accuracy
Language
english