Barriers for the performance of graph neural networks (GNN) in discrete random structures. A comment on <cit.>,<cit.>,<cit.>

06/05/2023
by   David Gamarnik, et al.
0

Recently graph neural network (GNN) based algorithms were proposed to solve a variety of combinatorial optimization problems, including Maximum Cut problem, Maximum Independent Set problem and similar other problems <cit.>,<cit.>. The publication <cit.> stirred a debate whether GNN based method was adequately benchmarked against best prior methods. In particular, critical commentaries <cit.> and <cit.> point out that simple greedy algorithm performs better than GNN in the setting of random graphs, and in fact stronger algorithmic performance can be reached with more sophisticated methods. A response from the authors <cit.> pointed out that GNN performance can be improved further by tuning up the parameters better. We do not intend to discuss the merits of arguments and counter-arguments in <cit.>,<cit.>,<cit.>,<cit.>. Rather in this note we establish a fundamental limitation for running GNN on random graphs considered in these references, for a broad range of choices of GNN architecture. These limitations arise from the presence of the Overlap Gap Property (OGP) phase transition, which is a barrier for many algorithms, both classical and quantum. As we demonstrate in this paper, it is also a barrier to GNN due to its local structure. We note that at the same time known algorithms ranging from simple greedy algorithms to more sophisticated algorithms based on message passing, provide best results for these problems up to the OGP phase transition. This leaves very little space for GNN to outperform the known algorithms, and based on this we side with the conclusions made in <cit.> and <cit.>.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset