Limitations of GNN
A compiled version of the recent insights I gained from the recent studies on theoretical properties of GNN.
A compiled version of the recent insights I gained from the recent studies on theoretical properties of GNN.
Medium
Limitations of Graph Neural Networks
Reading between the lines of the latest advancements in GML.
Graph Machine Learning research groups: Le Song
I do a series of posts on the groups in graph research. The second is Le Song.
His group had 7 accepted papers at ICLR 2020. The top-2 result after Sergey Levine.
Le Song (~1981)
- Affiliation: Georgia Institute of Technology;
- Education: Ph.D. at U. of Sydney in 2008 (supervised by Alex Smola);
- h-index: 59;
- Awards: best papers at ICML, NeurIPS, AISTATS;
- Interests: generative and adversarial graph models, social network analysis, diffusion models.
I do a series of posts on the groups in graph research. The second is Le Song.
His group had 7 accepted papers at ICLR 2020. The top-2 result after Sergey Levine.
Le Song (~1981)
- Affiliation: Georgia Institute of Technology;
- Education: Ph.D. at U. of Sydney in 2008 (supervised by Alex Smola);
- h-index: 59;
- Awards: best papers at ICML, NeurIPS, AISTATS;
- Interests: generative and adversarial graph models, social network analysis, diffusion models.
Chess is the Drosophila of Artificial Intelligence
While writing another post on the life of Andrei Leman, I found that John McCarthy, the creator of Lisp, author of the term Artificial Intelligence, and one of the first and most influential programmers has an interesting remark, where he discusses the importance of long-term research. Apparently this saying, Chess is the Drosophila of Artificial Intelligence, is attributed to a legendary Soviet scientist, Alexander Kronrod, who arguably founded the first AI lab in the USSR. Back in '70s there were some huge ICL machines (analog of IBM machines), people used punch cards to provide the input, and calculations lasted for days, without the possibility to run a program in parallel. Yet, Kronrod's lab, which included Weisfeiler, Leman, and many other famous CS scientists and engineers, were working among others on the development of the chess programs, which in 1974, by the name Kaissa, won the first international chess tournament among machines.
While writing another post on the life of Andrei Leman, I found that John McCarthy, the creator of Lisp, author of the term Artificial Intelligence, and one of the first and most influential programmers has an interesting remark, where he discusses the importance of long-term research. Apparently this saying, Chess is the Drosophila of Artificial Intelligence, is attributed to a legendary Soviet scientist, Alexander Kronrod, who arguably founded the first AI lab in the USSR. Back in '70s there were some huge ICL machines (analog of IBM machines), people used punch cards to provide the input, and calculations lasted for days, without the possibility to run a program in parallel. Yet, Kronrod's lab, which included Weisfeiler, Leman, and many other famous CS scientists and engineers, were working among others on the development of the chess programs, which in 1974, by the name Kaissa, won the first international chess tournament among machines.
Fresh picks from ArXiv
This week ranges applications of graphs from social studies to astrophysics to robotics π€ Also checkout latest survey of deep learning impact on scientific discovery by Eric Schmidt, ex-Google CEO π And lots of other coolest graph papers from last week.
Applications
β’ Deep Graph Matching via Blackbox Differentiation of Combinatorial Solvers β combinatorial optimization
β’ A Heterogeneous Dynamical Graph Neural Networks Approach to Quantify Scientific Impact β scientific impact
β’ Investigating Software Usage in the Social Sciences: A Knowledge Graph Approach β social sciences
β’ Graph Neural Networks for Particle Reconstruction in High Energy Physics detectors β physics
β’ Representing Multi-Robot Structure through Multimodal Graph Embedding for the Selection of Robot Teams β robotics
β’ Identification of Patterns in Cosmic-Ray Arrival Directions using Dynamic Graph Convolutional Neural Networks β astrophysics
Survey
β’ word2vec, node2vec, graph2vec, X2vec: Towards a Theory of Vector Embeddings of Structured Data by Martin Grohe
β’ A Survey of Deep Learning for Scientific Discovery by Eric Schmidt (ex-Google CEO)
β’ COVID-19 and Computer Audition: An Overview on What Speech & Sound Analysis Could Contribute in the SARS-CoV-2 Corona Crisis
CVPR
β’ Distillating Knowledge from Graph Convolutional Networks
GNN
β’ Bridging the Gap Between Spectral and Spatial Domains in Graph Neural Networks
Graph Theory
β’ The 1-2-3 Conjecture holds for graphs with large enough minimum degree
β’ Bounds for the rainbow disconnection number of graphs
This week ranges applications of graphs from social studies to astrophysics to robotics π€ Also checkout latest survey of deep learning impact on scientific discovery by Eric Schmidt, ex-Google CEO π And lots of other coolest graph papers from last week.
Applications
β’ Deep Graph Matching via Blackbox Differentiation of Combinatorial Solvers β combinatorial optimization
β’ A Heterogeneous Dynamical Graph Neural Networks Approach to Quantify Scientific Impact β scientific impact
β’ Investigating Software Usage in the Social Sciences: A Knowledge Graph Approach β social sciences
β’ Graph Neural Networks for Particle Reconstruction in High Energy Physics detectors β physics
β’ Representing Multi-Robot Structure through Multimodal Graph Embedding for the Selection of Robot Teams β robotics
β’ Identification of Patterns in Cosmic-Ray Arrival Directions using Dynamic Graph Convolutional Neural Networks β astrophysics
Survey
β’ word2vec, node2vec, graph2vec, X2vec: Towards a Theory of Vector Embeddings of Structured Data by Martin Grohe
β’ A Survey of Deep Learning for Scientific Discovery by Eric Schmidt (ex-Google CEO)
β’ COVID-19 and Computer Audition: An Overview on What Speech & Sound Analysis Could Contribute in the SARS-CoV-2 Corona Crisis
CVPR
β’ Distillating Knowledge from Graph Convolutional Networks
GNN
β’ Bridging the Gap Between Spectral and Spatial Domains in Graph Neural Networks
Graph Theory
β’ The 1-2-3 Conjecture holds for graphs with large enough minimum degree
β’ Bounds for the rainbow disconnection number of graphs
arXiv.org
Representing Multi-Robot Structure through Multimodal Graph...
Multi-robot systems of increasing size and complexity are used to solve
large-scale problems, such as area exploration and search and rescue. A key
decision in human-robot teaming is dividing a...
large-scale problems, such as area exploration and search and rescue. A key
decision in human-robot teaming is dividing a...
CS 520: Knowledge Graphs
Today starts a Stanford course on knowledge graphs. It's available for everyone for free via Zoom (although there is no guarantee to have the videos afterwards).
Today starts a Stanford course on knowledge graphs. It's available for everyone for free via Zoom (although there is no guarantee to have the videos afterwards).
Graph Isomorphism Software
Open-source software for finding isomorphism or canonical forms of graphs.
* Nauty/Traces
* Bliss
* saucy
* conauto
* Gi-ext
Open-source software for finding isomorphism or canonical forms of graphs.
* Nauty/Traces
* Bliss
* saucy
* conauto
* Gi-ext
pallini.di.uniroma1.it
Nauty Traces β Home
Nauty Traces Home: Graph canonical labeling and automorphism group computation for graph isomorphism
CS 520: Knowledge Graphs
There is no promise for the materials afterwards, but here is a page where presentations can be stored (so far there is a single presentation from Michael Galkin).
There is no promise for the materials afterwards, but here is a page where presentations can be stored (so far there is a single presentation from Michael Galkin).
βDeep Graph Library v0.4.3
In version 0.4.3, a popular GNN library DGL has now support for TensorFlow (originally supported PyTorch only), spin-off packages for Knowledge Graphs and Life Sciences (DGL-KE and DGL-LifeSci) and more flexible distributed training.
For me it's really interesting to see a big uplift in training time against PyTorch-BigGraph library, this looks great.
In version 0.4.3, a popular GNN library DGL has now support for TensorFlow (originally supported PyTorch only), spin-off packages for Knowledge Graphs and Life Sciences (DGL-KE and DGL-LifeSci) and more flexible distributed training.
For me it's really interesting to see a big uplift in training time against PyTorch-BigGraph library, this looks great.
Knowledge Graph course 2019
Michael Galkin had a course on knowledge graphs in 2019 and there are presentations of the course (though most in Russian) if you are interested.
Michael Galkin had a course on knowledge graphs in 2019 and there are presentations of the course (though most in Russian) if you are interested.
Network Epidemiology Online Workshop Series
In April, there are seminars, tutorials, and discussion groups organized by University of Maryland to study how network science can help fighting virus contamination. It's for free, although there is a limitation on the number of participants (~150). There are two tracks: just listen to the tutorials or additionally participate in the groups. Lectures through Zoom. Deadline for application: tomorrow, 7th April.
In April, there are seminars, tutorials, and discussion groups organized by University of Maryland to study how network science can help fighting virus contamination. It's for free, although there is a limitation on the number of participants (~150). There are two tracks: just listen to the tutorials or additionally participate in the groups. Lectures through Zoom. Deadline for application: tomorrow, 7th April.
CS 520 Knowledge Graphs Video Materials
It seems that they will make the lectures available on YouTube π
It seems that they will make the lectures available on YouTube π
YouTube
CS520: Knowledge Graph Seminar Session 1 (Spring 2020)
What is a Knowledge Graph?
Fresh picks from ArXiv
This week is full of CVPR papers with graphs, surveys on recommendation, and a historical note of the discovery of TSP algorithm is USSR π¨βπ
CVPR
β’ HOPE-Net: A Graph-based Model for Hand-Object Pose Estimation
β’ Graph Structured Network for Image-Text Matching
β’ Spatio-Temporal Graph for Video Captioning with Knowledge Distillation
β’ Multi-Modal Graph Neural Network for Joint Reasoning on Vision and Scene Text
β’ Disentangling and Unifying Graph Convolutions for Skeleton-Based Action Recognition
β’ Deep Homography Estimation for Dynamic Scenes
β’ Iterative Context-Aware Graph Inference for Visual Dialog
β’ Geometrically Principled Connections in Graph Neural Networks with Michael Bronstein
β’ LiDAR-based Online 3D Video Object Detection with Graph-based Message Passing and Spatiotemporal Transformer Attention
WebConf
β’ Graph Enhanced Representation Learning for News Recommendation
Survey
β’ Deep Learning on Knowledge Graph for Recommender System: A Survey
β’ Heterogeneous Network Representation Learning: Survey, Benchmark, Evaluation, and Beyond β essentially a survey on KG
β’ A historical note on the 3/2-approximation algorithm for the metric traveling salesman problem β show that Christofides algorithm used for TSP was discovered in USSR around the same time
β’ A Survey on Conversational Recommender Systems
Graph Theory
β’ Longest paths in random hypergraphs
β’ A Fast Algorithm for the Product Structure of Planar Graphs
β’ Counterexamples to "A Conjecture on Induced Subgraphs of Cayley Graphs"
This week is full of CVPR papers with graphs, surveys on recommendation, and a historical note of the discovery of TSP algorithm is USSR π¨βπ
CVPR
β’ HOPE-Net: A Graph-based Model for Hand-Object Pose Estimation
β’ Graph Structured Network for Image-Text Matching
β’ Spatio-Temporal Graph for Video Captioning with Knowledge Distillation
β’ Multi-Modal Graph Neural Network for Joint Reasoning on Vision and Scene Text
β’ Disentangling and Unifying Graph Convolutions for Skeleton-Based Action Recognition
β’ Deep Homography Estimation for Dynamic Scenes
β’ Iterative Context-Aware Graph Inference for Visual Dialog
β’ Geometrically Principled Connections in Graph Neural Networks with Michael Bronstein
β’ LiDAR-based Online 3D Video Object Detection with Graph-based Message Passing and Spatiotemporal Transformer Attention
WebConf
β’ Graph Enhanced Representation Learning for News Recommendation
Survey
β’ Deep Learning on Knowledge Graph for Recommender System: A Survey
β’ Heterogeneous Network Representation Learning: Survey, Benchmark, Evaluation, and Beyond β essentially a survey on KG
β’ A historical note on the 3/2-approximation algorithm for the metric traveling salesman problem β show that Christofides algorithm used for TSP was discovered in USSR around the same time
β’ A Survey on Conversational Recommender Systems
Graph Theory
β’ Longest paths in random hypergraphs
β’ A Fast Algorithm for the Product Structure of Planar Graphs
β’ Counterexamples to "A Conjecture on Induced Subgraphs of Cayley Graphs"
RL for TSP
There is already a rich literature on applying RL methods for Traveling Salesman Problem (TSP). In one line of research the solution is built incrementally, one node at a time. This is for example in a popular paper, Attention, Learn to Solve Routing Problems!. Recently another approach appeared: start with some solution to TSP and update this solution by swapping nodes. For example, Learning 2-opt Heuristics for the Traveling Salesman Problem via Deep Reinforcement Learning uses policy gradient method and shows that it significantly outperforms incremental approaches.
There is already a rich literature on applying RL methods for Traveling Salesman Problem (TSP). In one line of research the solution is built incrementally, one node at a time. This is for example in a popular paper, Attention, Learn to Solve Routing Problems!. Recently another approach appeared: start with some solution to TSP and update this solution by swapping nodes. For example, Learning 2-opt Heuristics for the Traveling Salesman Problem via Deep Reinforcement Learning uses policy gradient method and shows that it significantly outperforms incremental approaches.
Graph Representation Learning Workshop ICML 20
One of the places that gather a lot of researchers in GML is the workshop on Graph Representation Learning. This year it will appear at ICML 20. You can submit short papers (4 pages), including the works that you can resubmit later to a full conference elsewhere. This year there will be a separate track on using graphs for COVID-19 resolution. Deadline: 29th May.
One of the places that gather a lot of researchers in GML is the workshop on Graph Representation Learning. This year it will appear at ICML 20. You can submit short papers (4 pages), including the works that you can resubmit later to a full conference elsewhere. This year there will be a separate track on using graphs for COVID-19 resolution. Deadline: 29th May.
grlplus.github.io
Call for Papers
ICML 2020 Workshop
Some notes on reconstruction conjecture
Studying reconstruction conjecture is going down the rabbit hole π It seems so simple, but after a few steps its complexity unfolds and you find yourself reading another paper that provides some extra insights which only steer you away from the target.
Formulation. You have a graph. You delete one vertex from it, and put obtained subgraph to a list, called deck. Then you return to the original graph, delete another vertex and again add obtained subgraph to the deck. And so on, for all n different vertices. The main question is whether there are two non-isomorphic graphs that have the same deck.
Difficulty. I already wrote about this conjecture, showing that if you delete more than 1 vertex then there are indeed two different graphs with the same decks. That's the fact that you delete only one vertex makes it more difficult. Now think, you have n subgraphs, each has (n-1) original vertices, it seems you just have so much information as an input, O(n^3), to say everything about the original graph, which has only O(n^2). But no, it's still very hard to solve it.
So what if we have a deck instead and try to recover original graph. Can we do it fast? I'm maybe wrong, but it seems there is no known polynomial-time algorithms and even the complexity of this problem is not clear. This paper, for specific type of graphs, shows an algorithm O(n^8), which is the biggest factor I've seen for poly-time algorithms. This paper studies some similar problems related to reconstruction conjecture, showing that just checking that a graph has specific deck is already equivalent to graph isomorphism. Finally, there is a manual 1991 on all major achievements to date showing that reconstruction conjecture is solved for some families of graphs such as regular graphs.
Insights. After reading all this, I have the following thoughts.
* It would be cool to write a paper on the algorithm for reconstruction of the graph from its deck. I have some trivial idea how to do it via graph isomorphism solver and some heuristics. Actually the problem is not purely abstract and has a lot of important applications to chemoinformatics for example, so it's interesting.
* It would be cool to write a follow-up paper which uses RL to improve heuristics. Graph reconstruction is under radar compared to TSP, so it would be first-of-its-kind RL algorithm for such problem.
* Initially I had the idea that if I take hard instances for graph isomorphism and compute decks for such graphs then I would get the same decks for non-isomorphic graphs, solving reconstruction conjecture. The reason why I thought this was that hard non-isomorphic graphs are very similar to each other, so their subgraphs would be identical. But, regular graphs are already proved to be reconstructible, so many hard instances won't work (such as SRG or Miyazaki graphs). Then, there are "very hard" graphs, which are half-regular (there are only two values for degrees across all nodes). I tried those, but found that decks are completely different for non-isomorphic graphs. So my idea failed and I spent two days reading, thinking, and implementing it π
Studying reconstruction conjecture is going down the rabbit hole π It seems so simple, but after a few steps its complexity unfolds and you find yourself reading another paper that provides some extra insights which only steer you away from the target.
Formulation. You have a graph. You delete one vertex from it, and put obtained subgraph to a list, called deck. Then you return to the original graph, delete another vertex and again add obtained subgraph to the deck. And so on, for all n different vertices. The main question is whether there are two non-isomorphic graphs that have the same deck.
Difficulty. I already wrote about this conjecture, showing that if you delete more than 1 vertex then there are indeed two different graphs with the same decks. That's the fact that you delete only one vertex makes it more difficult. Now think, you have n subgraphs, each has (n-1) original vertices, it seems you just have so much information as an input, O(n^3), to say everything about the original graph, which has only O(n^2). But no, it's still very hard to solve it.
So what if we have a deck instead and try to recover original graph. Can we do it fast? I'm maybe wrong, but it seems there is no known polynomial-time algorithms and even the complexity of this problem is not clear. This paper, for specific type of graphs, shows an algorithm O(n^8), which is the biggest factor I've seen for poly-time algorithms. This paper studies some similar problems related to reconstruction conjecture, showing that just checking that a graph has specific deck is already equivalent to graph isomorphism. Finally, there is a manual 1991 on all major achievements to date showing that reconstruction conjecture is solved for some families of graphs such as regular graphs.
Insights. After reading all this, I have the following thoughts.
* It would be cool to write a paper on the algorithm for reconstruction of the graph from its deck. I have some trivial idea how to do it via graph isomorphism solver and some heuristics. Actually the problem is not purely abstract and has a lot of important applications to chemoinformatics for example, so it's interesting.
* It would be cool to write a follow-up paper which uses RL to improve heuristics. Graph reconstruction is under radar compared to TSP, so it would be first-of-its-kind RL algorithm for such problem.
* Initially I had the idea that if I take hard instances for graph isomorphism and compute decks for such graphs then I would get the same decks for non-isomorphic graphs, solving reconstruction conjecture. The reason why I thought this was that hard non-isomorphic graphs are very similar to each other, so their subgraphs would be identical. But, regular graphs are already proved to be reconstructible, so many hard instances won't work (such as SRG or Miyazaki graphs). Then, there are "very hard" graphs, which are half-regular (there are only two values for degrees across all nodes). I tried those, but found that decks are completely different for non-isomorphic graphs. So my idea failed and I spent two days reading, thinking, and implementing it π
Telegram
Graph Machine Learning
ββMy thoughts on Kelly's conjecture.
A Kelly's conjecture (a.k.a. reconstruction conjecture) states that for a graph with n vertices, a deck with n graphs on (n-1) vertices can reconstruct exactly a graph.
In other words, pick a graph, remove one vertexβ¦
A Kelly's conjecture (a.k.a. reconstruction conjecture) states that for a graph with n vertices, a deck with n graphs on (n-1) vertices can reconstruct exactly a graph.
In other words, pick a graph, remove one vertexβ¦
Graph Machine Learning research groups: Karsten Borgwardt
I do a series of posts on the groups in graph research. The third is Karsten Borgwardt. His group at ETH Zurich is working on computational biology and his lab is well known for development of some of commonly used graph kernels (e.g. Weisfeiler-Leman, graphlet, shortest path).
Karsten Borgwardt (1980)
- Affiliation: ETH Zurich;
- Education: Ph.D. at Ludwig-βMaximilians-University in Munich in 2007 (supervised by Hans-βPeter Kriegel);
- h-index: 47;
- Awards: Krupp Award, best papers at NeurIPS;
- Interests: graph kernels; molecular graph representation; computational biology
I do a series of posts on the groups in graph research. The third is Karsten Borgwardt. His group at ETH Zurich is working on computational biology and his lab is well known for development of some of commonly used graph kernels (e.g. Weisfeiler-Leman, graphlet, shortest path).
Karsten Borgwardt (1980)
- Affiliation: ETH Zurich;
- Education: Ph.D. at Ludwig-βMaximilians-University in Munich in 2007 (supervised by Hans-βPeter Kriegel);
- h-index: 47;
- Awards: Krupp Award, best papers at NeurIPS;
- Interests: graph kernels; molecular graph representation; computational biology
Machine Learning & Computational Biology Lab
Homepage - Machine Learning and Computational Biology Lab
Online Math Seminars
There is a list of upcoming math seminars around the world, that are due to the confinement will be organized online and can be accessed by anyone. Topics are broad, including combinatorics, group theory and graphs.
There is a list of upcoming math seminars around the world, that are due to the confinement will be organized online and can be accessed by anyone. Topics are broad, including combinatorics, group theory and graphs.
researchseminars.org
researchseminars.org - Browse talks
Welcome to researchseminars.org, a list of research seminars and conferences!
Fresh picks from ArXiv
This week features SIDMOD and CVPR accepted papers + ICML submissions; graph embeddings for text classification and surveys on reconstruction conjecture and double descent.
SIGMOD
β’ Exact Single-Source SimRank Computation on Large Graphs
β’ An Algorithm for Context-Free Path Queries over Graph Databases
β’ A1: A Distributed In-Memory Graph Database β Bing's graph database
GNN
β’ Principal Neighbourhood Aggregation for Graph Nets with Petar VeliΔkoviΔ; on the injection with continuous features
β’ VGCN-BERT: Augmenting BERT with Graph Embedding for Text Classification β application to NLP
CVPR
β’ Semantic Image Manipulation Using Scene Graphs
ICML
β’ Learning to Recognize Spatial Configurations of Objects with Graph Neural Networks
β’ Graph Highway Networks
β’ The general theory of permutation equivarant neural networks and higher order graph variational encoders
Graph Theory
β’ On reconstruction of graphs from the multiset of subgraphs obtained by deleting β vertices
Survey
β’ A Brief Prehistory of Double Descent
This week features SIDMOD and CVPR accepted papers + ICML submissions; graph embeddings for text classification and surveys on reconstruction conjecture and double descent.
SIGMOD
β’ Exact Single-Source SimRank Computation on Large Graphs
β’ An Algorithm for Context-Free Path Queries over Graph Databases
β’ A1: A Distributed In-Memory Graph Database β Bing's graph database
GNN
β’ Principal Neighbourhood Aggregation for Graph Nets with Petar VeliΔkoviΔ; on the injection with continuous features
β’ VGCN-BERT: Augmenting BERT with Graph Embedding for Text Classification β application to NLP
CVPR
β’ Semantic Image Manipulation Using Scene Graphs
ICML
β’ Learning to Recognize Spatial Configurations of Objects with Graph Neural Networks
β’ Graph Highway Networks
β’ The general theory of permutation equivarant neural networks and higher order graph variational encoders
Graph Theory
β’ On reconstruction of graphs from the multiset of subgraphs obtained by deleting β vertices
Survey
β’ A Brief Prehistory of Double Descent
KDD Cup 2020 AutoGraph Challenge
There is a kaggle-like challenge until 25th May on automl for graph embeddings. 5 datasets, $33.5K prize pool, node classification task, accuracy metric, time constraints on train and predict.
There is a kaggle-like challenge until 25th May on automl for graph embeddings. 5 datasets, $33.5K prize pool, node classification task, accuracy metric, time constraints on train and predict.
Discovering automorphism of a graph and its deck.
My hypothesis was that if you take some hard graph for graph isomorphism problem and remove one vertex, then the resulted graph will be much easier because the symmetry of the original graph will be broken. So I took the hardest known graphs for graph isomorphism and checked how much time it takes for determining automorphism group (which is similar to how hard it would be to run isomorphism test).
The results are quite interesting. Indeed for many subgraphs determining automorphism is 50-100 times easier. But surprisingly, there are subgraphs which are harder than the original graph.
In the image below, you can see results for 6 graphs from cfi-rigid-z2 with ~1000-2000 vertices and checked the runtime for the original graph and all possible subgraphs by deleting one vertex. You can see that while for the first two graphs (first row), all subgraphs are easier, for the next 4 graphs, there are smaller subgraphs that take 5x more time to solve that the original graph.
This could happen for three reasons: (1) nauty solver has some heuristics that worked better for the original graph than for the subgraphs (2) not stable running and rerunning it would result in different runtimes and (3) smaller graphs somehow indeed harder than the original graph. I think (2) is very unlikely and my guess is a combination of (1) and (3): removing a vertex makes equivalent vertices in the original graph to be non-equivalent in the subgraph which reduces the amount of pruning nauty does.
My hypothesis was that if you take some hard graph for graph isomorphism problem and remove one vertex, then the resulted graph will be much easier because the symmetry of the original graph will be broken. So I took the hardest known graphs for graph isomorphism and checked how much time it takes for determining automorphism group (which is similar to how hard it would be to run isomorphism test).
The results are quite interesting. Indeed for many subgraphs determining automorphism is 50-100 times easier. But surprisingly, there are subgraphs which are harder than the original graph.
In the image below, you can see results for 6 graphs from cfi-rigid-z2 with ~1000-2000 vertices and checked the runtime for the original graph and all possible subgraphs by deleting one vertex. You can see that while for the first two graphs (first row), all subgraphs are easier, for the next 4 graphs, there are smaller subgraphs that take 5x more time to solve that the original graph.
This could happen for three reasons: (1) nauty solver has some heuristics that worked better for the original graph than for the subgraphs (2) not stable running and rerunning it would result in different runtimes and (3) smaller graphs somehow indeed harder than the original graph. I think (2) is very unlikely and my guess is a combination of (1) and (3): removing a vertex makes equivalent vertices in the original graph to be non-equivalent in the subgraph which reduces the amount of pruning nauty does.
RWTH AACHEN UNIVERSITY
Benchmark Graphs for Practical Graph Isomorphism
The state-of-the-art solvers for the graph isomorphism problem (e.g. bliss , nauty/traces , conauto , saucy , etc.) can readily solve generic instances with tens of thousands of vertices. Indeed, experiments show that on inputs without particular combinatorialβ¦