Graph Machine Learning
6.7K subscribers
53 photos
11 files
808 links
Everything about graph theory, computer science, machine learning, etc.


If you have something worth sharing with the community, reach out @gimmeblues, @chaitjo.

Admins: Sergey Ivanov; Michael Galkin; Chaitanya K. Joshi
Download Telegram
Probabilistic Learning on Graphs via Contextual Architectures

This is a guest post by Federico Errica ([email protected]) about their new JMLR work called “Probabilistic Learning on Graphs via Contextual Architectures”.

Intro/TL;DR:
We propose a probabilistic methodology for representation learning on graph-structured data, in which a stack of Bayesian networks learns different distributions of a vertex’s neighbourhood. The main characteristics of our approach are (i) unsupervised, as it models the generation of node attributes; (ii) layer-wise training: (iii) incremental construction policy; (iv) maximum likelihood estimation with Expectation-Maximization. The model, called Contextual Graph Markov Model (CGMM), can be regarded as a probabilistic version of Deep Graph Networks (DGNs).

Each layer of the model implements a probabilistic version of neighbourhood aggregation. The hidden representation of each node is modelled as a categorical distribution. When aggregating neighbours, the incoming messages are the *frozen* posterior probabilities computed when training the previous layers. When discrete edge types are available, we can weight the contribution of nodes in different ways using the Switching Parent approximation. Moreover, each neighbour aggregation can be conditioned on an arbitrary subset of the previous layers.

By design, this incremental construction policy avoids the exploding/vanishing gradient effect. As a result, each layer exploits different sets of statistics when trying to maximize the likelihood of the nodes in each graph. We test the model on node and graph classification tasks. First, we generate unsupervised node/graph representations; then, we apply a standard ML classifier to output the right class. In turn, this leads to a critical analysis of some benchmarks used in the literature. Finally, we show that the performances of the model increase as we add more layers (up to 20).

Paper: https://www.jmlr.org/papers/v21/19-470.html
Code: https://github.com/diningphil/CGMM
Related reads: (i) https://doi.org/10.1016/j.neunet.2020.06.006 (ii) https://proceedings.mlr.press/v80/bacciu18a.html
Issue #1: Introduction, PAC Isometry, Over-Smoothing, and Evolution of the Field

Finally the first issue of a newsletter is out and I hope there will many more in the future. The most difficult of this is to find good stories for the email: it's somewhat different from posting on telegram and twitter, as you need to have more insights in a single story. So if you find something that could be relevant to the community, definitely send me a message.
Simple scalable graph neural networks

Michael Bronstein continues a marathon of great blog posts on GML. In a new post he describes their recent work on scaling GNNs to large network. There is a good introduction to sampling-based methods (e.g. SAGE, GraphSAINT, ClusterGCN), which sample a subgraph of a large graph and then train GNN only on a subgraph.

Then, he describes that it can be beneficial just precompute r-hop matrices, A^r X, and use MLP on these features. This way, you use topology of your graph and you apply mini-batch training with MLP.

What's cool is that the algorithm is already available in pytorch-geometric as a transform, which is implemented via sparseTensor matrix multiplication.
The Quantum Graph Recurrent Neural Network

This demonstration by pennylane investigates quantum graph recurrent neural networks (QGRNN), which are the quantum analogue of a classical graph recurrent neural network, and a subclass of the more general quantum graph neural network ansatz. Both the QGNN and QGRNN were introduced in this paper (2019) by Google X.
Drawing neural networks in LaTeX

There is a repo of good examples by Petar Veličković of how you can draw Tikz images in LaTeX. Here is an example of 1-layer GNN by Matthias Fey.
Graph Machine Learning research groups: Xavier Bresson

I do a series of posts on the groups in graph research, previous post is here. The 12th is Xavier Bresson, conference and tutorial organizers on graph machine learning.


Xavier Bresson (~1975)
- Affiliation: NTU Singapore
- Education: Ph.D. at EPFL, Switzerland in 2005 (supervised by Jean-Philippe Thiran);
- h-index: 38;
- Awards: Singapore NRF fellowship, best paper at IVMSP;
- Interests: signal processing on graphs, GNN, combinatorial optimization.
Temporal Graph Networks

Another blog post by Michael Bronstein with Emanuele Rossi on applying graph nets on dynamic graphs (represented as the stream of edges). Apparently this problem is much more realistic in many business contexts such as social networks and it has not been studied at depth until that paper.
Node regression problem

I asked on twitter what the available node regression data sets there and found quite a few interesting responses.

1. There are pure node regression data sets, but not so many. One can use Wikipedia, Pokec, or data sets from this paper. I hope to release a couple more data sets like these soon.

2. You can also find data sets in spatiotemporal prediction on graphs (eg. traffic forecasting). You are given graph + velocity on each lane and you are asked to predict velocity in the future. My opinion is that the problem is a toy problem: there are no features associated with the nodes (except for a speed). But you can take a look at DCRNN, STGCN, GaAN, Graph WaveNet, STGRAT, etc. models that deal with that.

3. You can find node regression in the work of simulating physics. A node is a particle, it has a few features (eg. position+velocity) you are asked to predict acceleration. This is an interesting problem, but I haven't found data sets. You probably need to write your own simulator.

4. Next scene prediction. Essentially the same as previous, but the objects can be anything: for example, a camera view in a self-driving car. You are asked to predict next position of every object. I don't know if anyone tried to solve this problem.

5. Action prediction for RL agent. NerveNet did it. Each object is a graph and you predict an action for each node.
Brief analytics of KDD papers

Here are some plots for upcoming KDD 2020 (only research track). It's interesting to compare it against ICML 2020. You can check out git repo with the analysis. Here is highlights:

1. Top affiliations at KDD are different than at ICML, with several "small" names at the top.
2. Leading authors are almost all from China.
3. There are more authors per paper (~4-5) at KDD than at ICML (~3-4). Only a single paper with a single author.
4. There are ~65 graph papers, with a handful batch on pure algorithms.
5. In total there are 217 research papers. Graphs comprise about 30% of all papers.
6. Wordcloud confirms: graphs are the most used word in titles.
7. "Geodesic Forests" is the shortest title appeared.