Graph Machine Learning
6.71K subscribers
53 photos
11 files
808 links
Everything about graph theory, computer science, machine learning, etc.


If you have something worth sharing with the community, reach out @gimmeblues, @chaitjo.

Admins: Sergey Ivanov; Michael Galkin; Chaitanya K. Joshi
Download Telegram
Graph Neural Networking Challenge 2021

An interesting competition, organized by Technical University of Catalonia (UPC) and ITU, about building GNNs to predict source-destination routing time. The goal is to test generalization abilities of GNNs: training on small graphs and testing on much larger graphs.
Udemy Graph Neural Network course

Online course at Udemy that covers the basics of representation learning on graphs (e.g. DeepWalk, node2vec) and popular GNN architectures, plus some PyG implementations.
Graphs at ICLR 2021

Very good digest of a few graph papers at ICLR 2021. Talks about new GNNS to solve overmoothing, over-squashing, heterophily, and attention problems.
PyTorch-Geometric Tutorial Talk

Today, I will speak about our ICLR work "Boost then Convolve: Gradient Boosting Meets Graph Neural Networks". If you want to learn more about how GBDT and GNN work, and how they can be applied successfully for node prediction tasks, please join here at 15 (Paris time).
Deep Learning DIY course

A very good deep learning course by Marc Lelarge that among other things cover graph ML: graph embeddings, signal processing, and GNNs. It comes with videos, slides, notebooks, and assignments.
Results of OGB large-scale challenge

OGB team announced the results of KDD 2021 cup challenge where teams competed in node classification, triplet prediction, and graph regression tasks. Short summaries are provided for the winning solutions and it's quite interesting to see the diversity of the proposed methods: some used ensembles of GNNs, some pretrained graph embeddings, some label propagation, among others. Notably, Baidu and DeepMind scored really well on these tasks. Congrats to the winners!
Graph Neural Networks User Group: June

June's meeting of GNN user group will include the following talks:

* 4:00 - 4:30 (PST): Binary Graph Neural Networks and Dynamic Graph Models (Mahdi Saleh, Imperial College London).
* 4:30 - 5:00 (PST): Simplifying large-scale visual analysis of tricky data & models with GPUs, graphs, and ML (Leo Meyerovich, Graphistry Inc)
* 5:00 - 5:30 (PST): Open Discussion and Networking

Join this Thursday!
Fresh picks from ArXiv
This week on ArXiv: 1000-layer GNN, solutions to OGB challenge, and theory behind GNN explanations 🤔

If I forgot to mention your paper, please shoot me a message and I will update the post.


Deep GNNs
* Training Graph Neural Networks with 1000 Layers ICML 2021
* Very Deep Graph Neural Networks Via Noise Regularisation with Petar Veličković, Peter Battaglia

Heterophily
* Improving Robustness of Graph Neural Networks with Heterophily-Inspired Designs with Danai Koutra

Knowledge graphs
* Query Embedding on Hyper-relational Knowledge Graphs with Mikhail Galkin

OGB-challenge
* Fast Quantum Property Prediction via Deeper 2D and 3D Graph Networks
* First Place Solution of KDD Cup 2021 & OGB Large-Scale Challenge Graph Prediction Track

Theory
* Towards a Rigorous Theoretical Analysis and Evaluation of GNN Explanations with Marinka Zitnik
* A unifying point of view on expressive power of GNNs

GNNs
* Stability of Graph Convolutional Neural Networks to Stochastic Perturbations with Alejandro Ribeiro
* TD-GEN: Graph Generation With Tree Decomposition
* Unsupervised Resource Allocation with Graph Neural Networks
* Equivariance-bridged SO(2)-Invariant Representation Learning using Graph Convolutional Network
* GemNet: Universal Directional Graph Neural Networks for Molecules with Stephan Günnemann
* Optimizing Graph Transformer Networks with Graph-based Techniques

Survey
* Systematic comparison of graph embedding methods in practical tasks
* Evaluating Modules in Graph Contrastive Learning
* A Survey on Mining and Analysis of Uncertain Graphs
Open Catalyst Challenge: Using AI to discover catalysts for renewable energy storage

Open Catalyst Project is an endeavor by Facebook and CMU to predict the energies between the molecules and the catalysts with the applications for discovering new energy solutions, which I wrote about in one of the coolest applications of GNNs. This year there is a competition for this project organized by the same team. Winners will be invited to NeurIPS 2021 to present their solutions.
Transferability of Spectral Graph Convolutional Neural Networks

A talk by Ron Levie (Ludwig Maximilian University of Munich, Germany) about spectral GNNs. As Xavier Bresson said: "His work aims at debunking the misconception that spectral nets are computationally expensive, unstable/do not generalize - which is not true (theoretically & in practice).". Good research for those who love math in GNN world.
Graph Machine Learning research groups: Johan Ugander

I do a series of posts on the groups in graph research, previous post is here. The 30th is Johan Ugander, a professor at Stanford, who was a post-doc at Microsoft Research Redmond 2014-2015 and held an affiliation with the Facebook Data Science team 2010-2014.

Johan Ugander (~1986)
- Affiliation: Stanford
- Education: Ph.D. at Cornell in 2014 (advisors: Jon Kleinberg)
- h-index 17
- Interests: social network analysis, algorithms on graphs, clustering
- Awards: Young Investigator Award, best paper awards (WebSci, WSDM, AAAI)
Compositional Tokenization for Knowledge Graphs

This is a guest post by Michael Galkin about their new paper of reducing the memory issues of existing approaches.

Pretty much all KG embedding algorithms are, in fact, shallow embedding algorithms. It means that each node is mapped to a unique vector - and as a basis for all downstream tasks you need to store the whole embedding matrix in memory. Already at the OGB scale (2.5-5M nodes) you’d need 2-10 GB VRAM on the embeddings only, not counting forward passes and backprop. The more nodes you have - the bigger the matrix, the more expensive GPU you need.

Looking back to 2015, it resembles word2vec and GloVe a lot - huge shallow word vocabularies of 0.5-3M words, every other word is OOV (out of vocab). Then, subword units arrived (as Byte-Pair Encoding or WordPiece) and dramatically reduced vocab sizes allowing building infinite combinations from a rather small tokens vocab (30-50K in BERT & GPT-2/3 ). Saved params are now better invested into a flurry of Transformer encoders.

If we treat nodes in a graph like “words”, what would be their “sub-word” units? Can we have a similar approach that would allow to bootstrap a representation of both seen and unseen nodes using the same vocab? We tackle those questions in our new work where we design NodePiece (pun intended), a compositional tokenization scheme for KGs where tokens are anchor nodes and relation types. Going from shallow to compositional encoding, we reduce embedding matrices 10-1000x times and still observe a competitive performance. Interestingly, sometimes you don’t even need trainable node embeddings to perform well on node classification and relation prediction, i.e, relations around the node are enough!

We encourage you to find even more details in the pre-print, Medum blog, and try out the code in Github repo.
Graph Neural Networks as Neural Diffusion PDEs

A new post by Michael Bronstein about the connection of GNNs and differential equations that govern diffusion on graphs. This gives new mathematical framework for studying different architectures on graphs as well as a blueprint for developing new ones.