Graph Machine Learning
6.7K subscribers
53 photos
11 files
808 links
Everything about graph theory, computer science, machine learning, etc.


If you have something worth sharing with the community, reach out @gimmeblues, @chaitjo.

Admins: Sergey Ivanov; Michael Galkin; Chaitanya K. Joshi
Download Telegram
AAAI 2021 stats

Dates: Feb 2-9
Where: Online
Price: ~$300

All papers can be found here. Graph categorization can be found here.

• 9034 submissions (vs 7737 in 2020)
• 1692 accepted (vs 1591 in 2020)
• 21% acceptance rate (vs 21% in 2020)
• 141 graph papers (8% of total)
ICLR 2021 stats

Dates: May 4-8
Where: Online

All papers can be found here. Graph papers can be found here.

• 2997 submissions (vs 2594 in 2020)
• 860 accepted (vs 687 in 2020)
• 29% acceptance rate (vs 26.5% in 2020)
• 50 graph papers (6% of total)
S+SSPR Workshop

An online event on GNN, adversarial learning and other topics, happening today and tomorrow, with great list of keynotes: Nicholas Carlini, Michael Bronstein, Max Welling, Fabio Roli. Program can be found here and streaming on YouTube (resuming at 15 Europe time).
Boost then Convolve: Gradient Boosting Meets Graph Neural Networks

In our new work at ICLR 2021, we explore how to apply Gradient Boosted Decision Trees to graphs. Surprisingly, I haven't encountered before papers that test performance of pure GBDT on graphs, for example for node classification.

GBDTs are usually used for heterogeneous data (e.g. in Kaggle competitions): the columns can be categorical, of different scale and meaning (e.g. income column vs age column). Such data is quite common in the real world, but most of the research graph datasets have sparse homogeneous nodes features (e.g. bag-of-words features or word embeddings). So we asked a question whether GNNs are efficient on graphs with heterogeneous features.

The first insight is that you can just pretrain GBDT on the node features and use the predictions of GBDT for training GNN model. This already gives a boost to GNN model.

Second, we proposed a scheme how to train GBDT and GNN end-to-end, and this would additionally boost performance.

Third, this combo of GBDT and GNN, which we call BGNN, converges much faster than GNN and therefore usually is faster to train than pure GNN.

Some limitations.
* BGNN works well with heterogeneous features. So Cora datasets and others with homogeneous features are still better of with plain GNN.
* The approach works for node regression and classification. We have some ideas how to extend it to link prediction or graph classification, but haven't worked it out yet. If you have some interest in continuing this line of work, let me know.

The code and datasets are available here.
Graph Machine Learning research groups: Stefanie Jegelka

I do a series of posts on the groups in graph research, previous post is here. The 22nd is Stefanie Jegelka, a professor at MIT working on submodular functions, DPP, and more recently on theoretical aspects of GNNs.

Stefanie Jegelka (~1986)
- Affiliation: MIT
- Education: Ph.D. at Max Planck Institute for Intelligent Systems, Tubingen and ETH Zurich in 2012 (advisors: Jeff Bilmes, Bernhard Scholkopf, Andreas Krause)
- h-index 33
- Awards: Joseph A Martore Award, NSF CAREER Award, best papers at ICML, NeurIPS
- Interests: generalization and expressivity of GNNs, clustering and graph partitioning
PhD position in Graph Neural Networks Modelling

Norwegian University of Science and Technology opened a PhD position for the thesis topic Interpretable Models with Graph Neural Networks to support the Green Transition of Critical Infrastructures. Deadline is 1 Feb 2021. 3-year contract, ~500K NOK per year before tax.
Course: ODS Knowledge Graphs

Michael Galkin starts a self-paced course on knowledge graphs. For now, it's only in Russian, with the plan to make it in English after the first iteration. The first introduction lecture is available on YouTube. You can join discussion group for all your questions and proposals: @kg_course. The first lecture starts this Thursday, more in the channel @kg_course.

Course curriculum:
* Knowledge representations (RDF, RDFS, OWL)
* Storage and queries (SPARQL, Graph DBs)
* Consistency (RDF*, SHACL, ShEx)
* Semantic Data Integration
* Graph theory intro
* KG embeddings
* GNNs for KGs
* Applications: Question Answering, Query Embeddings
GNN User Group events

First event at GNN User Group organized by DGL team (Amazon) and CuGraph team (Nvidia) starts tomorrow. Events should be organized monthly. The first talk is "A Framework For Differentiable Discovery of Graph Algorithms (Dr. Le Song, Georgia Tech)" + some networking event.
RoboGrammar: Graph Grammar for Terrain-Optimized Robot Design

(video) A recent work done at MIT for constructing different robot designs via graph grammar. Graph grammars were introduced in 1992 and defines a set of rules of transforming one graph to another. With this, a user can specify input robot components as well as the type of the terrain and graph grammar will produce possible robot designs. Next, a variation of A* algorithm is used to search for the optimal robot design for a given terrain. More on this in this article.
CS224W: Machine Learning with Graphs 2021

CS224W is one of the most popular graph courses by Jure Leskovec at Stanford. This year includes extra topics such as label propagation, scalability of GNNs, and graph nets for science and biology. The slides for the first 6 out of 20 lectures are available.
GML Newsletter: Interpolation and Extrapolation of Graph Neural Networks

The new issue of the newsletter is about generalization of GNNs. Compared to the study of expressive power, there are fewer works about generalization. Nonetheless, I gathered the most exciting research I found on this topic, which I hope will familiarize you with this research direction.
How many paths of length k exist in a graph?

In case you are preparing for the next interview, here is a nice post describing several solutions to a common interview problem: count the number of possible walks between two points in a graph. The problem is not as easy as it seems.
Tutorial: Graph Neural Networks: Models and Applications

A new tutorial covering robustness, attacks, scalability and self-supervised learning for GNN models at AAAI 2021. Slides and video are available.
Sberloga Talk

In case you speak Russian I will be presenting today our ICLR 2021 work about combination of GBDT with GNN on graphs with tabular features. The talk will be 19-00 MSK time. Zoom link will be shared soon at @sberlogawithgraphs. For more videos from Sberloga, subscribe here: https://www.youtube.com/c/SBERLOGA
Cleora Paper

I already wrote about Cleora, an unsupervised embedding library, now there is a paper explaining details of it. The algorithm is just some form of matrix multiplication, yet it shows better performance for link prediction metrics and running time than Pytorch-BigGraph, DeepWalk and others.
How to get started with Graph Machine Learning

In a new post, Aleksa Gordić talks in depth about graph ML, its applications and shares useful resources to get you started in this world.