Data Science by ODS.ai 🦜
46K subscribers
664 photos
77 videos
7 files
1.75K links
First Telegram Data Science channel. Covering all technical and popular staff about anything related to Data Science: AI, Big Data, Machine Learning, Statistics, general Math and the applications of former. To reach editors contact: @malev
Download Telegram
Forwarded from Graph Machine Learning
Awesome graph repos

Collections of methods and papers for specific graph topics.

Graph-based Deep Learning Literature β€” Links to Conference Publications and the top 10 most-cited publications, Related workshops, Surveys / Literature Reviews / Books in graph-based deep learning.

awesome-graph-classification β€” A collection of graph classification methods, covering embedding, deep learning, graph kernel and factorization papers with reference implementations.

Awesome-Graph-Neural-Networks β€” A collection of resources related with graph neural networks..

awesome-graph β€” A curated list of resources for graph databases and graph computing tools

awesome-knowledge-graph β€” A curated list of Knowledge Graph related learning materials, databases, tools and other resources.

awesome-knowledge-graph β€” A curated list of awesome knowledge graph tutorials, projects and communities.

Awesome-GNN-Recommendation β€” graph mining for recommender systems.

awesome-graph-attack-papers β€” links to works about adversarial attacks and defenses on graph data or GNNs.

Graph-Adversarial-Learning β€” Attack-related papers, Defense-related papers, Robustness Certification papers, etc., ranging from 2017 to 2021.

awesome-self-supervised-gnn β€” Papers about self-supervised learning on GNNs.

awesome-self-supervised-learning-for-graphs β€” A curated list for awesome self-supervised graph representation learning resources.

Awesome-Graph-Contrastive-Learning β€” Collection of resources related with Graph Contrastive Learning.
πŸ‘1
Data Science by ODS.ai 🦜
Starting -1 Data Science Breakfast as an audio chat
Starting 0 Data Breakfast as an audio chat in this channel in 15 minutes.

This is an informal online event where you can discuss anything related to Data Science (even vaguely related).
Live stream finished (1 hour)
Forwarded from Gradient Dude
This media is not supported in your browser
VIEW IN TELEGRAM
Researchers from Berkeley rolled out VideoGPT - a transformer that generates videos.

The results are not super "WOW", but the architecture is quite simple and now it can be a starting point for all future work in this direction. As you know, GPT-3 for text generation was also not built right away. So let's will wait for method acceleration and quality improvement.

πŸ“Paper
βš™οΈCode
🌐Project page
πŸƒDemo
Channel photo updated
Do you like our new logo?
Anonymous Poll
29%
Yeah
37%
No
34%
Can do better
For almost 5 years channel picture beared arbitrary picture found in google and now we updated it with a proper new channel logo generated by neural network. Do you like it?
Channel photo updated
The Annotated Transformer

3 years ago Alexander Rush created an incredible notebook supported the "Attention is All You Need" paper giving a possibility to dive in the implementation details and obtain your own transformer :)

We, SkoltechNLP group, within our Neual NLP 2021 course revisited this notebook for adapting it as a seminar. Of course, the original code was created 3 years ago and in some places is incompatible with new versions of required libraries. As a result, we created "runnable with 'Run all Cells' for April 2021" version of this notebook:
https://github.com/skoltech-nlp/annotated-transformer

So if you want to learn the Transformer and run an example in your computer or Colab, you can save your time and use current version of this great notebook. Also, we add some links to the amazing resources about Transformers that emerged during these years:
* Seq2Seq and Attention by Lena Voita;
* The Illustrated Transformer.

Enjoy your Transformer! And be free to ask any questions and leave comments.
​​MDETR: Modulated Detection for End-to-End Multi-Modal Understanding

Multi-modal reasoning systems rely on a pre-trained object detector to extract regions of interest from the image. However, this crucial module is typically used as a black box, trained independently of the downstream task and on a fixed vocabulary of objects and attributes.
The authors present an end-to-end approach to multi-modal reasoning systems, which works much better than using a separate pre-trained decoder.
They pre-train the network on 1.3M text-image pairs, mined from pre-existing multi-modal datasets having explicit alignment between phrases in text and objects in the image.
Fine-tuning this model achieves new SOTA results on phrase grounding, referring expression comprehension, and segmentation tasks. The approach could be extended to visual question answering.
Furthermore, the model is capable of handling the long tail of object categories.

Paper: https://arxiv.org/abs/2104.12763
Code: https://github.com/ashkamath/mdetr

A detailed unofficial overview of the paper: https://andlukyane.com/blog/paper-review-mdetr

#deeplearning #multimodalreasoning #transformer
​​Are Pre-trained Convolutions Better than Pre-trained Transformers?

In this paper, the authors from Google Research wanted to investigate whether CNN architectures can be competitive compared to transformers on NLP problems. It turns out that pre-trained CNN models outperform pre-trained Transformers on some tasks; they also train faster and scale better to longer sequences.

Overall, the findings outlined in this paper suggest that conflating pre-training and architectural advances is misguided and that both advances should be considered independently. The authors believe their research paves the way for a healthy amount of optimism in alternative architectures.

Paper: https://arxiv.org/abs/2105.03322

A detailed unofficial overview of the paper: https://andlukyane.com/blog/paper-review-cnnbettertransformers

#nlp #deeplearning #cnn #transformer #pretraining
Data Fest returns! πŸŽ‰ And pretty soon

πŸ“… Starting May 22nd and until June 19th we host an Online Fest just like we did last year:

πŸ”ΈOur YouTube livestream return to a zoo-forest with πŸ¦™πŸ¦Œ and this time 🐻a bear cub! (RU)

πŸ”ΈUnlimited networking in our spatial.chat - May 22nd will be the real community maelstrom (RU & EN)

πŸ”ΈTracks on our ODS.AI platform, with new types of activities and tons of new features (RU & EN)

Registration is live! Check out Data Fest 2021 website for the astonishing tracks we have in our programme and all the details 🀩
GAN Prior Embedded Network for Blind Face Restoration in the Wild

New proposed method allowed authors to improve the quality of old photoes

ArXiV: https://arxiv.org/abs/2105.06070
Github: https://github.com/yangxy/GPEN

#GAN #GPEN #blind_face_restoration #CV #DL
πŸ‘2
Forwarded from Binary Tree
Testing Python Applications with Pytest.

Pytest is a testing framework and test runner for Python. In this guide we will have a look at the most useful and common configuration and usage, including several pytest plugins and external libraries. Although Python comes with a unittest module in the standard library and there are other Python test frameworks like nose2 or Ward, pytest remains my favourite. The beauty of using simple functions instead of class hierarchies, one simple assert instead of many different assert functions, built-in parametrized testing, a nice system of fixtures and the number of available plugins makes it a pleasure to use.

#guide #testing #python #pytest
πŸ₯°2
Last Call: register to participate in the EMERGENCY DATAHACK

Online-hackathon for data-scientists and specialists in the fields of machine-learning, geography and geology.

Best solutions designed by the contestants during the event will be later utilized by the Ministry of the Russian Federation for Civil Defence, Emergencies and Elimination of Consequences of Natural Disasters (EMERCOM).

The contestants will be able to research and analyze data, for the first time provided by the Ministry. Also, the contestants will be able to work with data provided by the partners of the event: the Federal Service for Hydrometeorology (Roshydromet), the Federal Road Agency (Rosavtodor), GLONASS BDD, Tele2, Rostelecom, the Federal Water Resources Agency (Rosvodresources), the Main Directorate for Traffic Safety of Russia.

Date: May 28 – 30
Format: online
Registration: open until May 24 (the date is inclusive)

Link: https://emergencydatahack.ru

The aggregated prize fund for the event – 12 200 USD (in the national currency).