Network Analysis Resources & Updates
3.03K subscribers
877 photos
163 files
1.18K links
Are you seeking assistance or eager to collaborate?
Don't hesitate to dispatch your insights, inquiries, proposals, promotions, bulletins, announcements, and more to our channel overseer. We're all ears!

Contact: @Questioner2
Download Telegram
Distributed Graph Neural Network Training.pdf
1.9 MB
πŸ“ƒDistributed Graph Neural Network Training: A Survey

πŸ—“ Publish year: 2023

πŸ§‘β€πŸ’»Authors: YINGXIA SHAO، HONGZHENG LI, HONGBO YIN, XIZHI GU، WENTAO ZHANG,...

🏒Universities: Beijing University of Posts and Telecommunications, Carnegie Mellon University, The Hong Kong University of Science and Technology (Guangzhou), Peking University.

πŸ“Ž Study paper

πŸ“²Channel: @ComplexNetworkAnalysis
#paper #Graph #Distributed #Survey
πŸ‘3❀1
πŸ“ƒA Survey of Graph Neural Networks for Social Recommender Systems

πŸ—“ Publish year: 2024

πŸ§‘β€πŸ’»Authors: KARTIK SHARMA, YEON-CHANG LEE, SIVAGAMI NAMBI,...

🏒Universities: Georgia Institute of Technology, Ulsan National Institute of Science and Technology, Hanyang University.

πŸ“Ž Study paper

πŸ“²Channel: @ComplexNetworkAnalysis
#paper #Graph #GNN #Survey
πŸ‘1
🎞 Machine Learning with Graphs: hyperbolic graph embeddings

πŸ’₯Free recorded course by Prof. Jure Leskovec

πŸ’₯ This part focused on graph representation learning in Euclidean embedding spaces. In this lecture, we introduce hyperbolic embedding spaces, which are great for modeling hierarchical, tree-like graphs. Moreover, we introduce basics for hyperbolic geometry models, which leads to the idea of hyperbolic GNNs. More details can be found in the paper: Hyperbolic Graph Convolutional Neural Networks

πŸ“½ Watch

πŸ“²Channel: @ComplexNetworkAnalysis
#video #course #Graph #GNN #Machine_Learning
πŸ‘2
🎞 Machine Learning with Graphs: design space of graph neural networks

πŸ’₯Free recorded course by Prof. Jure Leskovec

πŸ’₯ This part discussed the important topic of GNN architecture design. Here, we introduce 3 key aspects in GNN design: (1) a general GNN design space, which includes intra-layer design, inter-layer design and learning configurations; (2) a GNN task space with similarity metrics so that we can characterize different GNN tasks and, therefore, transfer the best GNN models across tasks; (3) an effective GNN evaluation technique so that we can convincingly evaluate any GNN design question, such as β€œIs BatchNorm generally useful for GNNs?”. Overall, we provide the first systematic investigation of general guidelines for GNN design, understandings of GNN tasks, and how to transfer the best GNN designs across tasks. We release GraphGym as an easy-to-use code platform for GNN architectural design. More information can be found in the paper: Design Space for Graph Neural Networks

πŸ“½ Watch

πŸ“²Channel: @ComplexNetworkAnalysis
#video #course #Graph #GNN #Machine_Learning
πŸ“„ Graph Data Management and Graph Machine Learning: Synergies and Opportunities

πŸ—“
Publish year: 2025

πŸ§‘β€πŸ’»Authors: Arijit Kha, Xiangyu Ke, Yinghui Wu
🏒University:
- Aalborg University, Denmark
- Zhejiang University, China

- Case Western Reserve University, USA

πŸ“Ž Study the paper

⚑️Channel: @ComplexNetworkAnalysis
#review #graph #machine_learning #data_management
πŸ‘1
πŸ“š A curated list of awesome network analysis resources
πŸ’₯ GitBook website

🌐 Study

⚑️Channel: @ComplexNetworkAnalysis
#github #graph #visualization #book
πŸ“ƒ Methods of decomposition theory and graph labeling in the study of social network structure

πŸ—“ Publish year: 2024

πŸ§‘β€πŸ’»Authors: L Hulianytskyi, M Semeniuta, S Yakymenko
🏒Universities: Prospekt Universytetskyi,Ukraine

πŸ“Ž Study the paper

⚑️Channel: @ComplexNetworkAnalysis
#review #graph_labling #decomposition
❀1πŸ‘1
🎞 Machine Learning with Graphs: GraphSAGE Neighbor Sampling

πŸ’₯Free recorded course by Prof. Jure Leskovec

πŸ’₯ This part discussed Neighbor Sampling, That is a representative method used to scale up GNNs to large graphs. The key insight is that a K-layer GNN generates a node embedding by using only the nodes from the K-hop neighborhood around that node. Therefore, to generate embeddings of nodes in the mini-batch, only the K-hop neighborhood nodes and their features are needed to load onto a GPU, a tractable operation even if the original graph is large. To further reduce the computational cost, only a subset of neighboring nodes is sampled for GNNs to aggregate.


πŸ“½ Watch

πŸ“²Channel: @ComplexNetworkAnalysis
#video #course #Graph #GNN #Machine_Learning #GraphSAGE