SlideShare a Scribd company logo
1 of 17
Jin-Woo Jeong
Network Science Lab
Dept. of Mathematics
The Catholic University of Korea
E-mail: zeus0208b@gmail.com
Aditya Grover, Jure Leskovec
1
 Introduction
• Motivation
• Introduction
 Feature Learning Framework
• Classic search strategies
• Node2vec
• Random Walks
• Search bias 𝛼
• The node2vec algorithm
• Learning edge features
Experiments
• Case Study: Les Misérables network
• Multi-label classification
• Link prediction
Conclusion
Q/A
2
Introduction
Motivation
 Supervised machine learning algorithms in network prediction tasks require informative, discriminating, and
independent features. Typically, these features are constructed through hand-engineering based on
domain-specific expertise. However, this process is tedious and the resulting features may not generalize
well across different prediction tasks.
 Previous techniques fail to satisfactorily define and optimize a reasonable objective required for scalable
unsupervised feature learning in networks. Classic approaches based on linear and non-linear
dimensionality reduction techniques such as Principal Component Analysis, Multi-Dimensional Scaling and
their extensions optimize an objective that transforms a representative data matrix of the network such that
it maximizes the variance of the data representation. Consequently, these approaches invariably involve
eigendecomposition of the appropriate data matrix which is expensive for large real-world networks.
Moreover, the resulting latent representations give poor performance on various prediction tasks over
networks.
3
Introduction
Introduction
 A node could be organized based on communities they belong to (i.e., homophily); in other cases, the
organization could be based on the structural roles of nodes in the network (i.e., structural equivalence).
 Real-world networks commonly exhibit a mixture of such equivalences.
4
Introduction
Introduction
 they propose node2vec, a semi-supervised algorithm for scalable feature learning in networks. they
optimize a custom graph-based objective function using SGD motivated by prior work on natural language
processing.
 Intuitively, their approach returns feature representations that maximize the likelihood of preserving
network neighborhoods of nodes in a d-dimensional feature space. They use a 2nd order random walk
approach to generate (sample) network neighborhoods for nodes.
 Contributions
 1. They propose node2vec, an efficient scalable algorithm for feature learning in networks that
efficiently optimizes a novel network-aware, neighborhood preserving objective using SGD.
 2. They show how node2vec is in accordance with established principles in network science, providing
flexibility in discovering representations conforming to different equivalences.
 3. They extend node2vec and other feature learning methods based on neighborhood preserving
objectives, from nodes to pairs of nodes for edge-based prediction tasks.
 4. They empirically evaluate node2vec for multi-label classification and link prediction on several real-
world datasets.
5
Feature Learning Framework
Feature Learning Framework
 They formulate learning in networks as a maximum likelihood optimization problem.
• 𝐺 = 𝑉, 𝐸 be a given graph
• 𝑓 ∶ 𝑉 → ℝ𝑑
be the mapping function from nodes to feature representations.
• 𝑓 is a matrix of size 𝑉 × 𝑑 parameters.
• 𝑑 be a number of dimension of feature representation.
• ∀
𝑢 ∈ 𝑉, 𝑁𝑆 𝑢 ⊂ 𝑉 𝑁𝑆 𝑢 : a network neighborhood of node u generated through a neighborhood
sampling strategy S.
where
6
Feature Learning Framework
Classic search strategies
• Breadth-first Sampling (BFS)
• The neighborhood 𝑁𝑆 is restricted to nodes which are immediate neighbors of the source. For
example, in Figure 1 for a neighborhood of size k = 3, BFS samples nodes s1, s2, s3.
• Depth-first Sampling (DFS)
• The neighborhood consists of nodes sequentially sampled at increasing distances from the source
node. In Figure 1, DFS samples s4, s5, s6.
7
Feature Learning Framework
node2vec
• Random Walks
• 𝑢 : A source node
• 𝑙 : Walk of fixed length
• 𝑐𝑖 : 𝑖th node in the walk
• 𝑐0 = 𝑢
• 𝜋𝑣𝑥 : unnormalized transition probability
• 𝑍 : normalizing constant
8
Feature Learning Framework
node2vec
• Search bias 𝜶
𝜋𝑣𝑥 = 𝛼𝑝𝑞(𝑡, 𝑥) ∙ 𝜔𝑣𝑥
• Return parameter, p
• Parameter p controls the likelihood of immediately
revisiting a node in the walk.
• In-out parameter, q
• Parameter q allows the search to differentiate
between “inward” and “outward” nodes.
9
Feature Learning Framework
node2vec
 The node2vec algorithm
• SkipGram algorithm in DEEPWALK
Φ → 𝑓
𝑤 → 𝑘
𝑊𝑣𝑖 → 𝑤𝑎𝑙𝑘
10
Feature Learning Framework
Learning edge features
 Given two nodes 𝑢 and 𝑣, they define a binary operator ◦ over the corresponding feature vectors f(𝑢) and
f(𝑣) in order to generate a representation g(𝑢, 𝑣) such that 𝑔 ∶ 𝑉 × 𝑉 → ℝ𝑑′
where 𝑑′
is the representation
size for the pair g 𝑢, 𝑣 .
 They consider several choices for the operator ◦ such that 𝑑′ = 𝑑 which are summarized in Table 1
11
Experiments
Case Study: Les Misérables network
 They use a network where nodes correspond to characters in the novel Les Misérables and edges connect
coappearing characters. The network has 77 nodes and 254 edges.
 They set d = 16 and run node2vec to learn feature representation for every node in the network.
 The feature representations are clustered using k- means.
p = 1, q = 0.5 p = 1, q = 2
homophily structural equivalence
12
Experiments
Multi-label classification
13
Experiments
Multi-label classification
14
Experiments
Link prediction
15
Conclusion
Conclusion
 In this paper, they researched feature learning in networks as a search-based optimization problem. This
perspective provides us with several advantages. It can elucidate traditional search strategies based on the
balance between exploration and exploitation. Additionally, it imparts interpretability to the learned
representations when applied in prediction tasks.
 The search strategy in node2vec allows flexible exploration and control of network neighborhoods
through parameters p and q. While these search parameters have intuitive interpretations, we achieve the
best results on complex networks when we can learn them directly from data.
 From a practical standpoint, node2vec is scalable and robust. We demonstrated the superiority of
extending node embeddings over widely used heuristic scores for link prediction. Their methodology
allows additional binary operators beyond those listed in Table 1.
 They aim to explore the reasons behind the success of the Hadamard operator over others in future work,
as well as establish interpretable equivalence notions for edges based on the search parameters.
16
Q & A
Q / A

More Related Content

Similar to 240325_JW_labseminar[node2vec: Scalable Feature Learning for Networks].pptx

LPCNN: convolutional neural network for link prediction based on network stru...
LPCNN: convolutional neural network for link prediction based on network stru...LPCNN: convolutional neural network for link prediction based on network stru...
LPCNN: convolutional neural network for link prediction based on network stru...TELKOMNIKA JOURNAL
 
Adaptive Geographical Search in Networks
Adaptive Geographical Search in NetworksAdaptive Geographical Search in Networks
Adaptive Geographical Search in NetworksAndrea Wiggins
 
NS-CUK Seminar: H.B.Kim, Review on "Deep Gaussian Embedding of Graphs: Unsup...
NS-CUK Seminar: H.B.Kim,  Review on "Deep Gaussian Embedding of Graphs: Unsup...NS-CUK Seminar: H.B.Kim,  Review on "Deep Gaussian Embedding of Graphs: Unsup...
NS-CUK Seminar: H.B.Kim, Review on "Deep Gaussian Embedding of Graphs: Unsup...ssuser4b1f48
 
Graph Representation Learning
Graph Representation LearningGraph Representation Learning
Graph Representation LearningJure Leskovec
 
Understanding Large Social Networks | IRE Major Project | Team 57
Understanding Large Social Networks | IRE Major Project | Team 57 Understanding Large Social Networks | IRE Major Project | Team 57
Understanding Large Social Networks | IRE Major Project | Team 57 Raj Patel
 
network mining and representation learning
network mining and representation learningnetwork mining and representation learning
network mining and representation learningsun peiyuan
 
240115_Thanh_LabSeminar[Don't walk, skip! online learning of multi-scale netw...
240115_Thanh_LabSeminar[Don't walk, skip! online learning of multi-scale netw...240115_Thanh_LabSeminar[Don't walk, skip! online learning of multi-scale netw...
240115_Thanh_LabSeminar[Don't walk, skip! online learning of multi-scale netw...thanhdowork
 
A Generalization of Transformer Networks to Graphs.pptx
A Generalization of Transformer Networks to Graphs.pptxA Generalization of Transformer Networks to Graphs.pptx
A Generalization of Transformer Networks to Graphs.pptxssuser2624f71
 
NS-CUK Seminar: S.T.Nguyen, Review on "Improving Graph Neural Network Express...
NS-CUK Seminar: S.T.Nguyen, Review on "Improving Graph Neural Network Express...NS-CUK Seminar: S.T.Nguyen, Review on "Improving Graph Neural Network Express...
NS-CUK Seminar: S.T.Nguyen, Review on "Improving Graph Neural Network Express...ssuser4b1f48
 
LINE: Large-scale Information Network Embedding.pptx
LINE: Large-scale Information Network Embedding.pptxLINE: Large-scale Information Network Embedding.pptx
LINE: Large-scale Information Network Embedding.pptxssuser2624f71
 
Machine Learning for Efficient Neighbor Selection in ...
Machine Learning for Efficient Neighbor Selection in ...Machine Learning for Efficient Neighbor Selection in ...
Machine Learning for Efficient Neighbor Selection in ...butest
 
240401_Thuy_Labseminar[Train Once and Explain Everywhere: Pre-training Interp...
240401_Thuy_Labseminar[Train Once and Explain Everywhere: Pre-training Interp...240401_Thuy_Labseminar[Train Once and Explain Everywhere: Pre-training Interp...
240401_Thuy_Labseminar[Train Once and Explain Everywhere: Pre-training Interp...thanhdowork
 
EDGE-Net: Efficient Deep-learning Gradients Extraction Network
EDGE-Net: Efficient Deep-learning Gradients Extraction NetworkEDGE-Net: Efficient Deep-learning Gradients Extraction Network
EDGE-Net: Efficient Deep-learning Gradients Extraction Networkgerogepatton
 
EDGE-Net: Efficient Deep-learning Gradients Extraction Network
EDGE-Net: Efficient Deep-learning Gradients Extraction NetworkEDGE-Net: Efficient Deep-learning Gradients Extraction Network
EDGE-Net: Efficient Deep-learning Gradients Extraction Networkgerogepatton
 
Resnet.pdf
Resnet.pdfResnet.pdf
Resnet.pdfYanhuaSi
 
Recognition and Detection of Real-Time Objects Using Unified Network of Faste...
Recognition and Detection of Real-Time Objects Using Unified Network of Faste...Recognition and Detection of Real-Time Objects Using Unified Network of Faste...
Recognition and Detection of Real-Time Objects Using Unified Network of Faste...dbpublications
 
NS-CUK Seminar: J.H.Lee, Review on "Rethinking the Expressive Power of GNNs...
NS-CUK Seminar:  J.H.Lee,  Review on "Rethinking the Expressive Power of GNNs...NS-CUK Seminar:  J.H.Lee,  Review on "Rethinking the Expressive Power of GNNs...
NS-CUK Seminar: J.H.Lee, Review on "Rethinking the Expressive Power of GNNs...ssuser4b1f48
 
NS-CUK Seminar: J.H.Lee, Review on "Rethinking the Expressive Power of GNNs ...
NS-CUK Seminar: J.H.Lee,  Review on "Rethinking the Expressive Power of GNNs ...NS-CUK Seminar: J.H.Lee,  Review on "Rethinking the Expressive Power of GNNs ...
NS-CUK Seminar: J.H.Lee, Review on "Rethinking the Expressive Power of GNNs ...ssuser4b1f48
 

Similar to 240325_JW_labseminar[node2vec: Scalable Feature Learning for Networks].pptx (20)

LPCNN: convolutional neural network for link prediction based on network stru...
LPCNN: convolutional neural network for link prediction based on network stru...LPCNN: convolutional neural network for link prediction based on network stru...
LPCNN: convolutional neural network for link prediction based on network stru...
 
Adaptive Geographical Search in Networks
Adaptive Geographical Search in NetworksAdaptive Geographical Search in Networks
Adaptive Geographical Search in Networks
 
Deepwalk vs Node2vec
Deepwalk vs Node2vecDeepwalk vs Node2vec
Deepwalk vs Node2vec
 
NS-CUK Seminar: H.B.Kim, Review on "Deep Gaussian Embedding of Graphs: Unsup...
NS-CUK Seminar: H.B.Kim,  Review on "Deep Gaussian Embedding of Graphs: Unsup...NS-CUK Seminar: H.B.Kim,  Review on "Deep Gaussian Embedding of Graphs: Unsup...
NS-CUK Seminar: H.B.Kim, Review on "Deep Gaussian Embedding of Graphs: Unsup...
 
Graph Representation Learning
Graph Representation LearningGraph Representation Learning
Graph Representation Learning
 
Understanding Large Social Networks | IRE Major Project | Team 57
Understanding Large Social Networks | IRE Major Project | Team 57 Understanding Large Social Networks | IRE Major Project | Team 57
Understanding Large Social Networks | IRE Major Project | Team 57
 
network mining and representation learning
network mining and representation learningnetwork mining and representation learning
network mining and representation learning
 
240115_Thanh_LabSeminar[Don't walk, skip! online learning of multi-scale netw...
240115_Thanh_LabSeminar[Don't walk, skip! online learning of multi-scale netw...240115_Thanh_LabSeminar[Don't walk, skip! online learning of multi-scale netw...
240115_Thanh_LabSeminar[Don't walk, skip! online learning of multi-scale netw...
 
A Generalization of Transformer Networks to Graphs.pptx
A Generalization of Transformer Networks to Graphs.pptxA Generalization of Transformer Networks to Graphs.pptx
A Generalization of Transformer Networks to Graphs.pptx
 
NS-CUK Seminar: S.T.Nguyen, Review on "Improving Graph Neural Network Express...
NS-CUK Seminar: S.T.Nguyen, Review on "Improving Graph Neural Network Express...NS-CUK Seminar: S.T.Nguyen, Review on "Improving Graph Neural Network Express...
NS-CUK Seminar: S.T.Nguyen, Review on "Improving Graph Neural Network Express...
 
LINE: Large-scale Information Network Embedding.pptx
LINE: Large-scale Information Network Embedding.pptxLINE: Large-scale Information Network Embedding.pptx
LINE: Large-scale Information Network Embedding.pptx
 
Machine Learning for Efficient Neighbor Selection in ...
Machine Learning for Efficient Neighbor Selection in ...Machine Learning for Efficient Neighbor Selection in ...
Machine Learning for Efficient Neighbor Selection in ...
 
240401_Thuy_Labseminar[Train Once and Explain Everywhere: Pre-training Interp...
240401_Thuy_Labseminar[Train Once and Explain Everywhere: Pre-training Interp...240401_Thuy_Labseminar[Train Once and Explain Everywhere: Pre-training Interp...
240401_Thuy_Labseminar[Train Once and Explain Everywhere: Pre-training Interp...
 
K means report
K means reportK means report
K means report
 
EDGE-Net: Efficient Deep-learning Gradients Extraction Network
EDGE-Net: Efficient Deep-learning Gradients Extraction NetworkEDGE-Net: Efficient Deep-learning Gradients Extraction Network
EDGE-Net: Efficient Deep-learning Gradients Extraction Network
 
EDGE-Net: Efficient Deep-learning Gradients Extraction Network
EDGE-Net: Efficient Deep-learning Gradients Extraction NetworkEDGE-Net: Efficient Deep-learning Gradients Extraction Network
EDGE-Net: Efficient Deep-learning Gradients Extraction Network
 
Resnet.pdf
Resnet.pdfResnet.pdf
Resnet.pdf
 
Recognition and Detection of Real-Time Objects Using Unified Network of Faste...
Recognition and Detection of Real-Time Objects Using Unified Network of Faste...Recognition and Detection of Real-Time Objects Using Unified Network of Faste...
Recognition and Detection of Real-Time Objects Using Unified Network of Faste...
 
NS-CUK Seminar: J.H.Lee, Review on "Rethinking the Expressive Power of GNNs...
NS-CUK Seminar:  J.H.Lee,  Review on "Rethinking the Expressive Power of GNNs...NS-CUK Seminar:  J.H.Lee,  Review on "Rethinking the Expressive Power of GNNs...
NS-CUK Seminar: J.H.Lee, Review on "Rethinking the Expressive Power of GNNs...
 
NS-CUK Seminar: J.H.Lee, Review on "Rethinking the Expressive Power of GNNs ...
NS-CUK Seminar: J.H.Lee,  Review on "Rethinking the Expressive Power of GNNs ...NS-CUK Seminar: J.H.Lee,  Review on "Rethinking the Expressive Power of GNNs ...
NS-CUK Seminar: J.H.Lee, Review on "Rethinking the Expressive Power of GNNs ...
 

More from thanhdowork

[20240429_LabSeminar_Huy]Spatio-Temporal Graph Neural Point Process for Traff...
[20240429_LabSeminar_Huy]Spatio-Temporal Graph Neural Point Process for Traff...[20240429_LabSeminar_Huy]Spatio-Temporal Graph Neural Point Process for Traff...
[20240429_LabSeminar_Huy]Spatio-Temporal Graph Neural Point Process for Traff...thanhdowork
 
240429_Thanh_LabSeminar[TranSG: Transformer-Based Skeleton Graph Prototype Co...
240429_Thanh_LabSeminar[TranSG: Transformer-Based Skeleton Graph Prototype Co...240429_Thanh_LabSeminar[TranSG: Transformer-Based Skeleton Graph Prototype Co...
240429_Thanh_LabSeminar[TranSG: Transformer-Based Skeleton Graph Prototype Co...thanhdowork
 
240429_Thuy_Labseminar[Simplifying and Empowering Transformers for Large-Grap...
240429_Thuy_Labseminar[Simplifying and Empowering Transformers for Large-Grap...240429_Thuy_Labseminar[Simplifying and Empowering Transformers for Large-Grap...
240429_Thuy_Labseminar[Simplifying and Empowering Transformers for Large-Grap...thanhdowork
 
240422_Thanh_LabSeminar[Dynamic Graph Enhanced Contrastive Learning for Chest...
240422_Thanh_LabSeminar[Dynamic Graph Enhanced Contrastive Learning for Chest...240422_Thanh_LabSeminar[Dynamic Graph Enhanced Contrastive Learning for Chest...
240422_Thanh_LabSeminar[Dynamic Graph Enhanced Contrastive Learning for Chest...thanhdowork
 
[20240422_LabSeminar_Huy]Taming_Effect.pptx
[20240422_LabSeminar_Huy]Taming_Effect.pptx[20240422_LabSeminar_Huy]Taming_Effect.pptx
[20240422_LabSeminar_Huy]Taming_Effect.pptxthanhdowork
 
240422_Thuy_Labseminar[Large Graph Property Prediction via Graph Segment Trai...
240422_Thuy_Labseminar[Large Graph Property Prediction via Graph Segment Trai...240422_Thuy_Labseminar[Large Graph Property Prediction via Graph Segment Trai...
240422_Thuy_Labseminar[Large Graph Property Prediction via Graph Segment Trai...thanhdowork
 
[20240415_LabSeminar_Huy]Deciphering Spatio-Temporal Graph Forecasting: A Cau...
[20240415_LabSeminar_Huy]Deciphering Spatio-Temporal Graph Forecasting: A Cau...[20240415_LabSeminar_Huy]Deciphering Spatio-Temporal Graph Forecasting: A Cau...
[20240415_LabSeminar_Huy]Deciphering Spatio-Temporal Graph Forecasting: A Cau...thanhdowork
 
240315_Thanh_LabSeminar[G-TAD: Sub-Graph Localization for Temporal Action Det...
240315_Thanh_LabSeminar[G-TAD: Sub-Graph Localization for Temporal Action Det...240315_Thanh_LabSeminar[G-TAD: Sub-Graph Localization for Temporal Action Det...
240315_Thanh_LabSeminar[G-TAD: Sub-Graph Localization for Temporal Action Det...thanhdowork
 
240415_Thuy_Labseminar[Simple and Asymmetric Graph Contrastive Learning witho...
240415_Thuy_Labseminar[Simple and Asymmetric Graph Contrastive Learning witho...240415_Thuy_Labseminar[Simple and Asymmetric Graph Contrastive Learning witho...
240415_Thuy_Labseminar[Simple and Asymmetric Graph Contrastive Learning witho...thanhdowork
 
240115_Attention Is All You Need (2017 NIPS).pptx
240115_Attention Is All You Need (2017 NIPS).pptx240115_Attention Is All You Need (2017 NIPS).pptx
240115_Attention Is All You Need (2017 NIPS).pptxthanhdowork
 
240122_Attention Is All You Need (2017 NIPS)2.pptx
240122_Attention Is All You Need (2017 NIPS)2.pptx240122_Attention Is All You Need (2017 NIPS)2.pptx
240122_Attention Is All You Need (2017 NIPS)2.pptxthanhdowork
 
240226_Thanh_LabSeminar[Structure-Aware Transformer for Graph Representation ...
240226_Thanh_LabSeminar[Structure-Aware Transformer for Graph Representation ...240226_Thanh_LabSeminar[Structure-Aware Transformer for Graph Representation ...
240226_Thanh_LabSeminar[Structure-Aware Transformer for Graph Representation ...thanhdowork
 
[20240304_LabSeminar_Huy]DeepWalk: Online Learning of Social Representations....
[20240304_LabSeminar_Huy]DeepWalk: Online Learning of Social Representations....[20240304_LabSeminar_Huy]DeepWalk: Online Learning of Social Representations....
[20240304_LabSeminar_Huy]DeepWalk: Online Learning of Social Representations....thanhdowork
 
240304_Thanh_LabSeminar[Pure Transformers are Powerful Graph Learners].pptx
240304_Thanh_LabSeminar[Pure Transformers are Powerful Graph Learners].pptx240304_Thanh_LabSeminar[Pure Transformers are Powerful Graph Learners].pptx
240304_Thanh_LabSeminar[Pure Transformers are Powerful Graph Learners].pptxthanhdowork
 
240304_Thuy_Labseminar[SimGRACE: A Simple Framework for Graph Contrastive Lea...
240304_Thuy_Labseminar[SimGRACE: A Simple Framework for Graph Contrastive Lea...240304_Thuy_Labseminar[SimGRACE: A Simple Framework for Graph Contrastive Lea...
240304_Thuy_Labseminar[SimGRACE: A Simple Framework for Graph Contrastive Lea...thanhdowork
 
240311_JW_labseminar[Sequence to Sequence Learning with Neural Networks].pptx
240311_JW_labseminar[Sequence to Sequence Learning with Neural Networks].pptx240311_JW_labseminar[Sequence to Sequence Learning with Neural Networks].pptx
240311_JW_labseminar[Sequence to Sequence Learning with Neural Networks].pptxthanhdowork
 
[20240311_LabSeminar_Huy]LINE: Large-scale Information Network Embedding.pptx
[20240311_LabSeminar_Huy]LINE: Large-scale Information Network Embedding.pptx[20240311_LabSeminar_Huy]LINE: Large-scale Information Network Embedding.pptx
[20240311_LabSeminar_Huy]LINE: Large-scale Information Network Embedding.pptxthanhdowork
 
240311_Thanh_LabSeminar[Translating Embeddings for Modeling Multi-relational ...
240311_Thanh_LabSeminar[Translating Embeddings for Modeling Multi-relational ...240311_Thanh_LabSeminar[Translating Embeddings for Modeling Multi-relational ...
240311_Thanh_LabSeminar[Translating Embeddings for Modeling Multi-relational ...thanhdowork
 
240311_Thuy_Labseminar[Contrastive Multi-View Representation Learning on Grap...
240311_Thuy_Labseminar[Contrastive Multi-View Representation Learning on Grap...240311_Thuy_Labseminar[Contrastive Multi-View Representation Learning on Grap...
240311_Thuy_Labseminar[Contrastive Multi-View Representation Learning on Grap...thanhdowork
 
240318_JW_labseminar[Attention Is All You Need].pptx
240318_JW_labseminar[Attention Is All You Need].pptx240318_JW_labseminar[Attention Is All You Need].pptx
240318_JW_labseminar[Attention Is All You Need].pptxthanhdowork
 

More from thanhdowork (20)

[20240429_LabSeminar_Huy]Spatio-Temporal Graph Neural Point Process for Traff...
[20240429_LabSeminar_Huy]Spatio-Temporal Graph Neural Point Process for Traff...[20240429_LabSeminar_Huy]Spatio-Temporal Graph Neural Point Process for Traff...
[20240429_LabSeminar_Huy]Spatio-Temporal Graph Neural Point Process for Traff...
 
240429_Thanh_LabSeminar[TranSG: Transformer-Based Skeleton Graph Prototype Co...
240429_Thanh_LabSeminar[TranSG: Transformer-Based Skeleton Graph Prototype Co...240429_Thanh_LabSeminar[TranSG: Transformer-Based Skeleton Graph Prototype Co...
240429_Thanh_LabSeminar[TranSG: Transformer-Based Skeleton Graph Prototype Co...
 
240429_Thuy_Labseminar[Simplifying and Empowering Transformers for Large-Grap...
240429_Thuy_Labseminar[Simplifying and Empowering Transformers for Large-Grap...240429_Thuy_Labseminar[Simplifying and Empowering Transformers for Large-Grap...
240429_Thuy_Labseminar[Simplifying and Empowering Transformers for Large-Grap...
 
240422_Thanh_LabSeminar[Dynamic Graph Enhanced Contrastive Learning for Chest...
240422_Thanh_LabSeminar[Dynamic Graph Enhanced Contrastive Learning for Chest...240422_Thanh_LabSeminar[Dynamic Graph Enhanced Contrastive Learning for Chest...
240422_Thanh_LabSeminar[Dynamic Graph Enhanced Contrastive Learning for Chest...
 
[20240422_LabSeminar_Huy]Taming_Effect.pptx
[20240422_LabSeminar_Huy]Taming_Effect.pptx[20240422_LabSeminar_Huy]Taming_Effect.pptx
[20240422_LabSeminar_Huy]Taming_Effect.pptx
 
240422_Thuy_Labseminar[Large Graph Property Prediction via Graph Segment Trai...
240422_Thuy_Labseminar[Large Graph Property Prediction via Graph Segment Trai...240422_Thuy_Labseminar[Large Graph Property Prediction via Graph Segment Trai...
240422_Thuy_Labseminar[Large Graph Property Prediction via Graph Segment Trai...
 
[20240415_LabSeminar_Huy]Deciphering Spatio-Temporal Graph Forecasting: A Cau...
[20240415_LabSeminar_Huy]Deciphering Spatio-Temporal Graph Forecasting: A Cau...[20240415_LabSeminar_Huy]Deciphering Spatio-Temporal Graph Forecasting: A Cau...
[20240415_LabSeminar_Huy]Deciphering Spatio-Temporal Graph Forecasting: A Cau...
 
240315_Thanh_LabSeminar[G-TAD: Sub-Graph Localization for Temporal Action Det...
240315_Thanh_LabSeminar[G-TAD: Sub-Graph Localization for Temporal Action Det...240315_Thanh_LabSeminar[G-TAD: Sub-Graph Localization for Temporal Action Det...
240315_Thanh_LabSeminar[G-TAD: Sub-Graph Localization for Temporal Action Det...
 
240415_Thuy_Labseminar[Simple and Asymmetric Graph Contrastive Learning witho...
240415_Thuy_Labseminar[Simple and Asymmetric Graph Contrastive Learning witho...240415_Thuy_Labseminar[Simple and Asymmetric Graph Contrastive Learning witho...
240415_Thuy_Labseminar[Simple and Asymmetric Graph Contrastive Learning witho...
 
240115_Attention Is All You Need (2017 NIPS).pptx
240115_Attention Is All You Need (2017 NIPS).pptx240115_Attention Is All You Need (2017 NIPS).pptx
240115_Attention Is All You Need (2017 NIPS).pptx
 
240122_Attention Is All You Need (2017 NIPS)2.pptx
240122_Attention Is All You Need (2017 NIPS)2.pptx240122_Attention Is All You Need (2017 NIPS)2.pptx
240122_Attention Is All You Need (2017 NIPS)2.pptx
 
240226_Thanh_LabSeminar[Structure-Aware Transformer for Graph Representation ...
240226_Thanh_LabSeminar[Structure-Aware Transformer for Graph Representation ...240226_Thanh_LabSeminar[Structure-Aware Transformer for Graph Representation ...
240226_Thanh_LabSeminar[Structure-Aware Transformer for Graph Representation ...
 
[20240304_LabSeminar_Huy]DeepWalk: Online Learning of Social Representations....
[20240304_LabSeminar_Huy]DeepWalk: Online Learning of Social Representations....[20240304_LabSeminar_Huy]DeepWalk: Online Learning of Social Representations....
[20240304_LabSeminar_Huy]DeepWalk: Online Learning of Social Representations....
 
240304_Thanh_LabSeminar[Pure Transformers are Powerful Graph Learners].pptx
240304_Thanh_LabSeminar[Pure Transformers are Powerful Graph Learners].pptx240304_Thanh_LabSeminar[Pure Transformers are Powerful Graph Learners].pptx
240304_Thanh_LabSeminar[Pure Transformers are Powerful Graph Learners].pptx
 
240304_Thuy_Labseminar[SimGRACE: A Simple Framework for Graph Contrastive Lea...
240304_Thuy_Labseminar[SimGRACE: A Simple Framework for Graph Contrastive Lea...240304_Thuy_Labseminar[SimGRACE: A Simple Framework for Graph Contrastive Lea...
240304_Thuy_Labseminar[SimGRACE: A Simple Framework for Graph Contrastive Lea...
 
240311_JW_labseminar[Sequence to Sequence Learning with Neural Networks].pptx
240311_JW_labseminar[Sequence to Sequence Learning with Neural Networks].pptx240311_JW_labseminar[Sequence to Sequence Learning with Neural Networks].pptx
240311_JW_labseminar[Sequence to Sequence Learning with Neural Networks].pptx
 
[20240311_LabSeminar_Huy]LINE: Large-scale Information Network Embedding.pptx
[20240311_LabSeminar_Huy]LINE: Large-scale Information Network Embedding.pptx[20240311_LabSeminar_Huy]LINE: Large-scale Information Network Embedding.pptx
[20240311_LabSeminar_Huy]LINE: Large-scale Information Network Embedding.pptx
 
240311_Thanh_LabSeminar[Translating Embeddings for Modeling Multi-relational ...
240311_Thanh_LabSeminar[Translating Embeddings for Modeling Multi-relational ...240311_Thanh_LabSeminar[Translating Embeddings for Modeling Multi-relational ...
240311_Thanh_LabSeminar[Translating Embeddings for Modeling Multi-relational ...
 
240311_Thuy_Labseminar[Contrastive Multi-View Representation Learning on Grap...
240311_Thuy_Labseminar[Contrastive Multi-View Representation Learning on Grap...240311_Thuy_Labseminar[Contrastive Multi-View Representation Learning on Grap...
240311_Thuy_Labseminar[Contrastive Multi-View Representation Learning on Grap...
 
240318_JW_labseminar[Attention Is All You Need].pptx
240318_JW_labseminar[Attention Is All You Need].pptx240318_JW_labseminar[Attention Is All You Need].pptx
240318_JW_labseminar[Attention Is All You Need].pptx
 

Recently uploaded

ECONOMIC CONTEXT - LONG FORM TV DRAMA - PPT
ECONOMIC CONTEXT - LONG FORM TV DRAMA - PPTECONOMIC CONTEXT - LONG FORM TV DRAMA - PPT
ECONOMIC CONTEXT - LONG FORM TV DRAMA - PPTiammrhaywood
 
Organic Name Reactions for the students and aspirants of Chemistry12th.pptx
Organic Name Reactions  for the students and aspirants of Chemistry12th.pptxOrganic Name Reactions  for the students and aspirants of Chemistry12th.pptx
Organic Name Reactions for the students and aspirants of Chemistry12th.pptxVS Mahajan Coaching Centre
 
CARE OF CHILD IN INCUBATOR..........pptx
CARE OF CHILD IN INCUBATOR..........pptxCARE OF CHILD IN INCUBATOR..........pptx
CARE OF CHILD IN INCUBATOR..........pptxGaneshChakor2
 
Employee wellbeing at the workplace.pptx
Employee wellbeing at the workplace.pptxEmployee wellbeing at the workplace.pptx
Employee wellbeing at the workplace.pptxNirmalaLoungPoorunde1
 
Alper Gobel In Media Res Media Component
Alper Gobel In Media Res Media ComponentAlper Gobel In Media Res Media Component
Alper Gobel In Media Res Media ComponentInMediaRes1
 
mini mental status format.docx
mini    mental       status     format.docxmini    mental       status     format.docx
mini mental status format.docxPoojaSen20
 
Call Girls in Dwarka Mor Delhi Contact Us 9654467111
Call Girls in Dwarka Mor Delhi Contact Us 9654467111Call Girls in Dwarka Mor Delhi Contact Us 9654467111
Call Girls in Dwarka Mor Delhi Contact Us 9654467111Sapana Sha
 
Concept of Vouching. B.Com(Hons) /B.Compdf
Concept of Vouching. B.Com(Hons) /B.CompdfConcept of Vouching. B.Com(Hons) /B.Compdf
Concept of Vouching. B.Com(Hons) /B.CompdfUmakantAnnand
 
Mastering the Unannounced Regulatory Inspection
Mastering the Unannounced Regulatory InspectionMastering the Unannounced Regulatory Inspection
Mastering the Unannounced Regulatory InspectionSafetyChain Software
 
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptx
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptxPOINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptx
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptxSayali Powar
 
Solving Puzzles Benefits Everyone (English).pptx
Solving Puzzles Benefits Everyone (English).pptxSolving Puzzles Benefits Everyone (English).pptx
Solving Puzzles Benefits Everyone (English).pptxOH TEIK BIN
 
KSHARA STURA .pptx---KSHARA KARMA THERAPY (CAUSTIC THERAPY)————IMP.OF KSHARA ...
KSHARA STURA .pptx---KSHARA KARMA THERAPY (CAUSTIC THERAPY)————IMP.OF KSHARA ...KSHARA STURA .pptx---KSHARA KARMA THERAPY (CAUSTIC THERAPY)————IMP.OF KSHARA ...
KSHARA STURA .pptx---KSHARA KARMA THERAPY (CAUSTIC THERAPY)————IMP.OF KSHARA ...M56BOOKSTORE PRODUCT/SERVICE
 
The basics of sentences session 2pptx copy.pptx
The basics of sentences session 2pptx copy.pptxThe basics of sentences session 2pptx copy.pptx
The basics of sentences session 2pptx copy.pptxheathfieldcps1
 
A Critique of the Proposed National Education Policy Reform
A Critique of the Proposed National Education Policy ReformA Critique of the Proposed National Education Policy Reform
A Critique of the Proposed National Education Policy ReformChameera Dedduwage
 
Interactive Powerpoint_How to Master effective communication
Interactive Powerpoint_How to Master effective communicationInteractive Powerpoint_How to Master effective communication
Interactive Powerpoint_How to Master effective communicationnomboosow
 
Introduction to ArtificiaI Intelligence in Higher Education
Introduction to ArtificiaI Intelligence in Higher EducationIntroduction to ArtificiaI Intelligence in Higher Education
Introduction to ArtificiaI Intelligence in Higher Educationpboyjonauth
 
18-04-UA_REPORT_MEDIALITERAСY_INDEX-DM_23-1-final-eng.pdf
18-04-UA_REPORT_MEDIALITERAСY_INDEX-DM_23-1-final-eng.pdf18-04-UA_REPORT_MEDIALITERAСY_INDEX-DM_23-1-final-eng.pdf
18-04-UA_REPORT_MEDIALITERAСY_INDEX-DM_23-1-final-eng.pdfssuser54595a
 

Recently uploaded (20)

ECONOMIC CONTEXT - LONG FORM TV DRAMA - PPT
ECONOMIC CONTEXT - LONG FORM TV DRAMA - PPTECONOMIC CONTEXT - LONG FORM TV DRAMA - PPT
ECONOMIC CONTEXT - LONG FORM TV DRAMA - PPT
 
Organic Name Reactions for the students and aspirants of Chemistry12th.pptx
Organic Name Reactions  for the students and aspirants of Chemistry12th.pptxOrganic Name Reactions  for the students and aspirants of Chemistry12th.pptx
Organic Name Reactions for the students and aspirants of Chemistry12th.pptx
 
Model Call Girl in Bikash Puri Delhi reach out to us at 🔝9953056974🔝
Model Call Girl in Bikash Puri  Delhi reach out to us at 🔝9953056974🔝Model Call Girl in Bikash Puri  Delhi reach out to us at 🔝9953056974🔝
Model Call Girl in Bikash Puri Delhi reach out to us at 🔝9953056974🔝
 
CARE OF CHILD IN INCUBATOR..........pptx
CARE OF CHILD IN INCUBATOR..........pptxCARE OF CHILD IN INCUBATOR..........pptx
CARE OF CHILD IN INCUBATOR..........pptx
 
Employee wellbeing at the workplace.pptx
Employee wellbeing at the workplace.pptxEmployee wellbeing at the workplace.pptx
Employee wellbeing at the workplace.pptx
 
Alper Gobel In Media Res Media Component
Alper Gobel In Media Res Media ComponentAlper Gobel In Media Res Media Component
Alper Gobel In Media Res Media Component
 
mini mental status format.docx
mini    mental       status     format.docxmini    mental       status     format.docx
mini mental status format.docx
 
Call Girls in Dwarka Mor Delhi Contact Us 9654467111
Call Girls in Dwarka Mor Delhi Contact Us 9654467111Call Girls in Dwarka Mor Delhi Contact Us 9654467111
Call Girls in Dwarka Mor Delhi Contact Us 9654467111
 
Concept of Vouching. B.Com(Hons) /B.Compdf
Concept of Vouching. B.Com(Hons) /B.CompdfConcept of Vouching. B.Com(Hons) /B.Compdf
Concept of Vouching. B.Com(Hons) /B.Compdf
 
TataKelola dan KamSiber Kecerdasan Buatan v022.pdf
TataKelola dan KamSiber Kecerdasan Buatan v022.pdfTataKelola dan KamSiber Kecerdasan Buatan v022.pdf
TataKelola dan KamSiber Kecerdasan Buatan v022.pdf
 
Mastering the Unannounced Regulatory Inspection
Mastering the Unannounced Regulatory InspectionMastering the Unannounced Regulatory Inspection
Mastering the Unannounced Regulatory Inspection
 
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptx
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptxPOINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptx
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptx
 
Solving Puzzles Benefits Everyone (English).pptx
Solving Puzzles Benefits Everyone (English).pptxSolving Puzzles Benefits Everyone (English).pptx
Solving Puzzles Benefits Everyone (English).pptx
 
Staff of Color (SOC) Retention Efforts DDSD
Staff of Color (SOC) Retention Efforts DDSDStaff of Color (SOC) Retention Efforts DDSD
Staff of Color (SOC) Retention Efforts DDSD
 
KSHARA STURA .pptx---KSHARA KARMA THERAPY (CAUSTIC THERAPY)————IMP.OF KSHARA ...
KSHARA STURA .pptx---KSHARA KARMA THERAPY (CAUSTIC THERAPY)————IMP.OF KSHARA ...KSHARA STURA .pptx---KSHARA KARMA THERAPY (CAUSTIC THERAPY)————IMP.OF KSHARA ...
KSHARA STURA .pptx---KSHARA KARMA THERAPY (CAUSTIC THERAPY)————IMP.OF KSHARA ...
 
The basics of sentences session 2pptx copy.pptx
The basics of sentences session 2pptx copy.pptxThe basics of sentences session 2pptx copy.pptx
The basics of sentences session 2pptx copy.pptx
 
A Critique of the Proposed National Education Policy Reform
A Critique of the Proposed National Education Policy ReformA Critique of the Proposed National Education Policy Reform
A Critique of the Proposed National Education Policy Reform
 
Interactive Powerpoint_How to Master effective communication
Interactive Powerpoint_How to Master effective communicationInteractive Powerpoint_How to Master effective communication
Interactive Powerpoint_How to Master effective communication
 
Introduction to ArtificiaI Intelligence in Higher Education
Introduction to ArtificiaI Intelligence in Higher EducationIntroduction to ArtificiaI Intelligence in Higher Education
Introduction to ArtificiaI Intelligence in Higher Education
 
18-04-UA_REPORT_MEDIALITERAСY_INDEX-DM_23-1-final-eng.pdf
18-04-UA_REPORT_MEDIALITERAСY_INDEX-DM_23-1-final-eng.pdf18-04-UA_REPORT_MEDIALITERAСY_INDEX-DM_23-1-final-eng.pdf
18-04-UA_REPORT_MEDIALITERAСY_INDEX-DM_23-1-final-eng.pdf
 

240325_JW_labseminar[node2vec: Scalable Feature Learning for Networks].pptx

  • 1. Jin-Woo Jeong Network Science Lab Dept. of Mathematics The Catholic University of Korea E-mail: zeus0208b@gmail.com Aditya Grover, Jure Leskovec
  • 2. 1  Introduction • Motivation • Introduction  Feature Learning Framework • Classic search strategies • Node2vec • Random Walks • Search bias 𝛼 • The node2vec algorithm • Learning edge features Experiments • Case Study: Les Misérables network • Multi-label classification • Link prediction Conclusion Q/A
  • 3. 2 Introduction Motivation  Supervised machine learning algorithms in network prediction tasks require informative, discriminating, and independent features. Typically, these features are constructed through hand-engineering based on domain-specific expertise. However, this process is tedious and the resulting features may not generalize well across different prediction tasks.  Previous techniques fail to satisfactorily define and optimize a reasonable objective required for scalable unsupervised feature learning in networks. Classic approaches based on linear and non-linear dimensionality reduction techniques such as Principal Component Analysis, Multi-Dimensional Scaling and their extensions optimize an objective that transforms a representative data matrix of the network such that it maximizes the variance of the data representation. Consequently, these approaches invariably involve eigendecomposition of the appropriate data matrix which is expensive for large real-world networks. Moreover, the resulting latent representations give poor performance on various prediction tasks over networks.
  • 4. 3 Introduction Introduction  A node could be organized based on communities they belong to (i.e., homophily); in other cases, the organization could be based on the structural roles of nodes in the network (i.e., structural equivalence).  Real-world networks commonly exhibit a mixture of such equivalences.
  • 5. 4 Introduction Introduction  they propose node2vec, a semi-supervised algorithm for scalable feature learning in networks. they optimize a custom graph-based objective function using SGD motivated by prior work on natural language processing.  Intuitively, their approach returns feature representations that maximize the likelihood of preserving network neighborhoods of nodes in a d-dimensional feature space. They use a 2nd order random walk approach to generate (sample) network neighborhoods for nodes.  Contributions  1. They propose node2vec, an efficient scalable algorithm for feature learning in networks that efficiently optimizes a novel network-aware, neighborhood preserving objective using SGD.  2. They show how node2vec is in accordance with established principles in network science, providing flexibility in discovering representations conforming to different equivalences.  3. They extend node2vec and other feature learning methods based on neighborhood preserving objectives, from nodes to pairs of nodes for edge-based prediction tasks.  4. They empirically evaluate node2vec for multi-label classification and link prediction on several real- world datasets.
  • 6. 5 Feature Learning Framework Feature Learning Framework  They formulate learning in networks as a maximum likelihood optimization problem. • 𝐺 = 𝑉, 𝐸 be a given graph • 𝑓 ∶ 𝑉 → ℝ𝑑 be the mapping function from nodes to feature representations. • 𝑓 is a matrix of size 𝑉 × 𝑑 parameters. • 𝑑 be a number of dimension of feature representation. • ∀ 𝑢 ∈ 𝑉, 𝑁𝑆 𝑢 ⊂ 𝑉 𝑁𝑆 𝑢 : a network neighborhood of node u generated through a neighborhood sampling strategy S. where
  • 7. 6 Feature Learning Framework Classic search strategies • Breadth-first Sampling (BFS) • The neighborhood 𝑁𝑆 is restricted to nodes which are immediate neighbors of the source. For example, in Figure 1 for a neighborhood of size k = 3, BFS samples nodes s1, s2, s3. • Depth-first Sampling (DFS) • The neighborhood consists of nodes sequentially sampled at increasing distances from the source node. In Figure 1, DFS samples s4, s5, s6.
  • 8. 7 Feature Learning Framework node2vec • Random Walks • 𝑢 : A source node • 𝑙 : Walk of fixed length • 𝑐𝑖 : 𝑖th node in the walk • 𝑐0 = 𝑢 • 𝜋𝑣𝑥 : unnormalized transition probability • 𝑍 : normalizing constant
  • 9. 8 Feature Learning Framework node2vec • Search bias 𝜶 𝜋𝑣𝑥 = 𝛼𝑝𝑞(𝑡, 𝑥) ∙ 𝜔𝑣𝑥 • Return parameter, p • Parameter p controls the likelihood of immediately revisiting a node in the walk. • In-out parameter, q • Parameter q allows the search to differentiate between “inward” and “outward” nodes.
  • 10. 9 Feature Learning Framework node2vec  The node2vec algorithm • SkipGram algorithm in DEEPWALK Φ → 𝑓 𝑤 → 𝑘 𝑊𝑣𝑖 → 𝑤𝑎𝑙𝑘
  • 11. 10 Feature Learning Framework Learning edge features  Given two nodes 𝑢 and 𝑣, they define a binary operator ◦ over the corresponding feature vectors f(𝑢) and f(𝑣) in order to generate a representation g(𝑢, 𝑣) such that 𝑔 ∶ 𝑉 × 𝑉 → ℝ𝑑′ where 𝑑′ is the representation size for the pair g 𝑢, 𝑣 .  They consider several choices for the operator ◦ such that 𝑑′ = 𝑑 which are summarized in Table 1
  • 12. 11 Experiments Case Study: Les Misérables network  They use a network where nodes correspond to characters in the novel Les Misérables and edges connect coappearing characters. The network has 77 nodes and 254 edges.  They set d = 16 and run node2vec to learn feature representation for every node in the network.  The feature representations are clustered using k- means. p = 1, q = 0.5 p = 1, q = 2 homophily structural equivalence
  • 16. 15 Conclusion Conclusion  In this paper, they researched feature learning in networks as a search-based optimization problem. This perspective provides us with several advantages. It can elucidate traditional search strategies based on the balance between exploration and exploitation. Additionally, it imparts interpretability to the learned representations when applied in prediction tasks.  The search strategy in node2vec allows flexible exploration and control of network neighborhoods through parameters p and q. While these search parameters have intuitive interpretations, we achieve the best results on complex networks when we can learn them directly from data.  From a practical standpoint, node2vec is scalable and robust. We demonstrated the superiority of extending node embeddings over widely used heuristic scores for link prediction. Their methodology allows additional binary operators beyond those listed in Table 1.  They aim to explore the reasons behind the success of the Hadamard operator over others in future work, as well as establish interpretable equivalence notions for edges based on the search parameters.
  • 17. 16 Q & A Q / A

Editor's Notes

  1. thank you, the presentation is concluded