SlideShare a Scribd company logo
1 of 23
GRAPH NEURAL NETWORKS
FOR NODE CLASSIFICATION
AGENDA
• Abstract
• Background and Problem Definition
• Supervised Graph Neural Networks
• General Frame work of Graph Neural Networks (GNN)
• Graph Convolutional Networks (GCN)
• Graph Attention Networks
• Neural Message Passing Networks(MPNN)
• Continuous Graph Neural Networks
• Multi-Scale Spectral Graph Convolutional Networks
• Unsupervised Graph Neural Networks
• Variational Graph Auto-Encoders
• Deep Graph Infomax
• Over-smoothing Problem
ABSTRACT
• In this chapter, we focus on a fundamental task on graphs : node
classification.
BACKGROUND AND PROBLEM DEFINITION
• A fundamental task on graphs is node classification, which tries to
classify the nodes into a few predefined categories.
• To make effective prediction, a critical problem is to have very effective
node representations, which largely determine the performance of node
classification.
BACKGROUND AND PROBLEM DEFINITION
• N is the total number of nodes.
• C is the number of features for each node.
• F is the dimension of node representations.
The goal of graph neural networks is
to learn effective node representations
(denoted as H ∈ R N×F) by combining
the graph structure information and the node attributes, which are further used for
node classification.
Supervised Graph Neural Networks
• General Frame work of Graph Neural Networks (GNN)
• Graph Convolutional Networks (GCN)
• Graph Attention Networks
• Neural Message Passing Networks(MPNN)
• Continuous Graph Neural Networks
• Multi-Scale Spectral Graph Convolutional Networks
1- General Frame work of Graph Neural Networks (GNN)
• The essential idea of graph neural networks is to iteratively update the
node representations by combining the representations of their neighbors
and their own representations.
1- General Frame work of Graph Neural Networks (GNN)
Starting from the initial node representation H0 = X,
in each layer we have two important functions:
• AGGREGATE : which tries to aggregate the information from the neighbours
of each node;
• COMBINE : which tries to update the node representations by combining the
aggregated information from neighbors with the current node
representations.
2- Graph Convolutional Networks (GCN)
• Which is now the most popular graph neural network architecture due to its
simplicity and effectiveness in a variety of tasks and applications.
• Specifically, the node representations in each layer is updated according to a
propagation rule.
2- Graph Convolutional Networks (GCN)
• AGGREGATE function is defined as the weighted average of the
neighbor node representations. The weight of the neighbor j is
determined by the weight of the edge between i and j (i.e. Ai j
normalized by the degrees of the two nodes)
• COMBINE function is defined as the summation of the aggregated
messages and the node representation itself, in which the node
representation is normalized by its own degree.
3- Graph Attention Networks
• In GCNs, for a target node i, the importance of a neighbor j is determined
by the weight of their edge Ai j (normalized by their node degrees).
• However, in practice, the input graph may be noisy.
• The edge weights may not be able to reflect the true strength between two
nodes.
• As a result, a more principled approach would be to automatically learn the
importance of each neighbor. Graph Attention Networks (a.k.a. GAT) is built
on this idea and try to learn the importance of each neighbor based on the
Attention mechanism.
3- Graph Attention Networks
• Graph Attention Layer: defnes how to transfer the hidden node representations at layer k −1 (denoted
as H k−1) to the new node representations H.
• In order to guarantee suffcient expressive power to transform the lower-level node representations to
higher-level node representations, a shared linear transformation is applied to every node.
• Afterwards, self-attention is defned on the nodes, which measures the attention coeffcients for any
pair of nodes through a shared attentional mechanism (indicates the relationship strength between
node i and j).
• And to make the coeffcients comparable across different nodes, the attention coeffcients are usually
normalized with the softmax function:
• The new node representation is a linear combination of the neighboring node representations with the
weights determined by the attention coeffcients (with a potential nonlinear transformation)
4- Neural Message Passing Networks(MPNN)
• MPNN is actually very general, provides a general framework of
graph neural networks, and could be used for the task of node
classifcation as well.
• The essential idea of MPNN is formulating existing graph neural
networks as a general framework of neural message passing
among nodes.
• In MPNNs, there are two important functions including
Message and Updating function:
4- Neural Message Passing Networks(MPNN)
• The MPNN framework is very similar to the general framework
• The AGGREGATE function defned here is simply a summation of all the
messages from the neighbors.
• The COMBINE function is the same as the node Updating function: which
combines the aggregated messages from the neighbors and the node
representation itself.
5- Continuous Graph Neural Networks
• The above graph neural networks iteratively update the node representations with different
kinds of graph convolutional layers. Essentially, these approaches model the discrete
dynamics of node representations with GNNs.
• Xhonneux et al (2020) proposed the continuous graph neural networks (CGNNs), which
generalizes existing graph neural networks with discrete dynamics to continuous settings, i.e.,
trying to model the continuous dynamics of node representations.
• The key idea is how to characterize the continuous dynamics of node representations, i.e. the
derivatives of node representation w.r.t. time.
• The derivatives of the node representations are defned as a combination of the node
representation itself, the representations of its neighbors, and the initial status of the nodes.
5- Continuous Graph Neural Networks
• Specifcally, two different variants of node dynamics are
introduced.
• The frst model assumes that different dimensions of node
presentations (a.k.a. feature channels) are independent;
• The second model is more fexible, which allows different feature
channels to interact with each other.
6- Multi-Scale Spectral Graph Convolutional
Networks
UNSUPERVISED GRAPH NEURAL NETWORKS
• Variational Graph Auto-Encoders
• Deep graph infomax
VARIATIONAL GRAPH AUTO-ENCODERS
The VGAE model consists of an encoder, a decoder, and a prior .
• Encoder
The goal of the encoder is to learn a distribution of latent variables associated with each node
conditioning on the node features X and the adjacency matrix A. (qφ (Z|X,A)).
• Decoder
Given sampled latent variables, the decoder aims at predicting the connectivities among nodes.
Aopts a simple dot-product based predictor.
• Prior
The prior distributions over the latent variables are simply set to independent zero-mean Gaussians with unit variances,
This prior is fixed throughout the learning.
DEEP GRAPH INFOMAX
• Deep Graph Infomax is an unsupervised learning framework that
learns graph representations via the principle of mutual information
maximization.
OVER-SMOOTHING PROBLEM
• Training deep graph neural networks by stacking multiple layers of graph neural networks
usually yields inferior results, which is a common problem observed in many different
graph neural network architectures.
• This is mainly due to the problem of over-smoothing.
• The graph convolutional network is a special case of Laplacian smoothing:
• The Laplacian smoothing will push nodes belonging to the same clusters to take similar
representations, which are benefcial for downstream tasks such as node classifcation.
• However, when the GCNs go deep, the node representations suffer from the problem of
over-smoothing, i.e., all the nodes will have similar representations. As a result, the
performance on downstream tasks suffer as well.
OVER-SMOOTHING PROBLEM
• PairNorm: a method for alleviating the problem of over-smoothing
when GNNs go deep.
• The essential idea of PairNorm is to keep the total pairwise squared
distance (TPSD) of node representations unchanged, which is the
same as that of the original node feature X.

More Related Content

Similar to Chapter 4 better.pptx

intro-to-cnn-April_2020.pptx
intro-to-cnn-April_2020.pptxintro-to-cnn-April_2020.pptx
intro-to-cnn-April_2020.pptxssuser3aa461
 
240415_Thuy_Labseminar[Simple and Asymmetric Graph Contrastive Learning witho...
240415_Thuy_Labseminar[Simple and Asymmetric Graph Contrastive Learning witho...240415_Thuy_Labseminar[Simple and Asymmetric Graph Contrastive Learning witho...
240415_Thuy_Labseminar[Simple and Asymmetric Graph Contrastive Learning witho...thanhdowork
 
A Survey of Convolutional Neural Networks
A Survey of Convolutional Neural NetworksA Survey of Convolutional Neural Networks
A Survey of Convolutional Neural NetworksRimzim Thube
 
Review-image-segmentation-by-deep-learning
Review-image-segmentation-by-deep-learningReview-image-segmentation-by-deep-learning
Review-image-segmentation-by-deep-learningTrong-An Bui
 
201907 AutoML and Neural Architecture Search
201907 AutoML and Neural Architecture Search201907 AutoML and Neural Architecture Search
201907 AutoML and Neural Architecture SearchDaeJin Kim
 
Convolutional Neural Networks (CNN)
Convolutional Neural Networks (CNN)Convolutional Neural Networks (CNN)
Convolutional Neural Networks (CNN)Gaurav Mittal
 
[20240422_LabSeminar_Huy]Taming_Effect.pptx
[20240422_LabSeminar_Huy]Taming_Effect.pptx[20240422_LabSeminar_Huy]Taming_Effect.pptx
[20240422_LabSeminar_Huy]Taming_Effect.pptxthanhdowork
 
Towards Deep Attention in Graph Neural Networks: Problems and Remedies.pptx
Towards Deep Attention in Graph Neural Networks: Problems and Remedies.pptxTowards Deep Attention in Graph Neural Networks: Problems and Remedies.pptx
Towards Deep Attention in Graph Neural Networks: Problems and Remedies.pptxssuser2624f71
 
Weisfeiler and Leman Go Neural: Higher-order Graph Neural Networks, arXiv e-...
Weisfeiler and Leman Go Neural: Higher-order Graph Neural Networks, arXiv e-...Weisfeiler and Leman Go Neural: Higher-order Graph Neural Networks, arXiv e-...
Weisfeiler and Leman Go Neural: Higher-order Graph Neural Networks, arXiv e-...ssuser2624f71
 
Alberto Massidda - Scenes from a memory - Codemotion Rome 2019
Alberto Massidda - Scenes from a memory - Codemotion Rome 2019Alberto Massidda - Scenes from a memory - Codemotion Rome 2019
Alberto Massidda - Scenes from a memory - Codemotion Rome 2019Codemotion
 
Convolutional Neural Networks - Veronica Vilaplana - UPC Barcelona 2018
Convolutional Neural Networks - Veronica Vilaplana - UPC Barcelona 2018Convolutional Neural Networks - Veronica Vilaplana - UPC Barcelona 2018
Convolutional Neural Networks - Veronica Vilaplana - UPC Barcelona 2018Universitat Politècnica de Catalunya
 
Wits presentation 6_28072015
Wits presentation 6_28072015Wits presentation 6_28072015
Wits presentation 6_28072015Beatrice van Eden
 
Exploring Randomly Wired Neural Networks for Image Recognition
Exploring Randomly Wired Neural Networks for Image RecognitionExploring Randomly Wired Neural Networks for Image Recognition
Exploring Randomly Wired Neural Networks for Image RecognitionYongsu Baek
 
Neural Radiance Fields & Neural Rendering.pdf
Neural Radiance Fields & Neural Rendering.pdfNeural Radiance Fields & Neural Rendering.pdf
Neural Radiance Fields & Neural Rendering.pdfNavneetPaul2
 
RETHINKING THE EXPRESSIVE POWER OF GNNS VIA GRAPH BICONNECTIVITY.pptx
RETHINKING THE EXPRESSIVE POWER OF GNNS VIA GRAPH BICONNECTIVITY.pptxRETHINKING THE EXPRESSIVE POWER OF GNNS VIA GRAPH BICONNECTIVITY.pptx
RETHINKING THE EXPRESSIVE POWER OF GNNS VIA GRAPH BICONNECTIVITY.pptxssuser2624f71
 
PR-183: MixNet: Mixed Depthwise Convolutional Kernels
PR-183: MixNet: Mixed Depthwise Convolutional KernelsPR-183: MixNet: Mixed Depthwise Convolutional Kernels
PR-183: MixNet: Mixed Depthwise Convolutional KernelsJinwon Lee
 

Similar to Chapter 4 better.pptx (20)

NS-CUK Seminar: S.T.Nguyen, Review on "On Generalized Degree Fairness in Grap...
NS-CUK Seminar: S.T.Nguyen, Review on "On Generalized Degree Fairness in Grap...NS-CUK Seminar: S.T.Nguyen, Review on "On Generalized Degree Fairness in Grap...
NS-CUK Seminar: S.T.Nguyen, Review on "On Generalized Degree Fairness in Grap...
 
intro-to-cnn-April_2020.pptx
intro-to-cnn-April_2020.pptxintro-to-cnn-April_2020.pptx
intro-to-cnn-April_2020.pptx
 
240415_Thuy_Labseminar[Simple and Asymmetric Graph Contrastive Learning witho...
240415_Thuy_Labseminar[Simple and Asymmetric Graph Contrastive Learning witho...240415_Thuy_Labseminar[Simple and Asymmetric Graph Contrastive Learning witho...
240415_Thuy_Labseminar[Simple and Asymmetric Graph Contrastive Learning witho...
 
A Survey of Convolutional Neural Networks
A Survey of Convolutional Neural NetworksA Survey of Convolutional Neural Networks
A Survey of Convolutional Neural Networks
 
Review-image-segmentation-by-deep-learning
Review-image-segmentation-by-deep-learningReview-image-segmentation-by-deep-learning
Review-image-segmentation-by-deep-learning
 
201907 AutoML and Neural Architecture Search
201907 AutoML and Neural Architecture Search201907 AutoML and Neural Architecture Search
201907 AutoML and Neural Architecture Search
 
Deep Learning
Deep LearningDeep Learning
Deep Learning
 
Convolutional Neural Networks (CNN)
Convolutional Neural Networks (CNN)Convolutional Neural Networks (CNN)
Convolutional Neural Networks (CNN)
 
[20240422_LabSeminar_Huy]Taming_Effect.pptx
[20240422_LabSeminar_Huy]Taming_Effect.pptx[20240422_LabSeminar_Huy]Taming_Effect.pptx
[20240422_LabSeminar_Huy]Taming_Effect.pptx
 
Towards Deep Attention in Graph Neural Networks: Problems and Remedies.pptx
Towards Deep Attention in Graph Neural Networks: Problems and Remedies.pptxTowards Deep Attention in Graph Neural Networks: Problems and Remedies.pptx
Towards Deep Attention in Graph Neural Networks: Problems and Remedies.pptx
 
Weisfeiler and Leman Go Neural: Higher-order Graph Neural Networks, arXiv e-...
Weisfeiler and Leman Go Neural: Higher-order Graph Neural Networks, arXiv e-...Weisfeiler and Leman Go Neural: Higher-order Graph Neural Networks, arXiv e-...
Weisfeiler and Leman Go Neural: Higher-order Graph Neural Networks, arXiv e-...
 
Alberto Massidda - Scenes from a memory - Codemotion Rome 2019
Alberto Massidda - Scenes from a memory - Codemotion Rome 2019Alberto Massidda - Scenes from a memory - Codemotion Rome 2019
Alberto Massidda - Scenes from a memory - Codemotion Rome 2019
 
Convolutional Neural Networks - Veronica Vilaplana - UPC Barcelona 2018
Convolutional Neural Networks - Veronica Vilaplana - UPC Barcelona 2018Convolutional Neural Networks - Veronica Vilaplana - UPC Barcelona 2018
Convolutional Neural Networks - Veronica Vilaplana - UPC Barcelona 2018
 
Wits presentation 6_28072015
Wits presentation 6_28072015Wits presentation 6_28072015
Wits presentation 6_28072015
 
Exploring Randomly Wired Neural Networks for Image Recognition
Exploring Randomly Wired Neural Networks for Image RecognitionExploring Randomly Wired Neural Networks for Image Recognition
Exploring Randomly Wired Neural Networks for Image Recognition
 
Neural Radiance Fields & Neural Rendering.pdf
Neural Radiance Fields & Neural Rendering.pdfNeural Radiance Fields & Neural Rendering.pdf
Neural Radiance Fields & Neural Rendering.pdf
 
Multidimensional RNN
Multidimensional RNNMultidimensional RNN
Multidimensional RNN
 
NS-CUK Seminar: V.T.Hoang, Review on "Everything is Connected: Graph Neural N...
NS-CUK Seminar: V.T.Hoang, Review on "Everything is Connected: Graph Neural N...NS-CUK Seminar: V.T.Hoang, Review on "Everything is Connected: Graph Neural N...
NS-CUK Seminar: V.T.Hoang, Review on "Everything is Connected: Graph Neural N...
 
RETHINKING THE EXPRESSIVE POWER OF GNNS VIA GRAPH BICONNECTIVITY.pptx
RETHINKING THE EXPRESSIVE POWER OF GNNS VIA GRAPH BICONNECTIVITY.pptxRETHINKING THE EXPRESSIVE POWER OF GNNS VIA GRAPH BICONNECTIVITY.pptx
RETHINKING THE EXPRESSIVE POWER OF GNNS VIA GRAPH BICONNECTIVITY.pptx
 
PR-183: MixNet: Mixed Depthwise Convolutional Kernels
PR-183: MixNet: Mixed Depthwise Convolutional KernelsPR-183: MixNet: Mixed Depthwise Convolutional Kernels
PR-183: MixNet: Mixed Depthwise Convolutional Kernels
 

Recently uploaded

Breaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path MountBreaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path MountPuma Security, LLC
 
08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking Men08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking MenDelhi Call girls
 
08448380779 Call Girls In Friends Colony Women Seeking Men
08448380779 Call Girls In Friends Colony Women Seeking Men08448380779 Call Girls In Friends Colony Women Seeking Men
08448380779 Call Girls In Friends Colony Women Seeking MenDelhi Call girls
 
Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...
Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...
Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...HostedbyConfluent
 
Kotlin Multiplatform & Compose Multiplatform - Starter kit for pragmatics
Kotlin Multiplatform & Compose Multiplatform - Starter kit for pragmaticsKotlin Multiplatform & Compose Multiplatform - Starter kit for pragmatics
Kotlin Multiplatform & Compose Multiplatform - Starter kit for pragmaticscarlostorres15106
 
Slack Application Development 101 Slides
Slack Application Development 101 SlidesSlack Application Development 101 Slides
Slack Application Development 101 Slidespraypatel2
 
Install Stable Diffusion in windows machine
Install Stable Diffusion in windows machineInstall Stable Diffusion in windows machine
Install Stable Diffusion in windows machinePadma Pradeep
 
How to Remove Document Management Hurdles with X-Docs?
How to Remove Document Management Hurdles with X-Docs?How to Remove Document Management Hurdles with X-Docs?
How to Remove Document Management Hurdles with X-Docs?XfilesPro
 
Snow Chain-Integrated Tire for a Safe Drive on Winter Roads
Snow Chain-Integrated Tire for a Safe Drive on Winter RoadsSnow Chain-Integrated Tire for a Safe Drive on Winter Roads
Snow Chain-Integrated Tire for a Safe Drive on Winter RoadsHyundai Motor Group
 
SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024Scott Keck-Warren
 
GenCyber Cyber Security Day Presentation
GenCyber Cyber Security Day PresentationGenCyber Cyber Security Day Presentation
GenCyber Cyber Security Day PresentationMichael W. Hawkins
 
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | DelhiFULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhisoniya singh
 
Human Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR SystemsHuman Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR SystemsMark Billinghurst
 
Key Features Of Token Development (1).pptx
Key  Features Of Token  Development (1).pptxKey  Features Of Token  Development (1).pptx
Key Features Of Token Development (1).pptxLBM Solutions
 
Benefits Of Flutter Compared To Other Frameworks
Benefits Of Flutter Compared To Other FrameworksBenefits Of Flutter Compared To Other Frameworks
Benefits Of Flutter Compared To Other FrameworksSoftradix Technologies
 
How to convert PDF to text with Nanonets
How to convert PDF to text with NanonetsHow to convert PDF to text with Nanonets
How to convert PDF to text with Nanonetsnaman860154
 
Enhancing Worker Digital Experience: A Hands-on Workshop for Partners
Enhancing Worker Digital Experience: A Hands-on Workshop for PartnersEnhancing Worker Digital Experience: A Hands-on Workshop for Partners
Enhancing Worker Digital Experience: A Hands-on Workshop for PartnersThousandEyes
 
Unblocking The Main Thread Solving ANRs and Frozen Frames
Unblocking The Main Thread Solving ANRs and Frozen FramesUnblocking The Main Thread Solving ANRs and Frozen Frames
Unblocking The Main Thread Solving ANRs and Frozen FramesSinan KOZAK
 
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...Integration and Automation in Practice: CI/CD in Mule Integration and Automat...
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...Patryk Bandurski
 
Presentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreterPresentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreternaman860154
 

Recently uploaded (20)

Breaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path MountBreaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path Mount
 
08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking Men08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking Men
 
08448380779 Call Girls In Friends Colony Women Seeking Men
08448380779 Call Girls In Friends Colony Women Seeking Men08448380779 Call Girls In Friends Colony Women Seeking Men
08448380779 Call Girls In Friends Colony Women Seeking Men
 
Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...
Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...
Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...
 
Kotlin Multiplatform & Compose Multiplatform - Starter kit for pragmatics
Kotlin Multiplatform & Compose Multiplatform - Starter kit for pragmaticsKotlin Multiplatform & Compose Multiplatform - Starter kit for pragmatics
Kotlin Multiplatform & Compose Multiplatform - Starter kit for pragmatics
 
Slack Application Development 101 Slides
Slack Application Development 101 SlidesSlack Application Development 101 Slides
Slack Application Development 101 Slides
 
Install Stable Diffusion in windows machine
Install Stable Diffusion in windows machineInstall Stable Diffusion in windows machine
Install Stable Diffusion in windows machine
 
How to Remove Document Management Hurdles with X-Docs?
How to Remove Document Management Hurdles with X-Docs?How to Remove Document Management Hurdles with X-Docs?
How to Remove Document Management Hurdles with X-Docs?
 
Snow Chain-Integrated Tire for a Safe Drive on Winter Roads
Snow Chain-Integrated Tire for a Safe Drive on Winter RoadsSnow Chain-Integrated Tire for a Safe Drive on Winter Roads
Snow Chain-Integrated Tire for a Safe Drive on Winter Roads
 
SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024
 
GenCyber Cyber Security Day Presentation
GenCyber Cyber Security Day PresentationGenCyber Cyber Security Day Presentation
GenCyber Cyber Security Day Presentation
 
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | DelhiFULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
 
Human Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR SystemsHuman Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR Systems
 
Key Features Of Token Development (1).pptx
Key  Features Of Token  Development (1).pptxKey  Features Of Token  Development (1).pptx
Key Features Of Token Development (1).pptx
 
Benefits Of Flutter Compared To Other Frameworks
Benefits Of Flutter Compared To Other FrameworksBenefits Of Flutter Compared To Other Frameworks
Benefits Of Flutter Compared To Other Frameworks
 
How to convert PDF to text with Nanonets
How to convert PDF to text with NanonetsHow to convert PDF to text with Nanonets
How to convert PDF to text with Nanonets
 
Enhancing Worker Digital Experience: A Hands-on Workshop for Partners
Enhancing Worker Digital Experience: A Hands-on Workshop for PartnersEnhancing Worker Digital Experience: A Hands-on Workshop for Partners
Enhancing Worker Digital Experience: A Hands-on Workshop for Partners
 
Unblocking The Main Thread Solving ANRs and Frozen Frames
Unblocking The Main Thread Solving ANRs and Frozen FramesUnblocking The Main Thread Solving ANRs and Frozen Frames
Unblocking The Main Thread Solving ANRs and Frozen Frames
 
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...Integration and Automation in Practice: CI/CD in Mule Integration and Automat...
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...
 
Presentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreterPresentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreter
 

Chapter 4 better.pptx

  • 1. GRAPH NEURAL NETWORKS FOR NODE CLASSIFICATION
  • 2. AGENDA • Abstract • Background and Problem Definition • Supervised Graph Neural Networks • General Frame work of Graph Neural Networks (GNN) • Graph Convolutional Networks (GCN) • Graph Attention Networks • Neural Message Passing Networks(MPNN) • Continuous Graph Neural Networks • Multi-Scale Spectral Graph Convolutional Networks • Unsupervised Graph Neural Networks • Variational Graph Auto-Encoders • Deep Graph Infomax • Over-smoothing Problem
  • 3. ABSTRACT • In this chapter, we focus on a fundamental task on graphs : node classification.
  • 4. BACKGROUND AND PROBLEM DEFINITION • A fundamental task on graphs is node classification, which tries to classify the nodes into a few predefined categories. • To make effective prediction, a critical problem is to have very effective node representations, which largely determine the performance of node classification.
  • 5. BACKGROUND AND PROBLEM DEFINITION • N is the total number of nodes. • C is the number of features for each node. • F is the dimension of node representations. The goal of graph neural networks is to learn effective node representations (denoted as H ∈ R N×F) by combining the graph structure information and the node attributes, which are further used for node classification.
  • 6. Supervised Graph Neural Networks • General Frame work of Graph Neural Networks (GNN) • Graph Convolutional Networks (GCN) • Graph Attention Networks • Neural Message Passing Networks(MPNN) • Continuous Graph Neural Networks • Multi-Scale Spectral Graph Convolutional Networks
  • 7. 1- General Frame work of Graph Neural Networks (GNN) • The essential idea of graph neural networks is to iteratively update the node representations by combining the representations of their neighbors and their own representations.
  • 8. 1- General Frame work of Graph Neural Networks (GNN) Starting from the initial node representation H0 = X, in each layer we have two important functions: • AGGREGATE : which tries to aggregate the information from the neighbours of each node; • COMBINE : which tries to update the node representations by combining the aggregated information from neighbors with the current node representations.
  • 9.
  • 10. 2- Graph Convolutional Networks (GCN) • Which is now the most popular graph neural network architecture due to its simplicity and effectiveness in a variety of tasks and applications. • Specifically, the node representations in each layer is updated according to a propagation rule.
  • 11. 2- Graph Convolutional Networks (GCN) • AGGREGATE function is defined as the weighted average of the neighbor node representations. The weight of the neighbor j is determined by the weight of the edge between i and j (i.e. Ai j normalized by the degrees of the two nodes) • COMBINE function is defined as the summation of the aggregated messages and the node representation itself, in which the node representation is normalized by its own degree.
  • 12. 3- Graph Attention Networks • In GCNs, for a target node i, the importance of a neighbor j is determined by the weight of their edge Ai j (normalized by their node degrees). • However, in practice, the input graph may be noisy. • The edge weights may not be able to reflect the true strength between two nodes. • As a result, a more principled approach would be to automatically learn the importance of each neighbor. Graph Attention Networks (a.k.a. GAT) is built on this idea and try to learn the importance of each neighbor based on the Attention mechanism.
  • 13. 3- Graph Attention Networks • Graph Attention Layer: defnes how to transfer the hidden node representations at layer k −1 (denoted as H k−1) to the new node representations H. • In order to guarantee suffcient expressive power to transform the lower-level node representations to higher-level node representations, a shared linear transformation is applied to every node. • Afterwards, self-attention is defned on the nodes, which measures the attention coeffcients for any pair of nodes through a shared attentional mechanism (indicates the relationship strength between node i and j). • And to make the coeffcients comparable across different nodes, the attention coeffcients are usually normalized with the softmax function: • The new node representation is a linear combination of the neighboring node representations with the weights determined by the attention coeffcients (with a potential nonlinear transformation)
  • 14. 4- Neural Message Passing Networks(MPNN) • MPNN is actually very general, provides a general framework of graph neural networks, and could be used for the task of node classifcation as well. • The essential idea of MPNN is formulating existing graph neural networks as a general framework of neural message passing among nodes. • In MPNNs, there are two important functions including Message and Updating function:
  • 15. 4- Neural Message Passing Networks(MPNN) • The MPNN framework is very similar to the general framework • The AGGREGATE function defned here is simply a summation of all the messages from the neighbors. • The COMBINE function is the same as the node Updating function: which combines the aggregated messages from the neighbors and the node representation itself.
  • 16. 5- Continuous Graph Neural Networks • The above graph neural networks iteratively update the node representations with different kinds of graph convolutional layers. Essentially, these approaches model the discrete dynamics of node representations with GNNs. • Xhonneux et al (2020) proposed the continuous graph neural networks (CGNNs), which generalizes existing graph neural networks with discrete dynamics to continuous settings, i.e., trying to model the continuous dynamics of node representations. • The key idea is how to characterize the continuous dynamics of node representations, i.e. the derivatives of node representation w.r.t. time. • The derivatives of the node representations are defned as a combination of the node representation itself, the representations of its neighbors, and the initial status of the nodes.
  • 17. 5- Continuous Graph Neural Networks • Specifcally, two different variants of node dynamics are introduced. • The frst model assumes that different dimensions of node presentations (a.k.a. feature channels) are independent; • The second model is more fexible, which allows different feature channels to interact with each other.
  • 18. 6- Multi-Scale Spectral Graph Convolutional Networks
  • 19. UNSUPERVISED GRAPH NEURAL NETWORKS • Variational Graph Auto-Encoders • Deep graph infomax
  • 20. VARIATIONAL GRAPH AUTO-ENCODERS The VGAE model consists of an encoder, a decoder, and a prior . • Encoder The goal of the encoder is to learn a distribution of latent variables associated with each node conditioning on the node features X and the adjacency matrix A. (qφ (Z|X,A)). • Decoder Given sampled latent variables, the decoder aims at predicting the connectivities among nodes. Aopts a simple dot-product based predictor. • Prior The prior distributions over the latent variables are simply set to independent zero-mean Gaussians with unit variances, This prior is fixed throughout the learning.
  • 21. DEEP GRAPH INFOMAX • Deep Graph Infomax is an unsupervised learning framework that learns graph representations via the principle of mutual information maximization.
  • 22. OVER-SMOOTHING PROBLEM • Training deep graph neural networks by stacking multiple layers of graph neural networks usually yields inferior results, which is a common problem observed in many different graph neural network architectures. • This is mainly due to the problem of over-smoothing. • The graph convolutional network is a special case of Laplacian smoothing: • The Laplacian smoothing will push nodes belonging to the same clusters to take similar representations, which are benefcial for downstream tasks such as node classifcation. • However, when the GCNs go deep, the node representations suffer from the problem of over-smoothing, i.e., all the nodes will have similar representations. As a result, the performance on downstream tasks suffer as well.
  • 23. OVER-SMOOTHING PROBLEM • PairNorm: a method for alleviating the problem of over-smoothing when GNNs go deep. • The essential idea of PairNorm is to keep the total pairwise squared distance (TPSD) of node representations unchanged, which is the same as that of the original node feature X.