SlideShare a Scribd company logo
1 of 14
Ho-Beom Kim
Network Science Lab
Dept. of Mathematics
The Catholic University of Korea
E-mail: hobeom2001@catholic.ac.kr
2023 / 10 / 30
DWIVEDI, Vijay Prakash
AAAI 2021
2
Introduction
Problem Statements
• How to deal with the sparsity characteristics of graph?
• What positional encoding technique will be used to represent the position of the node in the graph?
3
Introduction
Contribution
1. They find that the most fruitful ideas from the transformers literature in NLP can be applied in a more
efficient way and posit that sparsity and positional encodings are two key aspects in the development
of a Graph Transformer
2. They put forward a generalization of transformer networks to homogeneous graphs of arbitrary
structure, namely Graph Transformer, and an extended version of Graph Transformer with edge
features that allows the usage of explicit domain information as edge features.
3. Their method includes an elegant way to fuse node positional features using Laplacian eigenvectors
for graph datasets, inspired from the heavy usage of positional encodings in NLP transformer models
and recent research on node positional features in GNNs.
4. Thier experiments demonstrate that the proposed model surpasses baseline isotropic and anisotropic
GNNs.
4
Methodology
Graph Sparsity
• Sparsity can be a great inductive bias in learning graphs
• When a graph is sparse, it means that there are not many edges between nodes.
• In the original transformer, all words become a 'fully connected graph of words' through the attention
process.
• This is because it is difficult to find meaningful interactions or connections between words in a
sentence.
• The words handled by the nlp transformer are smaller than tens or hundreds of nodes when expressed
graphically, so calculations are possible even when fully connected.
• In the case of the graph dataset, arbitrary connectivity structures (i.e. edges) can be used, and the node
size can reach millions or billions.
• It would be impossible or very inefficient to use the attention method used in existing transformers in
the graph.
5
Methodology
Overall Architecture
6
Methodology
Positional Encodings
• The transformer in NLP used positional encoding, which can convey information about where each word
token is located in the entire sequence and the distance between words.
• Until now, GNN has been trained to learn the structural information of nodes that are invariant in the node's
position in graph learning.
• Node attends to local node neighbors
• Use Laplacian eigenvectors for positional encoding
Symmetric normalized Laplacian matrix
Λ : eigenvalues
U : eigenvectors
7
Methodology
Graph Transformer Architecture - Input
node feature αi , edge feature βij
• Positional encoding is also embedded in d dimension through linear projection.
• And add it to the node feature. Positional encoding is added only to the node features of
the input layer.
8
Methodology
Graph Transformer Architecture – Graph Transformer Layer
The attention outputs
After passing through the Feed Forward Network, it goes through residual connections and
normalization layers. Normalization uses LayerNorm or BatchNorm.
9
Methodology
Graph Transformer Architecture – Graph Transformer Layer with edge features
Edge attention output
Node attention output
10
Methodology
Graph Transformer Architecture – Graph Transformer Layer with edge features
The outputs and are then passed to separate
Feed Forward Networks preceded and succeeded by
residual connections and normalization layers
11
Experiments
Datasets
• ZINC – Graph Regression
• ZINC is a molecular dataset with the task of graph property regression for constrained solubility
• PATTERN – Node Classification
• PATTERN is a node classification dataset generated using the Stochastic Block Models
• CLUSTER – Node Classification
• CLUSTER is also a synthetically generated dataset using SBM model.
12
Experiments
Results of GraphTransformer (GT) on all datasets
13
Experiments
Comparison of our best performing scores / Analysis of GraphTransformer using different PE schemes
14
Conclusion
Conclusion / Future works
• Their experiments consistently showed that the presence of Laplacian eigenvectors as
node positional encodings and batch normalization, in place of layer normalization,
around the transformer feed forward layers enhanced the transformer universally on
all experiments.
• In future works, they are interested in building upon the graph transformer along
aspects such as efficient training on single large graphs, applicability on
heterogeneous domains, etc., and perform efficient graph representation learning
keeping in account the recent innovations in graph inductive biases.

More Related Content

Similar to A Generalization of Transformer Networks to Graphs.pptx

240325_JW_labseminar[node2vec: Scalable Feature Learning for Networks].pptx
240325_JW_labseminar[node2vec: Scalable Feature Learning for Networks].pptx240325_JW_labseminar[node2vec: Scalable Feature Learning for Networks].pptx
240325_JW_labseminar[node2vec: Scalable Feature Learning for Networks].pptxthanhdowork
 
Deep learning for 3-D Scene Reconstruction and Modeling
Deep learning for 3-D Scene Reconstruction and Modeling Deep learning for 3-D Scene Reconstruction and Modeling
Deep learning for 3-D Scene Reconstruction and Modeling Yu Huang
 
NS-CUK Seminar: S.T.Nguyen, Review on "Do We Really Need Complicated Model Ar...
NS-CUK Seminar: S.T.Nguyen, Review on "Do We Really Need Complicated Model Ar...NS-CUK Seminar: S.T.Nguyen, Review on "Do We Really Need Complicated Model Ar...
NS-CUK Seminar: S.T.Nguyen, Review on "Do We Really Need Complicated Model Ar...ssuser4b1f48
 
Graph Attention Networks.pptx
Graph Attention Networks.pptxGraph Attention Networks.pptx
Graph Attention Networks.pptxssuser2624f71
 
240115_Attention Is All You Need (2017 NIPS).pptx
240115_Attention Is All You Need (2017 NIPS).pptx240115_Attention Is All You Need (2017 NIPS).pptx
240115_Attention Is All You Need (2017 NIPS).pptxthanhdowork
 
Graph convolutional neural networks for web-scale recommender systems.pptx
Graph convolutional neural networks for web-scale recommender systems.pptxGraph convolutional neural networks for web-scale recommender systems.pptx
Graph convolutional neural networks for web-scale recommender systems.pptxssuser2624f71
 
SPECFORMER: SPECTRAL GRAPH NEURAL NETWORKS MEET TRANSFORMERS.pptx
SPECFORMER: SPECTRAL GRAPH NEURAL NETWORKS MEET TRANSFORMERS.pptxSPECFORMER: SPECTRAL GRAPH NEURAL NETWORKS MEET TRANSFORMERS.pptx
SPECFORMER: SPECTRAL GRAPH NEURAL NETWORKS MEET TRANSFORMERS.pptxssuser2624f71
 
PyData Los Angeles 2020 (Abhilash Majumder)
PyData Los Angeles 2020 (Abhilash Majumder)PyData Los Angeles 2020 (Abhilash Majumder)
PyData Los Angeles 2020 (Abhilash Majumder)Abhilash Majumder
 
NS-CUK Joint Journal Club : S.T.Nguyen, Review on "Graph Neural Networks for ...
NS-CUK Joint Journal Club : S.T.Nguyen, Review on "Graph Neural Networks for ...NS-CUK Joint Journal Club : S.T.Nguyen, Review on "Graph Neural Networks for ...
NS-CUK Joint Journal Club : S.T.Nguyen, Review on "Graph Neural Networks for ...ssuser4b1f48
 
NS-CUK Seminar: S.T.Nguyen, Review on "Hierarchical Graph Convolutional Netwo...
NS-CUK Seminar: S.T.Nguyen, Review on "Hierarchical Graph Convolutional Netwo...NS-CUK Seminar: S.T.Nguyen, Review on "Hierarchical Graph Convolutional Netwo...
NS-CUK Seminar: S.T.Nguyen, Review on "Hierarchical Graph Convolutional Netwo...ssuser4b1f48
 
NS-CUK Seminar: J.H.Lee, Review on "Graph Propagation Transformer for Graph R...
NS-CUK Seminar: J.H.Lee, Review on "Graph Propagation Transformer for Graph R...NS-CUK Seminar: J.H.Lee, Review on "Graph Propagation Transformer for Graph R...
NS-CUK Seminar: J.H.Lee, Review on "Graph Propagation Transformer for Graph R...ssuser4b1f48
 
240226_Thanh_LabSeminar[Structure-Aware Transformer for Graph Representation ...
240226_Thanh_LabSeminar[Structure-Aware Transformer for Graph Representation ...240226_Thanh_LabSeminar[Structure-Aware Transformer for Graph Representation ...
240226_Thanh_LabSeminar[Structure-Aware Transformer for Graph Representation ...thanhdowork
 
240226_Thanh_LabSeminar[Structure-Aware Transformer for Graph Representation ...
240226_Thanh_LabSeminar[Structure-Aware Transformer for Graph Representation ...240226_Thanh_LabSeminar[Structure-Aware Transformer for Graph Representation ...
240226_Thanh_LabSeminar[Structure-Aware Transformer for Graph Representation ...thanhdowork
 
[20240304_LabSeminar_Huy]DeepWalk: Online Learning of Social Representations....
[20240304_LabSeminar_Huy]DeepWalk: Online Learning of Social Representations....[20240304_LabSeminar_Huy]DeepWalk: Online Learning of Social Representations....
[20240304_LabSeminar_Huy]DeepWalk: Online Learning of Social Representations....thanhdowork
 
NS-CUK Journal club: HBKim, Review on "Neural Graph Collaborative Filtering",...
NS-CUK Journal club: HBKim, Review on "Neural Graph Collaborative Filtering",...NS-CUK Journal club: HBKim, Review on "Neural Graph Collaborative Filtering",...
NS-CUK Journal club: HBKim, Review on "Neural Graph Collaborative Filtering",...ssuser4b1f48
 

Similar to A Generalization of Transformer Networks to Graphs.pptx (20)

240325_JW_labseminar[node2vec: Scalable Feature Learning for Networks].pptx
240325_JW_labseminar[node2vec: Scalable Feature Learning for Networks].pptx240325_JW_labseminar[node2vec: Scalable Feature Learning for Networks].pptx
240325_JW_labseminar[node2vec: Scalable Feature Learning for Networks].pptx
 
Deep learning for 3-D Scene Reconstruction and Modeling
Deep learning for 3-D Scene Reconstruction and Modeling Deep learning for 3-D Scene Reconstruction and Modeling
Deep learning for 3-D Scene Reconstruction and Modeling
 
NS-CUK Seminar: S.T.Nguyen, Review on "Do We Really Need Complicated Model Ar...
NS-CUK Seminar: S.T.Nguyen, Review on "Do We Really Need Complicated Model Ar...NS-CUK Seminar: S.T.Nguyen, Review on "Do We Really Need Complicated Model Ar...
NS-CUK Seminar: S.T.Nguyen, Review on "Do We Really Need Complicated Model Ar...
 
Graph Attention Networks.pptx
Graph Attention Networks.pptxGraph Attention Networks.pptx
Graph Attention Networks.pptx
 
240115_Attention Is All You Need (2017 NIPS).pptx
240115_Attention Is All You Need (2017 NIPS).pptx240115_Attention Is All You Need (2017 NIPS).pptx
240115_Attention Is All You Need (2017 NIPS).pptx
 
Graph convolutional neural networks for web-scale recommender systems.pptx
Graph convolutional neural networks for web-scale recommender systems.pptxGraph convolutional neural networks for web-scale recommender systems.pptx
Graph convolutional neural networks for web-scale recommender systems.pptx
 
Machine Learning @NECST
Machine Learning @NECSTMachine Learning @NECST
Machine Learning @NECST
 
SPECFORMER: SPECTRAL GRAPH NEURAL NETWORKS MEET TRANSFORMERS.pptx
SPECFORMER: SPECTRAL GRAPH NEURAL NETWORKS MEET TRANSFORMERS.pptxSPECFORMER: SPECTRAL GRAPH NEURAL NETWORKS MEET TRANSFORMERS.pptx
SPECFORMER: SPECTRAL GRAPH NEURAL NETWORKS MEET TRANSFORMERS.pptx
 
PyData Los Angeles 2020 (Abhilash Majumder)
PyData Los Angeles 2020 (Abhilash Majumder)PyData Los Angeles 2020 (Abhilash Majumder)
PyData Los Angeles 2020 (Abhilash Majumder)
 
NS-CUK Joint Journal Club : S.T.Nguyen, Review on "Graph Neural Networks for ...
NS-CUK Joint Journal Club : S.T.Nguyen, Review on "Graph Neural Networks for ...NS-CUK Joint Journal Club : S.T.Nguyen, Review on "Graph Neural Networks for ...
NS-CUK Joint Journal Club : S.T.Nguyen, Review on "Graph Neural Networks for ...
 
NS-CUK Seminar: S.T.Nguyen, Review on "Hierarchical Graph Convolutional Netwo...
NS-CUK Seminar: S.T.Nguyen, Review on "Hierarchical Graph Convolutional Netwo...NS-CUK Seminar: S.T.Nguyen, Review on "Hierarchical Graph Convolutional Netwo...
NS-CUK Seminar: S.T.Nguyen, Review on "Hierarchical Graph Convolutional Netwo...
 
Conformer review
Conformer reviewConformer review
Conformer review
 
NS-CUK Seminar: J.H.Lee, Review on "Graph Propagation Transformer for Graph R...
NS-CUK Seminar: J.H.Lee, Review on "Graph Propagation Transformer for Graph R...NS-CUK Seminar: J.H.Lee, Review on "Graph Propagation Transformer for Graph R...
NS-CUK Seminar: J.H.Lee, Review on "Graph Propagation Transformer for Graph R...
 
240226_Thanh_LabSeminar[Structure-Aware Transformer for Graph Representation ...
240226_Thanh_LabSeminar[Structure-Aware Transformer for Graph Representation ...240226_Thanh_LabSeminar[Structure-Aware Transformer for Graph Representation ...
240226_Thanh_LabSeminar[Structure-Aware Transformer for Graph Representation ...
 
240226_Thanh_LabSeminar[Structure-Aware Transformer for Graph Representation ...
240226_Thanh_LabSeminar[Structure-Aware Transformer for Graph Representation ...240226_Thanh_LabSeminar[Structure-Aware Transformer for Graph Representation ...
240226_Thanh_LabSeminar[Structure-Aware Transformer for Graph Representation ...
 
Data argumentation
Data argumentationData argumentation
Data argumentation
 
[20240304_LabSeminar_Huy]DeepWalk: Online Learning of Social Representations....
[20240304_LabSeminar_Huy]DeepWalk: Online Learning of Social Representations....[20240304_LabSeminar_Huy]DeepWalk: Online Learning of Social Representations....
[20240304_LabSeminar_Huy]DeepWalk: Online Learning of Social Representations....
 
DaViT.pdf
DaViT.pdfDaViT.pdf
DaViT.pdf
 
Chapter 3.pptx
Chapter 3.pptxChapter 3.pptx
Chapter 3.pptx
 
NS-CUK Journal club: HBKim, Review on "Neural Graph Collaborative Filtering",...
NS-CUK Journal club: HBKim, Review on "Neural Graph Collaborative Filtering",...NS-CUK Journal club: HBKim, Review on "Neural Graph Collaborative Filtering",...
NS-CUK Journal club: HBKim, Review on "Neural Graph Collaborative Filtering",...
 

More from ssuser2624f71

Vector and Matrix operationsVector and Matrix operations
Vector and Matrix operationsVector and Matrix operationsVector and Matrix operationsVector and Matrix operations
Vector and Matrix operationsVector and Matrix operationsssuser2624f71
 
240219_RNN, LSTM code.pptxdddddddddddddddd
240219_RNN, LSTM code.pptxdddddddddddddddd240219_RNN, LSTM code.pptxdddddddddddddddd
240219_RNN, LSTM code.pptxddddddddddddddddssuser2624f71
 
Sparse Graph Attention Networks 2021.pptx
Sparse Graph Attention Networks 2021.pptxSparse Graph Attention Networks 2021.pptx
Sparse Graph Attention Networks 2021.pptxssuser2624f71
 
인공지능 로봇 윤리_1229_9차시.pptx
인공지능 로봇 윤리_1229_9차시.pptx인공지능 로봇 윤리_1229_9차시.pptx
인공지능 로봇 윤리_1229_9차시.pptxssuser2624f71
 
인공지능 로봇 윤리_1228_8차시.pptx
인공지능 로봇 윤리_1228_8차시.pptx인공지능 로봇 윤리_1228_8차시.pptx
인공지능 로봇 윤리_1228_8차시.pptxssuser2624f71
 
인공지능 로봇 윤리_1227_7차시.pptx
인공지능 로봇 윤리_1227_7차시.pptx인공지능 로봇 윤리_1227_7차시.pptx
인공지능 로봇 윤리_1227_7차시.pptxssuser2624f71
 
인공지능 로봇 윤리_1226_6차시.pptx
인공지능 로봇 윤리_1226_6차시.pptx인공지능 로봇 윤리_1226_6차시.pptx
인공지능 로봇 윤리_1226_6차시.pptxssuser2624f71
 
인공지능 로봇 윤리_1222_5차시.pptx
인공지능 로봇 윤리_1222_5차시.pptx인공지능 로봇 윤리_1222_5차시.pptx
인공지능 로봇 윤리_1222_5차시.pptxssuser2624f71
 
인공지능 로봇 윤리_1221_4차시.pptx
인공지능 로봇 윤리_1221_4차시.pptx인공지능 로봇 윤리_1221_4차시.pptx
인공지능 로봇 윤리_1221_4차시.pptxssuser2624f71
 
인공지능 로봇 윤리_1220_3차시.pptx
인공지능 로봇 윤리_1220_3차시.pptx인공지능 로봇 윤리_1220_3차시.pptx
인공지능 로봇 윤리_1220_3차시.pptxssuser2624f71
 
인공지능 로봇 윤리_1219_2차시.pptx
인공지능 로봇 윤리_1219_2차시.pptx인공지능 로봇 윤리_1219_2차시.pptx
인공지능 로봇 윤리_1219_2차시.pptxssuser2624f71
 
인공지능 로봇 윤리_1218_1차시.pptx
인공지능 로봇 윤리_1218_1차시.pptx인공지능 로봇 윤리_1218_1차시.pptx
인공지능 로봇 윤리_1218_1차시.pptxssuser2624f71
 
디지털인문학9차시.pptx
디지털인문학9차시.pptx디지털인문학9차시.pptx
디지털인문학9차시.pptxssuser2624f71
 
디지털인문학8차시.pptx
디지털인문학8차시.pptx디지털인문학8차시.pptx
디지털인문학8차시.pptxssuser2624f71
 
디지털인문학7차시.pptx
디지털인문학7차시.pptx디지털인문학7차시.pptx
디지털인문학7차시.pptxssuser2624f71
 
디지털인문학6차시.pptx
디지털인문학6차시.pptx디지털인문학6차시.pptx
디지털인문학6차시.pptxssuser2624f71
 
디지털인문학 5차시.pptx
디지털인문학 5차시.pptx디지털인문학 5차시.pptx
디지털인문학 5차시.pptxssuser2624f71
 
디지털인문학4차시.pptx
디지털인문학4차시.pptx디지털인문학4차시.pptx
디지털인문학4차시.pptxssuser2624f71
 
디지털인문학3차시.pptx
디지털인문학3차시.pptx디지털인문학3차시.pptx
디지털인문학3차시.pptxssuser2624f71
 
디지털인문학2차시.pptx
디지털인문학2차시.pptx디지털인문학2차시.pptx
디지털인문학2차시.pptxssuser2624f71
 

More from ssuser2624f71 (20)

Vector and Matrix operationsVector and Matrix operations
Vector and Matrix operationsVector and Matrix operationsVector and Matrix operationsVector and Matrix operations
Vector and Matrix operationsVector and Matrix operations
 
240219_RNN, LSTM code.pptxdddddddddddddddd
240219_RNN, LSTM code.pptxdddddddddddddddd240219_RNN, LSTM code.pptxdddddddddddddddd
240219_RNN, LSTM code.pptxdddddddddddddddd
 
Sparse Graph Attention Networks 2021.pptx
Sparse Graph Attention Networks 2021.pptxSparse Graph Attention Networks 2021.pptx
Sparse Graph Attention Networks 2021.pptx
 
인공지능 로봇 윤리_1229_9차시.pptx
인공지능 로봇 윤리_1229_9차시.pptx인공지능 로봇 윤리_1229_9차시.pptx
인공지능 로봇 윤리_1229_9차시.pptx
 
인공지능 로봇 윤리_1228_8차시.pptx
인공지능 로봇 윤리_1228_8차시.pptx인공지능 로봇 윤리_1228_8차시.pptx
인공지능 로봇 윤리_1228_8차시.pptx
 
인공지능 로봇 윤리_1227_7차시.pptx
인공지능 로봇 윤리_1227_7차시.pptx인공지능 로봇 윤리_1227_7차시.pptx
인공지능 로봇 윤리_1227_7차시.pptx
 
인공지능 로봇 윤리_1226_6차시.pptx
인공지능 로봇 윤리_1226_6차시.pptx인공지능 로봇 윤리_1226_6차시.pptx
인공지능 로봇 윤리_1226_6차시.pptx
 
인공지능 로봇 윤리_1222_5차시.pptx
인공지능 로봇 윤리_1222_5차시.pptx인공지능 로봇 윤리_1222_5차시.pptx
인공지능 로봇 윤리_1222_5차시.pptx
 
인공지능 로봇 윤리_1221_4차시.pptx
인공지능 로봇 윤리_1221_4차시.pptx인공지능 로봇 윤리_1221_4차시.pptx
인공지능 로봇 윤리_1221_4차시.pptx
 
인공지능 로봇 윤리_1220_3차시.pptx
인공지능 로봇 윤리_1220_3차시.pptx인공지능 로봇 윤리_1220_3차시.pptx
인공지능 로봇 윤리_1220_3차시.pptx
 
인공지능 로봇 윤리_1219_2차시.pptx
인공지능 로봇 윤리_1219_2차시.pptx인공지능 로봇 윤리_1219_2차시.pptx
인공지능 로봇 윤리_1219_2차시.pptx
 
인공지능 로봇 윤리_1218_1차시.pptx
인공지능 로봇 윤리_1218_1차시.pptx인공지능 로봇 윤리_1218_1차시.pptx
인공지능 로봇 윤리_1218_1차시.pptx
 
디지털인문학9차시.pptx
디지털인문학9차시.pptx디지털인문학9차시.pptx
디지털인문학9차시.pptx
 
디지털인문학8차시.pptx
디지털인문학8차시.pptx디지털인문학8차시.pptx
디지털인문학8차시.pptx
 
디지털인문학7차시.pptx
디지털인문학7차시.pptx디지털인문학7차시.pptx
디지털인문학7차시.pptx
 
디지털인문학6차시.pptx
디지털인문학6차시.pptx디지털인문학6차시.pptx
디지털인문학6차시.pptx
 
디지털인문학 5차시.pptx
디지털인문학 5차시.pptx디지털인문학 5차시.pptx
디지털인문학 5차시.pptx
 
디지털인문학4차시.pptx
디지털인문학4차시.pptx디지털인문학4차시.pptx
디지털인문학4차시.pptx
 
디지털인문학3차시.pptx
디지털인문학3차시.pptx디지털인문학3차시.pptx
디지털인문학3차시.pptx
 
디지털인문학2차시.pptx
디지털인문학2차시.pptx디지털인문학2차시.pptx
디지털인문학2차시.pptx
 

Recently uploaded

Roles & Responsibilities in Pharmacovigilance
Roles & Responsibilities in PharmacovigilanceRoles & Responsibilities in Pharmacovigilance
Roles & Responsibilities in PharmacovigilanceSamikshaHamane
 
How to Make a Pirate ship Primary Education.pptx
How to Make a Pirate ship Primary Education.pptxHow to Make a Pirate ship Primary Education.pptx
How to Make a Pirate ship Primary Education.pptxmanuelaromero2013
 
Organic Name Reactions for the students and aspirants of Chemistry12th.pptx
Organic Name Reactions  for the students and aspirants of Chemistry12th.pptxOrganic Name Reactions  for the students and aspirants of Chemistry12th.pptx
Organic Name Reactions for the students and aspirants of Chemistry12th.pptxVS Mahajan Coaching Centre
 
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptx
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptxPOINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptx
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptxSayali Powar
 
Full Stack Web Development Course for Beginners
Full Stack Web Development Course  for BeginnersFull Stack Web Development Course  for Beginners
Full Stack Web Development Course for BeginnersSabitha Banu
 
call girls in Kamla Market (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️
call girls in Kamla Market (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️call girls in Kamla Market (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️
call girls in Kamla Market (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️9953056974 Low Rate Call Girls In Saket, Delhi NCR
 
“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...
“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...
“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...Marc Dusseiller Dusjagr
 
Proudly South Africa powerpoint Thorisha.pptx
Proudly South Africa powerpoint Thorisha.pptxProudly South Africa powerpoint Thorisha.pptx
Proudly South Africa powerpoint Thorisha.pptxthorishapillay1
 
What is Model Inheritance in Odoo 17 ERP
What is Model Inheritance in Odoo 17 ERPWhat is Model Inheritance in Odoo 17 ERP
What is Model Inheritance in Odoo 17 ERPCeline George
 
Employee wellbeing at the workplace.pptx
Employee wellbeing at the workplace.pptxEmployee wellbeing at the workplace.pptx
Employee wellbeing at the workplace.pptxNirmalaLoungPoorunde1
 
Blooming Together_ Growing a Community Garden Worksheet.docx
Blooming Together_ Growing a Community Garden Worksheet.docxBlooming Together_ Growing a Community Garden Worksheet.docx
Blooming Together_ Growing a Community Garden Worksheet.docxUnboundStockton
 
Solving Puzzles Benefits Everyone (English).pptx
Solving Puzzles Benefits Everyone (English).pptxSolving Puzzles Benefits Everyone (English).pptx
Solving Puzzles Benefits Everyone (English).pptxOH TEIK BIN
 
DATA STRUCTURE AND ALGORITHM for beginners
DATA STRUCTURE AND ALGORITHM for beginnersDATA STRUCTURE AND ALGORITHM for beginners
DATA STRUCTURE AND ALGORITHM for beginnersSabitha Banu
 
Presiding Officer Training module 2024 lok sabha elections
Presiding Officer Training module 2024 lok sabha electionsPresiding Officer Training module 2024 lok sabha elections
Presiding Officer Training module 2024 lok sabha electionsanshu789521
 
ECONOMIC CONTEXT - LONG FORM TV DRAMA - PPT
ECONOMIC CONTEXT - LONG FORM TV DRAMA - PPTECONOMIC CONTEXT - LONG FORM TV DRAMA - PPT
ECONOMIC CONTEXT - LONG FORM TV DRAMA - PPTiammrhaywood
 
MARGINALIZATION (Different learners in Marginalized Group
MARGINALIZATION (Different learners in Marginalized GroupMARGINALIZATION (Different learners in Marginalized Group
MARGINALIZATION (Different learners in Marginalized GroupJonathanParaisoCruz
 
Types of Journalistic Writing Grade 8.pptx
Types of Journalistic Writing Grade 8.pptxTypes of Journalistic Writing Grade 8.pptx
Types of Journalistic Writing Grade 8.pptxEyham Joco
 
EPANDING THE CONTENT OF AN OUTLINE using notes.pptx
EPANDING THE CONTENT OF AN OUTLINE using notes.pptxEPANDING THE CONTENT OF AN OUTLINE using notes.pptx
EPANDING THE CONTENT OF AN OUTLINE using notes.pptxRaymartEstabillo3
 
Capitol Tech U Doctoral Presentation - April 2024.pptx
Capitol Tech U Doctoral Presentation - April 2024.pptxCapitol Tech U Doctoral Presentation - April 2024.pptx
Capitol Tech U Doctoral Presentation - April 2024.pptxCapitolTechU
 

Recently uploaded (20)

Roles & Responsibilities in Pharmacovigilance
Roles & Responsibilities in PharmacovigilanceRoles & Responsibilities in Pharmacovigilance
Roles & Responsibilities in Pharmacovigilance
 
How to Make a Pirate ship Primary Education.pptx
How to Make a Pirate ship Primary Education.pptxHow to Make a Pirate ship Primary Education.pptx
How to Make a Pirate ship Primary Education.pptx
 
Organic Name Reactions for the students and aspirants of Chemistry12th.pptx
Organic Name Reactions  for the students and aspirants of Chemistry12th.pptxOrganic Name Reactions  for the students and aspirants of Chemistry12th.pptx
Organic Name Reactions for the students and aspirants of Chemistry12th.pptx
 
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptx
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptxPOINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptx
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptx
 
Full Stack Web Development Course for Beginners
Full Stack Web Development Course  for BeginnersFull Stack Web Development Course  for Beginners
Full Stack Web Development Course for Beginners
 
call girls in Kamla Market (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️
call girls in Kamla Market (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️call girls in Kamla Market (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️
call girls in Kamla Market (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️
 
“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...
“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...
“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...
 
Proudly South Africa powerpoint Thorisha.pptx
Proudly South Africa powerpoint Thorisha.pptxProudly South Africa powerpoint Thorisha.pptx
Proudly South Africa powerpoint Thorisha.pptx
 
What is Model Inheritance in Odoo 17 ERP
What is Model Inheritance in Odoo 17 ERPWhat is Model Inheritance in Odoo 17 ERP
What is Model Inheritance in Odoo 17 ERP
 
Employee wellbeing at the workplace.pptx
Employee wellbeing at the workplace.pptxEmployee wellbeing at the workplace.pptx
Employee wellbeing at the workplace.pptx
 
Blooming Together_ Growing a Community Garden Worksheet.docx
Blooming Together_ Growing a Community Garden Worksheet.docxBlooming Together_ Growing a Community Garden Worksheet.docx
Blooming Together_ Growing a Community Garden Worksheet.docx
 
Solving Puzzles Benefits Everyone (English).pptx
Solving Puzzles Benefits Everyone (English).pptxSolving Puzzles Benefits Everyone (English).pptx
Solving Puzzles Benefits Everyone (English).pptx
 
DATA STRUCTURE AND ALGORITHM for beginners
DATA STRUCTURE AND ALGORITHM for beginnersDATA STRUCTURE AND ALGORITHM for beginners
DATA STRUCTURE AND ALGORITHM for beginners
 
Presiding Officer Training module 2024 lok sabha elections
Presiding Officer Training module 2024 lok sabha electionsPresiding Officer Training module 2024 lok sabha elections
Presiding Officer Training module 2024 lok sabha elections
 
ECONOMIC CONTEXT - LONG FORM TV DRAMA - PPT
ECONOMIC CONTEXT - LONG FORM TV DRAMA - PPTECONOMIC CONTEXT - LONG FORM TV DRAMA - PPT
ECONOMIC CONTEXT - LONG FORM TV DRAMA - PPT
 
MARGINALIZATION (Different learners in Marginalized Group
MARGINALIZATION (Different learners in Marginalized GroupMARGINALIZATION (Different learners in Marginalized Group
MARGINALIZATION (Different learners in Marginalized Group
 
Types of Journalistic Writing Grade 8.pptx
Types of Journalistic Writing Grade 8.pptxTypes of Journalistic Writing Grade 8.pptx
Types of Journalistic Writing Grade 8.pptx
 
EPANDING THE CONTENT OF AN OUTLINE using notes.pptx
EPANDING THE CONTENT OF AN OUTLINE using notes.pptxEPANDING THE CONTENT OF AN OUTLINE using notes.pptx
EPANDING THE CONTENT OF AN OUTLINE using notes.pptx
 
Model Call Girl in Tilak Nagar Delhi reach out to us at 🔝9953056974🔝
Model Call Girl in Tilak Nagar Delhi reach out to us at 🔝9953056974🔝Model Call Girl in Tilak Nagar Delhi reach out to us at 🔝9953056974🔝
Model Call Girl in Tilak Nagar Delhi reach out to us at 🔝9953056974🔝
 
Capitol Tech U Doctoral Presentation - April 2024.pptx
Capitol Tech U Doctoral Presentation - April 2024.pptxCapitol Tech U Doctoral Presentation - April 2024.pptx
Capitol Tech U Doctoral Presentation - April 2024.pptx
 

A Generalization of Transformer Networks to Graphs.pptx

  • 1. Ho-Beom Kim Network Science Lab Dept. of Mathematics The Catholic University of Korea E-mail: hobeom2001@catholic.ac.kr 2023 / 10 / 30 DWIVEDI, Vijay Prakash AAAI 2021
  • 2. 2 Introduction Problem Statements • How to deal with the sparsity characteristics of graph? • What positional encoding technique will be used to represent the position of the node in the graph?
  • 3. 3 Introduction Contribution 1. They find that the most fruitful ideas from the transformers literature in NLP can be applied in a more efficient way and posit that sparsity and positional encodings are two key aspects in the development of a Graph Transformer 2. They put forward a generalization of transformer networks to homogeneous graphs of arbitrary structure, namely Graph Transformer, and an extended version of Graph Transformer with edge features that allows the usage of explicit domain information as edge features. 3. Their method includes an elegant way to fuse node positional features using Laplacian eigenvectors for graph datasets, inspired from the heavy usage of positional encodings in NLP transformer models and recent research on node positional features in GNNs. 4. Thier experiments demonstrate that the proposed model surpasses baseline isotropic and anisotropic GNNs.
  • 4. 4 Methodology Graph Sparsity • Sparsity can be a great inductive bias in learning graphs • When a graph is sparse, it means that there are not many edges between nodes. • In the original transformer, all words become a 'fully connected graph of words' through the attention process. • This is because it is difficult to find meaningful interactions or connections between words in a sentence. • The words handled by the nlp transformer are smaller than tens or hundreds of nodes when expressed graphically, so calculations are possible even when fully connected. • In the case of the graph dataset, arbitrary connectivity structures (i.e. edges) can be used, and the node size can reach millions or billions. • It would be impossible or very inefficient to use the attention method used in existing transformers in the graph.
  • 6. 6 Methodology Positional Encodings • The transformer in NLP used positional encoding, which can convey information about where each word token is located in the entire sequence and the distance between words. • Until now, GNN has been trained to learn the structural information of nodes that are invariant in the node's position in graph learning. • Node attends to local node neighbors • Use Laplacian eigenvectors for positional encoding Symmetric normalized Laplacian matrix Λ : eigenvalues U : eigenvectors
  • 7. 7 Methodology Graph Transformer Architecture - Input node feature αi , edge feature βij • Positional encoding is also embedded in d dimension through linear projection. • And add it to the node feature. Positional encoding is added only to the node features of the input layer.
  • 8. 8 Methodology Graph Transformer Architecture – Graph Transformer Layer The attention outputs After passing through the Feed Forward Network, it goes through residual connections and normalization layers. Normalization uses LayerNorm or BatchNorm.
  • 9. 9 Methodology Graph Transformer Architecture – Graph Transformer Layer with edge features Edge attention output Node attention output
  • 10. 10 Methodology Graph Transformer Architecture – Graph Transformer Layer with edge features The outputs and are then passed to separate Feed Forward Networks preceded and succeeded by residual connections and normalization layers
  • 11. 11 Experiments Datasets • ZINC – Graph Regression • ZINC is a molecular dataset with the task of graph property regression for constrained solubility • PATTERN – Node Classification • PATTERN is a node classification dataset generated using the Stochastic Block Models • CLUSTER – Node Classification • CLUSTER is also a synthetically generated dataset using SBM model.
  • 13. 13 Experiments Comparison of our best performing scores / Analysis of GraphTransformer using different PE schemes
  • 14. 14 Conclusion Conclusion / Future works • Their experiments consistently showed that the presence of Laplacian eigenvectors as node positional encodings and batch normalization, in place of layer normalization, around the transformer feed forward layers enhanced the transformer universally on all experiments. • In future works, they are interested in building upon the graph transformer along aspects such as efficient training on single large graphs, applicability on heterogeneous domains, etc., and perform efficient graph representation learning keeping in account the recent innovations in graph inductive biases.