SlideShare a Scribd company logo
Neural Relation Extraction Within and Across Sentence
Boundaries
Dependency Based Neural Architectures for Relation Extraction
Pankaj Gupta1,2
, Subburam Rajaram2
, Hinrich Sch¨utze1
& Thomas Runkler2
1
CIS, University of Munich (LMU), Germany
2
Corporate Technology, Machine-Intelligence, Siemens AG Munich, Germany
pankaj.gupta@campus.lmu.de | pankaj.gupta@siemens.com
Introduction
Precisely extract relationships in entities within and across sentence boundaries via
neural architectures based on dependency parse trees by modeling:
• shortest dependency path (SDP) using bidirectional RNN (biRNN)
• augmented dependency path (ADP) using Recursive NN (RecNN)
Binary Relation Extraction(RE)
- Identify semantic relationship between a pair of nominals or entities e1 and e2 in a text snippet, S
Problem Statement / Motivation
NOISY text in-between entities spanning sentence boundaries Ñ POOR PRECISION
Therefore, the need for a robust system that
• tackles false positives in inter-sentential RE Ñ good precision
• maintains a better balance in precision and recall Ñ improved F1 score
Evaluation and Analysis
• Quantitative evaluation on four datasets from medical (BioNLP ST 2011, 2013
and 2016) and news (MUC6) domain for intra- and inter-sentential relationships
train
Model
Evaluation for different values of sentence range k
param k  0 k ¤ 1 k ¤ 2 k ¤ 3
pr P R F1 pr P R F1 pr P R F1 pr P R F1
k  0
SVM 363 .474 .512 .492 821 .249 .606 .354 1212 .199 .678 .296 1517 .153 .684 .250
graphLSTM 473 .472 .668 .554 993 .213 .632 .319 1345 .166 .660 .266 2191 .121 .814 .218
i-biLSTM 480 .475 .674 .556 998 .220 .652 .328 1376 .165 .668 .265 1637 .132 .640 .219
i-biRNN 286 .517 .437 .474 425 .301 .378 .335 540 .249 .398 .307 570 .239 .401 .299
iDepNN-SDP 297 .519 .457 .486 553 .313 .510 .388 729 .240 .518 .328 832 .209 .516 .298
iDepNN-ADP 266 .526 .414 .467 476 .311 .438 .364 607 .251 .447 .320 669 .226 .447 .300
k ¤ 1
SVM 471 .464 .645 .540 888 .284 .746 .411 1109 .238 .779 .365 1196 .221 .779 .344
graphLSTM 406 .502 .607 .548 974 .226 .657 .336 1503 .165 .732 .268 2177 .126 .813 .218
i-biLSTM 417 .505 .628 .556 1101 .224 .730 .343 1690 .162 .818 .273 1969 .132 .772 .226
i-biRNN 376 .489 .544 .515 405 .393 .469 .427 406 .391 .469 .426 433 .369 .472 .414
iDepNN-SDP 303 .561 .503 .531 525 .358 .555 .435 660 .292 .569 .387 724 .265 .568 .362
iDepNN-ADP 292 .570 .491 .527 428 .402 .509 .449 497 .356 .522 .423 517 .341 .521 .412
k ¤ 2
SVM 495 .461 .675 .547 1016 .259 .780 .389 1296 .218 .834 .345 1418 .199 .834 .321
graphLSTM 442 .485 .637 .551 1016 .232 .702 .347 1334 .182 .723 .292 1758 .136 .717 .230
i-biLSTM 404 .487 .582 .531 940 .245 .682 .360 1205 .185 .661 .289 2146 .128 .816 .222
i-biRNN 288 .566 .482 .521 462 .376 .515 .435 556 .318 .524 .396 601 ..296 .525 .378
iDepNN-SDP 335 .537 .531 .534 633 .319 .598 .416 832 .258 .634 .367 941 .228 .633 .335
iDepNN-ADP 309 .538 .493 .514 485 .365 .525 .431 572 .320 .542 .402 603 .302 .540 .387
k ¤ 3
SVM 507 .458 .686 .549 1172 .234 .811 .363 1629 .186 .894 .308 1874 .162 .897 .275
graphLSTM 429 .491 .624 .550 1082 .230 .740 .351 1673 .167 .833 .280 2126 .124 .787 .214
i-biLSTM 417 .478 .582 .526 1142 .224 .758 .345 1218 .162 .833 .273 2091 .128 .800 .223
i-biRNN 405 .464 .559 .507 622 .324 .601 .422 654 .310 .604 .410 655 .311 .607 .410
iDepNN-SDP 351 .533 .552 .542 651 .315 .605 .414 842 .251 .622 .357 928 .227 .622 .333
iDepNN-ADP 313 .553 .512 .532 541 .355 .568 .437 654 .315 .601 .415 687 .300 .601 .401
k ¤ 1 ensemble 480 .478 .680 .561 837 .311 .769 .443 1003 .268 .794 .401 1074 .252 .797 .382
Table 1: BioNLP ST 2016 Dataset: Performance of the intra-and-inter-sentential training/evaluation
for different k. Underline: Better precision by iDepNN-ADP over iDepNN-SDP, graphLSTM[1] and
SVM. pr: Count of predictions. Ensemble of SVM, ibiRNN, iDepNN-SDP and iDepNN-ADP.
Ensemble with Threshold on Prediction Probability: (1) Exploit the precision and recall bias of
the different models via an ensemble approach, similar to TurkuNLP (Mehryary et al. 2016) and UMS
(Deleger et al. 2016) systems. (2) Aggregate the prediction outputs of the ibiRNN, iDepNN-SDP and
iDepNN-ADP and SVM classifiers, i.e., a relation to hold if any classifier has predicted it.
Neural Architectures for Intra-and Inter-sentential Relationships
A unified inter-sentential dependency-based neural network (iDepNN) models:
• iSDP with biRNN. The architecture is named as iDepNN-SDP.
• subtrees for each word on the iSDP with RecNN. The architecture (iDepNN-SDP
+ subtrees) is named as iDepNN-Augmented Dependency Path (iDepNN-ADP).
Ñ iDepNN-ADP offers precise structure and complementary information to iSDP in
classifying inter-sentential relationship
Each word is associated with a dependency relation r, e.g., r = dobj, during
the bottom-up construction of the subtree. For each r, a transformation matrix
Wr € Rd1
¢pd d1
q is learned. The subtree embedding is computed as:
cw  fp
°
q€Childrenpwq WRpw,qq
¤ pq   bq and pq  rxq, cqs
where Rpw,qq is the dependency relation between word w and its child word q and
b € Rd1
is a bias. This process continues recursively up to the root on the iSDP.
Same sentence One sentence apart
Three sentences apartTwo sentences apart
0
200
400
600
800
0 1 2 3 0 1 2 3
True Positive False Negative False Positive
SVM GraphLSTM
1000
count
Sentence Range for Evaluation Samples
172
164
191
205
133
616
230
109
981
1200
1400
1600
232
107
1285
0
200
400
600
800
1000
1200
1400
1600
218
120
252
252
86
636
264
75
845
264
75
932
Same sentence One sentence apart
Three sentences apartTwo sentences apart
223
117
250
211
128
782
223
117
456
265
75
1926
0 1 2 3 0 1 2 3
232
105
275
274
63
899
303
36
1326
303
35
1570
228
110
267
263
75
752
282
56
1014
282
56
1136
2000
220
120
754
248
92
1255
274
66
1903
235
103
781
243
97
1091
204
136
202
214
125
228
239
101
1518
279
61
1394
210
130
218
249
91
834
263
77
1863
Same sentence One sentence apart
Three sentences apartTwo sentences apart
134
198
126
147
190
328
151
187
456
151
187
518
162
176
227
166
172
284
166
172
297
177
160
308
182
155
389
155
182
112
166
171
143
182
155
421
204
136
550
172
166
160
191
147
409
204
136
623
0
200
400
600
800
1000
1200
1400
1600
0
200
400
600
800
1000
1200
1400
1600
iDepNN-ADP
0 1 2 3 0 1 2 3
Figure 1: Error Analysis: Count of True Positive, False Negative and False Positive. Observe the
fewer number of false positives in iDepNN-ADP, compared to both SVM and graphLSTM.
Conclusion  Key Takeaways
• Novel neural architectures iDepNN for Inter-sentential RE to precisely extract relations within and
across sentence boundaries and demonstrate a better balance in precision and recall
• Gain of 5.2% (0.587 vs 0.558) in F1 over the winning team (out of 11 teams) in BioNLP ST 2016
• Code/Data at: https://github.com/pgcool/Cross-sentence-Relation-Extraction-iDepNN
References
[1] Peng Nanyun, Hoifung Poon, Chris Quirk, Kristina Toutanova, and Wen-tau Yih. Cross-sentence n-ary relation ex-
traction with graph lstms. In Transactions of the Association for Computational Linguistics. 2017.

More Related Content

More from Pankaj Gupta, PhD

textTOvec: Deep Contextualized Neural Autoregressive Topic Models of Language...
textTOvec: Deep Contextualized Neural Autoregressive Topic Models of Language...textTOvec: Deep Contextualized Neural Autoregressive Topic Models of Language...
textTOvec: Deep Contextualized Neural Autoregressive Topic Models of Language...
Pankaj Gupta, PhD
 
Neural NLP Models of Information Extraction
Neural NLP Models of Information ExtractionNeural NLP Models of Information Extraction
Neural NLP Models of Information Extraction
Pankaj Gupta, PhD
 
Poster: Document Informed Neural Autoregressive Topic Models with Distributio...
Poster: Document Informed Neural Autoregressive Topic Models with Distributio...Poster: Document Informed Neural Autoregressive Topic Models with Distributio...
Poster: Document Informed Neural Autoregressive Topic Models with Distributio...
Pankaj Gupta, PhD
 
Document Informed Neural Autoregressive Topic Models with Distributional Prior
Document Informed Neural Autoregressive Topic Models with Distributional PriorDocument Informed Neural Autoregressive Topic Models with Distributional Prior
Document Informed Neural Autoregressive Topic Models with Distributional Prior
Pankaj Gupta, PhD
 
Neural Relation ExtractionWithin and Across Sentence Boundaries
Neural Relation ExtractionWithin and Across Sentence BoundariesNeural Relation ExtractionWithin and Across Sentence Boundaries
Neural Relation ExtractionWithin and Across Sentence Boundaries
Pankaj Gupta, PhD
 
Deep Learning for Information Extraction in Natural Language Text
Deep Learning for Information Extraction in Natural Language TextDeep Learning for Information Extraction in Natural Language Text
Deep Learning for Information Extraction in Natural Language Text
Pankaj Gupta, PhD
 
textTOvec: DEEP CONTEXTUALIZED NEURAL AUTOREGRESSIVE TOPIC MODELS OF LANGUAGE...
textTOvec: DEEP CONTEXTUALIZED NEURAL AUTOREGRESSIVE TOPIC MODELS OF LANGUAGE...textTOvec: DEEP CONTEXTUALIZED NEURAL AUTOREGRESSIVE TOPIC MODELS OF LANGUAGE...
textTOvec: DEEP CONTEXTUALIZED NEURAL AUTOREGRESSIVE TOPIC MODELS OF LANGUAGE...
Pankaj Gupta, PhD
 
Pankaj Gupta CV / Resume
Pankaj Gupta CV / ResumePankaj Gupta CV / Resume
Pankaj Gupta CV / Resume
Pankaj Gupta, PhD
 
LISA: Explaining RNN Judgments via Layer-wIse Semantic Accumulation and Examp...
LISA: Explaining RNN Judgments via Layer-wIse Semantic Accumulation and Examp...LISA: Explaining RNN Judgments via Layer-wIse Semantic Accumulation and Examp...
LISA: Explaining RNN Judgments via Layer-wIse Semantic Accumulation and Examp...
Pankaj Gupta, PhD
 
Replicated Siamese LSTM in Ticketing System for Similarity Learning and Retri...
Replicated Siamese LSTM in Ticketing System for Similarity Learning and Retri...Replicated Siamese LSTM in Ticketing System for Similarity Learning and Retri...
Replicated Siamese LSTM in Ticketing System for Similarity Learning and Retri...
Pankaj Gupta, PhD
 
Joint Bootstrapping Machines for High Confidence Relation Extraction
Joint Bootstrapping Machines for High Confidence Relation ExtractionJoint Bootstrapping Machines for High Confidence Relation Extraction
Joint Bootstrapping Machines for High Confidence Relation Extraction
Pankaj Gupta, PhD
 
RNN-RSM (Topics over Time) | NAACL2018 conference talk
RNN-RSM (Topics over Time) | NAACL2018 conference talkRNN-RSM (Topics over Time) | NAACL2018 conference talk
RNN-RSM (Topics over Time) | NAACL2018 conference talk
Pankaj Gupta, PhD
 
Lecture 07: Representation and Distributional Learning by Pankaj Gupta
Lecture 07: Representation and Distributional Learning by Pankaj GuptaLecture 07: Representation and Distributional Learning by Pankaj Gupta
Lecture 07: Representation and Distributional Learning by Pankaj Gupta
Pankaj Gupta, PhD
 
Lecture 05: Recurrent Neural Networks / Deep Learning by Pankaj Gupta
Lecture 05: Recurrent Neural Networks / Deep Learning by Pankaj GuptaLecture 05: Recurrent Neural Networks / Deep Learning by Pankaj Gupta
Lecture 05: Recurrent Neural Networks / Deep Learning by Pankaj Gupta
Pankaj Gupta, PhD
 

More from Pankaj Gupta, PhD (14)

textTOvec: Deep Contextualized Neural Autoregressive Topic Models of Language...
textTOvec: Deep Contextualized Neural Autoregressive Topic Models of Language...textTOvec: Deep Contextualized Neural Autoregressive Topic Models of Language...
textTOvec: Deep Contextualized Neural Autoregressive Topic Models of Language...
 
Neural NLP Models of Information Extraction
Neural NLP Models of Information ExtractionNeural NLP Models of Information Extraction
Neural NLP Models of Information Extraction
 
Poster: Document Informed Neural Autoregressive Topic Models with Distributio...
Poster: Document Informed Neural Autoregressive Topic Models with Distributio...Poster: Document Informed Neural Autoregressive Topic Models with Distributio...
Poster: Document Informed Neural Autoregressive Topic Models with Distributio...
 
Document Informed Neural Autoregressive Topic Models with Distributional Prior
Document Informed Neural Autoregressive Topic Models with Distributional PriorDocument Informed Neural Autoregressive Topic Models with Distributional Prior
Document Informed Neural Autoregressive Topic Models with Distributional Prior
 
Neural Relation ExtractionWithin and Across Sentence Boundaries
Neural Relation ExtractionWithin and Across Sentence BoundariesNeural Relation ExtractionWithin and Across Sentence Boundaries
Neural Relation ExtractionWithin and Across Sentence Boundaries
 
Deep Learning for Information Extraction in Natural Language Text
Deep Learning for Information Extraction in Natural Language TextDeep Learning for Information Extraction in Natural Language Text
Deep Learning for Information Extraction in Natural Language Text
 
textTOvec: DEEP CONTEXTUALIZED NEURAL AUTOREGRESSIVE TOPIC MODELS OF LANGUAGE...
textTOvec: DEEP CONTEXTUALIZED NEURAL AUTOREGRESSIVE TOPIC MODELS OF LANGUAGE...textTOvec: DEEP CONTEXTUALIZED NEURAL AUTOREGRESSIVE TOPIC MODELS OF LANGUAGE...
textTOvec: DEEP CONTEXTUALIZED NEURAL AUTOREGRESSIVE TOPIC MODELS OF LANGUAGE...
 
Pankaj Gupta CV / Resume
Pankaj Gupta CV / ResumePankaj Gupta CV / Resume
Pankaj Gupta CV / Resume
 
LISA: Explaining RNN Judgments via Layer-wIse Semantic Accumulation and Examp...
LISA: Explaining RNN Judgments via Layer-wIse Semantic Accumulation and Examp...LISA: Explaining RNN Judgments via Layer-wIse Semantic Accumulation and Examp...
LISA: Explaining RNN Judgments via Layer-wIse Semantic Accumulation and Examp...
 
Replicated Siamese LSTM in Ticketing System for Similarity Learning and Retri...
Replicated Siamese LSTM in Ticketing System for Similarity Learning and Retri...Replicated Siamese LSTM in Ticketing System for Similarity Learning and Retri...
Replicated Siamese LSTM in Ticketing System for Similarity Learning and Retri...
 
Joint Bootstrapping Machines for High Confidence Relation Extraction
Joint Bootstrapping Machines for High Confidence Relation ExtractionJoint Bootstrapping Machines for High Confidence Relation Extraction
Joint Bootstrapping Machines for High Confidence Relation Extraction
 
RNN-RSM (Topics over Time) | NAACL2018 conference talk
RNN-RSM (Topics over Time) | NAACL2018 conference talkRNN-RSM (Topics over Time) | NAACL2018 conference talk
RNN-RSM (Topics over Time) | NAACL2018 conference talk
 
Lecture 07: Representation and Distributional Learning by Pankaj Gupta
Lecture 07: Representation and Distributional Learning by Pankaj GuptaLecture 07: Representation and Distributional Learning by Pankaj Gupta
Lecture 07: Representation and Distributional Learning by Pankaj Gupta
 
Lecture 05: Recurrent Neural Networks / Deep Learning by Pankaj Gupta
Lecture 05: Recurrent Neural Networks / Deep Learning by Pankaj GuptaLecture 05: Recurrent Neural Networks / Deep Learning by Pankaj Gupta
Lecture 05: Recurrent Neural Networks / Deep Learning by Pankaj Gupta
 

Recently uploaded

Analysis insight about a Flyball dog competition team's performance
Analysis insight about a Flyball dog competition team's performanceAnalysis insight about a Flyball dog competition team's performance
Analysis insight about a Flyball dog competition team's performance
roli9797
 
My burning issue is homelessness K.C.M.O.
My burning issue is homelessness K.C.M.O.My burning issue is homelessness K.C.M.O.
My burning issue is homelessness K.C.M.O.
rwarrenll
 
06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Dat...
06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Dat...06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Dat...
06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Dat...
Timothy Spann
 
一比一原版(Dalhousie毕业证书)达尔豪斯大学毕业证如何办理
一比一原版(Dalhousie毕业证书)达尔豪斯大学毕业证如何办理一比一原版(Dalhousie毕业证书)达尔豪斯大学毕业证如何办理
一比一原版(Dalhousie毕业证书)达尔豪斯大学毕业证如何办理
mzpolocfi
 
做(mqu毕业证书)麦考瑞大学毕业证硕士文凭证书学费发票原版一模一样
做(mqu毕业证书)麦考瑞大学毕业证硕士文凭证书学费发票原版一模一样做(mqu毕业证书)麦考瑞大学毕业证硕士文凭证书学费发票原版一模一样
做(mqu毕业证书)麦考瑞大学毕业证硕士文凭证书学费发票原版一模一样
axoqas
 
06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Dat...
06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Dat...06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Dat...
06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Dat...
Timothy Spann
 
Data_and_Analytics_Essentials_Architect_an_Analytics_Platform.pptx
Data_and_Analytics_Essentials_Architect_an_Analytics_Platform.pptxData_and_Analytics_Essentials_Architect_an_Analytics_Platform.pptx
Data_and_Analytics_Essentials_Architect_an_Analytics_Platform.pptx
AnirbanRoy608946
 
Ch03-Managing the Object-Oriented Information Systems Project a.pdf
Ch03-Managing the Object-Oriented Information Systems Project a.pdfCh03-Managing the Object-Oriented Information Systems Project a.pdf
Ch03-Managing the Object-Oriented Information Systems Project a.pdf
haila53
 
一比一原版(爱大毕业证书)爱丁堡大学毕业证如何办理
一比一原版(爱大毕业证书)爱丁堡大学毕业证如何办理一比一原版(爱大毕业证书)爱丁堡大学毕业证如何办理
一比一原版(爱大毕业证书)爱丁堡大学毕业证如何办理
g4dpvqap0
 
一比一原版(UniSA毕业证书)南澳大学毕业证如何办理
一比一原版(UniSA毕业证书)南澳大学毕业证如何办理一比一原版(UniSA毕业证书)南澳大学毕业证如何办理
一比一原版(UniSA毕业证书)南澳大学毕业证如何办理
slg6lamcq
 
一比一原版(UIUC毕业证)伊利诺伊大学|厄巴纳-香槟分校毕业证如何办理
一比一原版(UIUC毕业证)伊利诺伊大学|厄巴纳-香槟分校毕业证如何办理一比一原版(UIUC毕业证)伊利诺伊大学|厄巴纳-香槟分校毕业证如何办理
一比一原版(UIUC毕业证)伊利诺伊大学|厄巴纳-香槟分校毕业证如何办理
ahzuo
 
Unleashing the Power of Data_ Choosing a Trusted Analytics Platform.pdf
Unleashing the Power of Data_ Choosing a Trusted Analytics Platform.pdfUnleashing the Power of Data_ Choosing a Trusted Analytics Platform.pdf
Unleashing the Power of Data_ Choosing a Trusted Analytics Platform.pdf
Enterprise Wired
 
Adjusting OpenMP PageRank : SHORT REPORT / NOTES
Adjusting OpenMP PageRank : SHORT REPORT / NOTESAdjusting OpenMP PageRank : SHORT REPORT / NOTES
Adjusting OpenMP PageRank : SHORT REPORT / NOTES
Subhajit Sahu
 
一比一原版(Bradford毕业证书)布拉德福德大学毕业证如何办理
一比一原版(Bradford毕业证书)布拉德福德大学毕业证如何办理一比一原版(Bradford毕业证书)布拉德福德大学毕业证如何办理
一比一原版(Bradford毕业证书)布拉德福德大学毕业证如何办理
mbawufebxi
 
Global Situational Awareness of A.I. and where its headed
Global Situational Awareness of A.I. and where its headedGlobal Situational Awareness of A.I. and where its headed
Global Situational Awareness of A.I. and where its headed
vikram sood
 
Nanandann Nilekani's ppt On India's .pdf
Nanandann Nilekani's ppt On India's .pdfNanandann Nilekani's ppt On India's .pdf
Nanandann Nilekani's ppt On India's .pdf
eddie19851
 
Algorithmic optimizations for Dynamic Levelwise PageRank (from STICD) : SHORT...
Algorithmic optimizations for Dynamic Levelwise PageRank (from STICD) : SHORT...Algorithmic optimizations for Dynamic Levelwise PageRank (from STICD) : SHORT...
Algorithmic optimizations for Dynamic Levelwise PageRank (from STICD) : SHORT...
Subhajit Sahu
 
一比一原版(BCU毕业证书)伯明翰城市大学毕业证如何办理
一比一原版(BCU毕业证书)伯明翰城市大学毕业证如何办理一比一原版(BCU毕业证书)伯明翰城市大学毕业证如何办理
一比一原版(BCU毕业证书)伯明翰城市大学毕业证如何办理
dwreak4tg
 
Enhanced Enterprise Intelligence with your personal AI Data Copilot.pdf
Enhanced Enterprise Intelligence with your personal AI Data Copilot.pdfEnhanced Enterprise Intelligence with your personal AI Data Copilot.pdf
Enhanced Enterprise Intelligence with your personal AI Data Copilot.pdf
GetInData
 
Malana- Gimlet Market Analysis (Portfolio 2)
Malana- Gimlet Market Analysis (Portfolio 2)Malana- Gimlet Market Analysis (Portfolio 2)
Malana- Gimlet Market Analysis (Portfolio 2)
TravisMalana
 

Recently uploaded (20)

Analysis insight about a Flyball dog competition team's performance
Analysis insight about a Flyball dog competition team's performanceAnalysis insight about a Flyball dog competition team's performance
Analysis insight about a Flyball dog competition team's performance
 
My burning issue is homelessness K.C.M.O.
My burning issue is homelessness K.C.M.O.My burning issue is homelessness K.C.M.O.
My burning issue is homelessness K.C.M.O.
 
06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Dat...
06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Dat...06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Dat...
06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Dat...
 
一比一原版(Dalhousie毕业证书)达尔豪斯大学毕业证如何办理
一比一原版(Dalhousie毕业证书)达尔豪斯大学毕业证如何办理一比一原版(Dalhousie毕业证书)达尔豪斯大学毕业证如何办理
一比一原版(Dalhousie毕业证书)达尔豪斯大学毕业证如何办理
 
做(mqu毕业证书)麦考瑞大学毕业证硕士文凭证书学费发票原版一模一样
做(mqu毕业证书)麦考瑞大学毕业证硕士文凭证书学费发票原版一模一样做(mqu毕业证书)麦考瑞大学毕业证硕士文凭证书学费发票原版一模一样
做(mqu毕业证书)麦考瑞大学毕业证硕士文凭证书学费发票原版一模一样
 
06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Dat...
06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Dat...06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Dat...
06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Dat...
 
Data_and_Analytics_Essentials_Architect_an_Analytics_Platform.pptx
Data_and_Analytics_Essentials_Architect_an_Analytics_Platform.pptxData_and_Analytics_Essentials_Architect_an_Analytics_Platform.pptx
Data_and_Analytics_Essentials_Architect_an_Analytics_Platform.pptx
 
Ch03-Managing the Object-Oriented Information Systems Project a.pdf
Ch03-Managing the Object-Oriented Information Systems Project a.pdfCh03-Managing the Object-Oriented Information Systems Project a.pdf
Ch03-Managing the Object-Oriented Information Systems Project a.pdf
 
一比一原版(爱大毕业证书)爱丁堡大学毕业证如何办理
一比一原版(爱大毕业证书)爱丁堡大学毕业证如何办理一比一原版(爱大毕业证书)爱丁堡大学毕业证如何办理
一比一原版(爱大毕业证书)爱丁堡大学毕业证如何办理
 
一比一原版(UniSA毕业证书)南澳大学毕业证如何办理
一比一原版(UniSA毕业证书)南澳大学毕业证如何办理一比一原版(UniSA毕业证书)南澳大学毕业证如何办理
一比一原版(UniSA毕业证书)南澳大学毕业证如何办理
 
一比一原版(UIUC毕业证)伊利诺伊大学|厄巴纳-香槟分校毕业证如何办理
一比一原版(UIUC毕业证)伊利诺伊大学|厄巴纳-香槟分校毕业证如何办理一比一原版(UIUC毕业证)伊利诺伊大学|厄巴纳-香槟分校毕业证如何办理
一比一原版(UIUC毕业证)伊利诺伊大学|厄巴纳-香槟分校毕业证如何办理
 
Unleashing the Power of Data_ Choosing a Trusted Analytics Platform.pdf
Unleashing the Power of Data_ Choosing a Trusted Analytics Platform.pdfUnleashing the Power of Data_ Choosing a Trusted Analytics Platform.pdf
Unleashing the Power of Data_ Choosing a Trusted Analytics Platform.pdf
 
Adjusting OpenMP PageRank : SHORT REPORT / NOTES
Adjusting OpenMP PageRank : SHORT REPORT / NOTESAdjusting OpenMP PageRank : SHORT REPORT / NOTES
Adjusting OpenMP PageRank : SHORT REPORT / NOTES
 
一比一原版(Bradford毕业证书)布拉德福德大学毕业证如何办理
一比一原版(Bradford毕业证书)布拉德福德大学毕业证如何办理一比一原版(Bradford毕业证书)布拉德福德大学毕业证如何办理
一比一原版(Bradford毕业证书)布拉德福德大学毕业证如何办理
 
Global Situational Awareness of A.I. and where its headed
Global Situational Awareness of A.I. and where its headedGlobal Situational Awareness of A.I. and where its headed
Global Situational Awareness of A.I. and where its headed
 
Nanandann Nilekani's ppt On India's .pdf
Nanandann Nilekani's ppt On India's .pdfNanandann Nilekani's ppt On India's .pdf
Nanandann Nilekani's ppt On India's .pdf
 
Algorithmic optimizations for Dynamic Levelwise PageRank (from STICD) : SHORT...
Algorithmic optimizations for Dynamic Levelwise PageRank (from STICD) : SHORT...Algorithmic optimizations for Dynamic Levelwise PageRank (from STICD) : SHORT...
Algorithmic optimizations for Dynamic Levelwise PageRank (from STICD) : SHORT...
 
一比一原版(BCU毕业证书)伯明翰城市大学毕业证如何办理
一比一原版(BCU毕业证书)伯明翰城市大学毕业证如何办理一比一原版(BCU毕业证书)伯明翰城市大学毕业证如何办理
一比一原版(BCU毕业证书)伯明翰城市大学毕业证如何办理
 
Enhanced Enterprise Intelligence with your personal AI Data Copilot.pdf
Enhanced Enterprise Intelligence with your personal AI Data Copilot.pdfEnhanced Enterprise Intelligence with your personal AI Data Copilot.pdf
Enhanced Enterprise Intelligence with your personal AI Data Copilot.pdf
 
Malana- Gimlet Market Analysis (Portfolio 2)
Malana- Gimlet Market Analysis (Portfolio 2)Malana- Gimlet Market Analysis (Portfolio 2)
Malana- Gimlet Market Analysis (Portfolio 2)
 

Poster: Neural Relation ExtractionWithin and Across Sentence Boundaries

  • 1. Neural Relation Extraction Within and Across Sentence Boundaries Dependency Based Neural Architectures for Relation Extraction Pankaj Gupta1,2 , Subburam Rajaram2 , Hinrich Sch¨utze1 & Thomas Runkler2 1 CIS, University of Munich (LMU), Germany 2 Corporate Technology, Machine-Intelligence, Siemens AG Munich, Germany pankaj.gupta@campus.lmu.de | pankaj.gupta@siemens.com Introduction Precisely extract relationships in entities within and across sentence boundaries via neural architectures based on dependency parse trees by modeling: • shortest dependency path (SDP) using bidirectional RNN (biRNN) • augmented dependency path (ADP) using Recursive NN (RecNN) Binary Relation Extraction(RE) - Identify semantic relationship between a pair of nominals or entities e1 and e2 in a text snippet, S Problem Statement / Motivation NOISY text in-between entities spanning sentence boundaries Ñ POOR PRECISION Therefore, the need for a robust system that • tackles false positives in inter-sentential RE Ñ good precision • maintains a better balance in precision and recall Ñ improved F1 score Evaluation and Analysis • Quantitative evaluation on four datasets from medical (BioNLP ST 2011, 2013 and 2016) and news (MUC6) domain for intra- and inter-sentential relationships train Model Evaluation for different values of sentence range k param k 0 k ¤ 1 k ¤ 2 k ¤ 3 pr P R F1 pr P R F1 pr P R F1 pr P R F1 k 0 SVM 363 .474 .512 .492 821 .249 .606 .354 1212 .199 .678 .296 1517 .153 .684 .250 graphLSTM 473 .472 .668 .554 993 .213 .632 .319 1345 .166 .660 .266 2191 .121 .814 .218 i-biLSTM 480 .475 .674 .556 998 .220 .652 .328 1376 .165 .668 .265 1637 .132 .640 .219 i-biRNN 286 .517 .437 .474 425 .301 .378 .335 540 .249 .398 .307 570 .239 .401 .299 iDepNN-SDP 297 .519 .457 .486 553 .313 .510 .388 729 .240 .518 .328 832 .209 .516 .298 iDepNN-ADP 266 .526 .414 .467 476 .311 .438 .364 607 .251 .447 .320 669 .226 .447 .300 k ¤ 1 SVM 471 .464 .645 .540 888 .284 .746 .411 1109 .238 .779 .365 1196 .221 .779 .344 graphLSTM 406 .502 .607 .548 974 .226 .657 .336 1503 .165 .732 .268 2177 .126 .813 .218 i-biLSTM 417 .505 .628 .556 1101 .224 .730 .343 1690 .162 .818 .273 1969 .132 .772 .226 i-biRNN 376 .489 .544 .515 405 .393 .469 .427 406 .391 .469 .426 433 .369 .472 .414 iDepNN-SDP 303 .561 .503 .531 525 .358 .555 .435 660 .292 .569 .387 724 .265 .568 .362 iDepNN-ADP 292 .570 .491 .527 428 .402 .509 .449 497 .356 .522 .423 517 .341 .521 .412 k ¤ 2 SVM 495 .461 .675 .547 1016 .259 .780 .389 1296 .218 .834 .345 1418 .199 .834 .321 graphLSTM 442 .485 .637 .551 1016 .232 .702 .347 1334 .182 .723 .292 1758 .136 .717 .230 i-biLSTM 404 .487 .582 .531 940 .245 .682 .360 1205 .185 .661 .289 2146 .128 .816 .222 i-biRNN 288 .566 .482 .521 462 .376 .515 .435 556 .318 .524 .396 601 ..296 .525 .378 iDepNN-SDP 335 .537 .531 .534 633 .319 .598 .416 832 .258 .634 .367 941 .228 .633 .335 iDepNN-ADP 309 .538 .493 .514 485 .365 .525 .431 572 .320 .542 .402 603 .302 .540 .387 k ¤ 3 SVM 507 .458 .686 .549 1172 .234 .811 .363 1629 .186 .894 .308 1874 .162 .897 .275 graphLSTM 429 .491 .624 .550 1082 .230 .740 .351 1673 .167 .833 .280 2126 .124 .787 .214 i-biLSTM 417 .478 .582 .526 1142 .224 .758 .345 1218 .162 .833 .273 2091 .128 .800 .223 i-biRNN 405 .464 .559 .507 622 .324 .601 .422 654 .310 .604 .410 655 .311 .607 .410 iDepNN-SDP 351 .533 .552 .542 651 .315 .605 .414 842 .251 .622 .357 928 .227 .622 .333 iDepNN-ADP 313 .553 .512 .532 541 .355 .568 .437 654 .315 .601 .415 687 .300 .601 .401 k ¤ 1 ensemble 480 .478 .680 .561 837 .311 .769 .443 1003 .268 .794 .401 1074 .252 .797 .382 Table 1: BioNLP ST 2016 Dataset: Performance of the intra-and-inter-sentential training/evaluation for different k. Underline: Better precision by iDepNN-ADP over iDepNN-SDP, graphLSTM[1] and SVM. pr: Count of predictions. Ensemble of SVM, ibiRNN, iDepNN-SDP and iDepNN-ADP. Ensemble with Threshold on Prediction Probability: (1) Exploit the precision and recall bias of the different models via an ensemble approach, similar to TurkuNLP (Mehryary et al. 2016) and UMS (Deleger et al. 2016) systems. (2) Aggregate the prediction outputs of the ibiRNN, iDepNN-SDP and iDepNN-ADP and SVM classifiers, i.e., a relation to hold if any classifier has predicted it. Neural Architectures for Intra-and Inter-sentential Relationships A unified inter-sentential dependency-based neural network (iDepNN) models: • iSDP with biRNN. The architecture is named as iDepNN-SDP. • subtrees for each word on the iSDP with RecNN. The architecture (iDepNN-SDP + subtrees) is named as iDepNN-Augmented Dependency Path (iDepNN-ADP). Ñ iDepNN-ADP offers precise structure and complementary information to iSDP in classifying inter-sentential relationship Each word is associated with a dependency relation r, e.g., r = dobj, during the bottom-up construction of the subtree. For each r, a transformation matrix Wr € Rd1 ¢pd d1 q is learned. The subtree embedding is computed as: cw fp ° q€Childrenpwq WRpw,qq ¤ pq   bq and pq rxq, cqs where Rpw,qq is the dependency relation between word w and its child word q and b € Rd1 is a bias. This process continues recursively up to the root on the iSDP. Same sentence One sentence apart Three sentences apartTwo sentences apart 0 200 400 600 800 0 1 2 3 0 1 2 3 True Positive False Negative False Positive SVM GraphLSTM 1000 count Sentence Range for Evaluation Samples 172 164 191 205 133 616 230 109 981 1200 1400 1600 232 107 1285 0 200 400 600 800 1000 1200 1400 1600 218 120 252 252 86 636 264 75 845 264 75 932 Same sentence One sentence apart Three sentences apartTwo sentences apart 223 117 250 211 128 782 223 117 456 265 75 1926 0 1 2 3 0 1 2 3 232 105 275 274 63 899 303 36 1326 303 35 1570 228 110 267 263 75 752 282 56 1014 282 56 1136 2000 220 120 754 248 92 1255 274 66 1903 235 103 781 243 97 1091 204 136 202 214 125 228 239 101 1518 279 61 1394 210 130 218 249 91 834 263 77 1863 Same sentence One sentence apart Three sentences apartTwo sentences apart 134 198 126 147 190 328 151 187 456 151 187 518 162 176 227 166 172 284 166 172 297 177 160 308 182 155 389 155 182 112 166 171 143 182 155 421 204 136 550 172 166 160 191 147 409 204 136 623 0 200 400 600 800 1000 1200 1400 1600 0 200 400 600 800 1000 1200 1400 1600 iDepNN-ADP 0 1 2 3 0 1 2 3 Figure 1: Error Analysis: Count of True Positive, False Negative and False Positive. Observe the fewer number of false positives in iDepNN-ADP, compared to both SVM and graphLSTM. Conclusion Key Takeaways • Novel neural architectures iDepNN for Inter-sentential RE to precisely extract relations within and across sentence boundaries and demonstrate a better balance in precision and recall • Gain of 5.2% (0.587 vs 0.558) in F1 over the winning team (out of 11 teams) in BioNLP ST 2016 • Code/Data at: https://github.com/pgcool/Cross-sentence-Relation-Extraction-iDepNN References [1] Peng Nanyun, Hoifung Poon, Chris Quirk, Kristina Toutanova, and Wen-tau Yih. Cross-sentence n-ary relation ex- traction with graph lstms. In Transactions of the Association for Computational Linguistics. 2017.