Presentation at the AIST'17 conference by Dmitry Ustalov. Authors of the original paper: Dmitry Ustalov, Mikhail Chernoskutov, Chris Biemann, Alexander Panchenko.
Monolingual Phrase Alignment on Parse Forests (EMNLP2017 presentation)Yuki Arase
Slides for my EMNLP2017 presentation (edited for slideshare).
Paper: http://aclweb.org/anthology/D/D17/D17-1001.pdf
Supplementary material: http://aclweb.org/anthology/attachments/D/D17/D17-1001.Attachment.zip
Abstract: We propose an efficient method to conduct phrase alignment on parse forests for paraphrase detection. Unlike previous studies, our method identifies syntactic paraphrases under linguistically motivated grammar. In addition, it allows phrases to non-compositionally align to handle paraphrases with non-homographic phrase correspondences.
A dataset that provides gold parse trees and their phrase alignments is created. The experimental results confirm that the proposed method conducts highly accurate phrase alignment compared to human performance.
Monolingual Phrase Alignment on Parse Forests (EMNLP2017 presentation)Yuki Arase
Slides for my EMNLP2017 presentation (edited for slideshare).
Paper: http://aclweb.org/anthology/D/D17/D17-1001.pdf
Supplementary material: http://aclweb.org/anthology/attachments/D/D17/D17-1001.Attachment.zip
Abstract: We propose an efficient method to conduct phrase alignment on parse forests for paraphrase detection. Unlike previous studies, our method identifies syntactic paraphrases under linguistically motivated grammar. In addition, it allows phrases to non-compositionally align to handle paraphrases with non-homographic phrase correspondences.
A dataset that provides gold parse trees and their phrase alignments is created. The experimental results confirm that the proposed method conducts highly accurate phrase alignment compared to human performance.
International Journal of Computational Engineering Research(IJCER) is an intentional online Journal in English monthly publishing journal. This Journal publish original research work that contributes significantly to further the scientific knowledge in engineering and Technology.
Basics of Language Science, Probabilities & Statistics and Artificial Neural ...Sanjib Narzary
Artificial Neural Network plays important role in Linguistics and Language Science as a whole. We often forget the basics of underlying principle behind these. This presentation provides many important characteristics of language, basics of mathematical concepts behind understanding neural network and concepts of Artificial Neural network which is the foundation of Deep Learning.
Deep Learning for Information Retrieval: Models, Progress, & OpportunitiesMatthew Lease
Talk given at the 8th Forum for Information Retrieval Evaluation (FIRE, http://fire.irsi.res.in/fire/2016/), December 10, 2016, and at the Qatar Computing Research Institute (QCRI), December 15, 2016.
Graph's not dead: from unsupervised induction of linguistic structures from t...Alexander Panchenko
In this invited talk, presented at the Dialogue'2018 conference, I argue for the usefulness of graph representations for NLP in the deep learning era. In the lecture, it is described how to extract symbolic linguistic structures, such as word senses and semantic frames in an unsupervised way from text corpora using graph-based algorithms and distributional semantics.
Building a Web-Scale Dependency-Parsed Corpus from Common CrawlAlexander Panchenko
We present DepCC, the largest-to-date linguistically analyzed corpus in English including 365 million documents, composed of 252 billion tokens and 7.5 billion of named entity occurrences in 14.3 billion sentences from a web-scale crawl of the Common Crawl project. The sentences are processed with a dependency parser and with a named entity tagger and contain provenance information, enabling various applications ranging from training syntax-based word embeddings to open information extraction and question answering. We built an index of all sentences and their linguistic meta-data enabling quick search across the corpus. We demonstrate the utility of this corpus on the verb similarity task by showing that a distributional model trained on our corpus yields better results than models trained on smaller corpora, like Wikipedia. This distributional model outperforms the state of art models of verb similarity trained on smaller corpora on the SimVerb3500 dataset.
http://www.lrec-conf.org/proceedings/lrec2018/summaries/215.html
More Related Content
Similar to Fighting with Sparsity of the Synonymy Dictionaries for Automatic Synset Induction
International Journal of Computational Engineering Research(IJCER) is an intentional online Journal in English monthly publishing journal. This Journal publish original research work that contributes significantly to further the scientific knowledge in engineering and Technology.
Basics of Language Science, Probabilities & Statistics and Artificial Neural ...Sanjib Narzary
Artificial Neural Network plays important role in Linguistics and Language Science as a whole. We often forget the basics of underlying principle behind these. This presentation provides many important characteristics of language, basics of mathematical concepts behind understanding neural network and concepts of Artificial Neural network which is the foundation of Deep Learning.
Deep Learning for Information Retrieval: Models, Progress, & OpportunitiesMatthew Lease
Talk given at the 8th Forum for Information Retrieval Evaluation (FIRE, http://fire.irsi.res.in/fire/2016/), December 10, 2016, and at the Qatar Computing Research Institute (QCRI), December 15, 2016.
Graph's not dead: from unsupervised induction of linguistic structures from t...Alexander Panchenko
In this invited talk, presented at the Dialogue'2018 conference, I argue for the usefulness of graph representations for NLP in the deep learning era. In the lecture, it is described how to extract symbolic linguistic structures, such as word senses and semantic frames in an unsupervised way from text corpora using graph-based algorithms and distributional semantics.
Building a Web-Scale Dependency-Parsed Corpus from Common CrawlAlexander Panchenko
We present DepCC, the largest-to-date linguistically analyzed corpus in English including 365 million documents, composed of 252 billion tokens and 7.5 billion of named entity occurrences in 14.3 billion sentences from a web-scale crawl of the Common Crawl project. The sentences are processed with a dependency parser and with a named entity tagger and contain provenance information, enabling various applications ranging from training syntax-based word embeddings to open information extraction and question answering. We built an index of all sentences and their linguistic meta-data enabling quick search across the corpus. We demonstrate the utility of this corpus on the verb similarity task by showing that a distributional model trained on our corpus yields better results than models trained on smaller corpora, like Wikipedia. This distributional model outperforms the state of art models of verb similarity trained on smaller corpora on the SimVerb3500 dataset.
http://www.lrec-conf.org/proceedings/lrec2018/summaries/215.html
Improving Hypernymy Extraction with Distributional Semantic ClassesAlexander Panchenko
http://www.lrec-conf.org/proceedings/lrec2018/pdf/234.pdf
In this paper, we show how distributionally-induced semantic classes can be helpful for extracting hypernyms. We present methods for inducing sense-aware semantic classes using distributional semantics and using these induced semantic classes for filtering noisy hypernymy relations. Denoising of hypernyms is performed by labeling each semantic class with its hypernyms. On the one hand, this allows us to filter out wrong extractions using the global structure of distributionally similar senses. On the other hand, we infer missing hypernyms via label propagation to cluster terms. We conduct a large-scale crowdsourcing study showing that processing of automatically extracted hypernyms using our approach improves the quality of the hypernymy extraction in terms of both precision and recall. Furthermore, we show the utility of our method in the domain taxonomy induction task, achieving the state-of-the-art results on a SemEval'16 task on taxonomy induction.
The paper was presented at the LREC'2018 conference in Miyazaki, Japan.
Inducing Interpretable Word Senses for WSD and Enrichment of Lexical ResourcesAlexander Panchenko
In this talk, we will discuss induction of sparse and dense
word sense representations using graph-based approaches and
distributional models. Induced senses are represented by a vector, but
also a set of hypernyms, images, and usage examples, derived in an
unsupervised and knowledge-free manner, which ensure interpretability
of the discovered senses by humans. We showcase the usage of the
induced representations for the tasks of word sense disambiguation and
enrichment of lexical resources, such as WordNet.
Using Linked Disambiguated Distributional Networks for Word Sense DisambiguationAlexander Panchenko
We introduce a new method for unsupervised knowledge-based word sense disambiguation (WSD) based on a resource that links two types of sense-aware lexical networks: one is induced from a corpus using distributional semantics, the other is manually constructed. The combination of two networks reduces the sparsity of sense representations used for WSD. We evaluate these enriched representations within two lexical sample sense disambiguation benchmarks. Our results indicate that (1) features extracted from the corpus-based resource help to significantly outperform a model based solely on the lexical resource; (2) our method achieves results comparable or better to four state-of-the-art unsupervised knowledge-based WSD systems including three hybrid systems that also rely on text corpora. In contrast to these hybrid methods, our approach does not require access to web search engines, texts mapped to a sense inventory, or machine translation systems.
See the full paper at: http://www.aclweb.org/anthology/W/W17/W17-1909.pdf
Panchenko A., Faralli S., Ponzetto S. P., and Biemann C. (2017): Using Linked Disambiguated Distributional Networks for Word Sense Disambiguation. In Proceedings of the Workshop on Sense, Concept and Entity Representations and their Applications (SENSE) co-located with the 15th Conference of the European Chapter of the Association for Computational Linguistics (EACL'2017). Valencia, Spain. Association for Computational Linguistics
Unsupervised Does Not Mean Uninterpretable: The Case for Word Sense Induction...Alexander Panchenko
The current trend in NLP is the use of highly opaque models, e.g. neural networks and word embeddings. While these models yield state-of-the-art results on a range of tasks, their drawback is
poor interpretability. On the example of word sense induction and disambiguation (WSID), we show that it is possible to develop an interpretable model that matches the state-of-the-art models in accuracy. Namely, we present an unsupervised, knowledge-free WSID approach, which is interpretable at three levels: word sense inventory, sense feature representations, and disambiguation procedure. Experiments show that our model performs on par with state-of-the-art word sense embeddings and other unsupervised systems while offering the possibility to justify
its decisions in human-readable form.
Noun Sense Induction and Disambiguation using Graph-Based Distributional Sema...Alexander Panchenko
We introduce an approach to word sense
induction and disambiguation. The method
is unsupervised and knowledge-free: sense
representations are learned from distributional
evidence and subsequently used to
disambiguate word instances in context.
These sense representations are obtained
by clustering dependency-based secondorder
similarity networks. We then add
features for disambiguation from heterogeneous
sources such as window-based and
sentence-wide co-occurrences, and explore
various schemes to combine these context
clues. Our method reaches a performance
comparable to the state-of-the-art unsupervised
word sense disambiguation systems
including top participants of the SemEval
2013 word sense induction task and two
more recent state-of-the-art neural word
sense induction systems
Full paper:
https://www.lt.informatik.tu-darmstadt.de/fileadmin/user_upload/Group_LangTech/publications/konvens2016panchenko.pdf
Ayush Kumar, Sarah Kohail, Amit Kumar, Asif Ekbal, Chris Biemann
IIT Patna, India
TU Darmstadt, Germany
Presented by: Alexander Panchenko, TU Darmstadt, Germany
A sentiment index measures the average emotional level in a corpus. We introduce four such indexes and use them to gauge average “positiveness” of a population during some period based on posts in a social network. This article for the first time presents a text-, rather than word-based sentiment index. Furthermore, this study presents the first large-scale study of the sentiment index of the Russian-speaking Facebook. Our results are consistent with the prior experiments for English language.
Semantic relations, such as synonyms, hypernyms and co-hyponyms proved to be useful for text processing applications, including text similarity, query expansion, question answering and word sense disambiguation. Such relations are practical because of the gap between lexical surface of the text and its meaning. Indeed, the same concept is often represented by different terms. However, existing resources often do not cover a vocabulary required by a given system. Manual resource construction is prohibitively expensive for many projects.
On the other hand, precision of the existing extractors still do not meet quality of the handcrafted resources. All these factors motivate the development of novel extraction methods. In this work we developed several similarity measures for semantic relation extraction. The main research question we address, is how to improve precision and coverage of such measures. First, we perform a large-scale study the baseline techniques. Second, we propose four novel measures. One of them significantly outperforms the baselines, the others perform comparably to the state-of-the-art techniques. Finally, we successfully apply one of the novel measures in two text processing systems.
Detecting Gender by Full Name: Experiments with the Russian LanguageAlexander Panchenko
This paper describes a method that detects gender of a person by his/her full name. While some approaches were proposed for English language, little has been done so far for Russian. We fill this gap and present a large-scale experiment on a dataset of 100,000 Russian full names from Facebook. Our method is based on three types of features (word endings, character $n$-grams and dictionary of names) combined within a linear supervised model. Experiments show that the proposed simple and computationally efficient approach yields excellent results achieving accuracy up to 96\%.
Вычислительная лексическая семантика: метрики семантической близости и их при...Alexander Panchenko
Вычислительная лексическая семантика: метрики семантической близости и их приложения
Серия лекций в НИУ ВШЭ, факультет бизнес-информатики и прикладной математики (Нижний Новгород)
Observation of Io’s Resurfacing via Plume Deposition Using Ground-based Adapt...Sérgio Sacani
Since volcanic activity was first discovered on Io from Voyager images in 1979, changes
on Io’s surface have been monitored from both spacecraft and ground-based telescopes.
Here, we present the highest spatial resolution images of Io ever obtained from a groundbased telescope. These images, acquired by the SHARK-VIS instrument on the Large
Binocular Telescope, show evidence of a major resurfacing event on Io’s trailing hemisphere. When compared to the most recent spacecraft images, the SHARK-VIS images
show that a plume deposit from a powerful eruption at Pillan Patera has covered part
of the long-lived Pele plume deposit. Although this type of resurfacing event may be common on Io, few have been detected due to the rarity of spacecraft visits and the previously low spatial resolution available from Earth-based telescopes. The SHARK-VIS instrument ushers in a new era of high resolution imaging of Io’s surface using adaptive
optics at visible wavelengths.
Professional air quality monitoring systems provide immediate, on-site data for analysis, compliance, and decision-making.
Monitor common gases, weather parameters, particulates.
Deep Behavioral Phenotyping in Systems Neuroscience for Functional Atlasing a...Ana Luísa Pinho
Functional Magnetic Resonance Imaging (fMRI) provides means to characterize brain activations in response to behavior. However, cognitive neuroscience has been limited to group-level effects referring to the performance of specific tasks. To obtain the functional profile of elementary cognitive mechanisms, the combination of brain responses to many tasks is required. Yet, to date, both structural atlases and parcellation-based activations do not fully account for cognitive function and still present several limitations. Further, they do not adapt overall to individual characteristics. In this talk, I will give an account of deep-behavioral phenotyping strategies, namely data-driven methods in large task-fMRI datasets, to optimize functional brain-data collection and improve inference of effects-of-interest related to mental processes. Key to this approach is the employment of fast multi-functional paradigms rich on features that can be well parametrized and, consequently, facilitate the creation of psycho-physiological constructs to be modelled with imaging data. Particular emphasis will be given to music stimuli when studying high-order cognitive mechanisms, due to their ecological nature and quality to enable complex behavior compounded by discrete entities. I will also discuss how deep-behavioral phenotyping and individualized models applied to neuroimaging data can better account for the subject-specific organization of domain-general cognitive systems in the human brain. Finally, the accumulation of functional brain signatures brings the possibility to clarify relationships among tasks and create a univocal link between brain systems and mental functions through: (1) the development of ontologies proposing an organization of cognitive processes; and (2) brain-network taxonomies describing functional specialization. To this end, tools to improve commensurability in cognitive science are necessary, such as public repositories, ontology-based platforms and automated meta-analysis tools. I will thus discuss some brain-atlasing resources currently under development, and their applicability in cognitive as well as clinical neuroscience.
This presentation explores a brief idea about the structural and functional attributes of nucleotides, the structure and function of genetic materials along with the impact of UV rays and pH upon them.
Remote Sensing and Computational, Evolutionary, Supercomputing, and Intellige...University of Maribor
Slides from talk:
Aleš Zamuda: Remote Sensing and Computational, Evolutionary, Supercomputing, and Intelligent Systems.
11th International Conference on Electrical, Electronics and Computer Engineering (IcETRAN), Niš, 3-6 June 2024
Inter-Society Networking Panel GRSS/MTT-S/CIS Panel Session: Promoting Connection and Cooperation
https://www.etran.rs/2024/en/home-english/
Phenomics assisted breeding in crop improvementIshaGoswami9
As the population is increasing and will reach about 9 billion upto 2050. Also due to climate change, it is difficult to meet the food requirement of such a large population. Facing the challenges presented by resource shortages, climate
change, and increasing global population, crop yield and quality need to be improved in a sustainable way over the coming decades. Genetic improvement by breeding is the best way to increase crop productivity. With the rapid progression of functional
genomics, an increasing number of crop genomes have been sequenced and dozens of genes influencing key agronomic traits have been identified. However, current genome sequence information has not been adequately exploited for understanding
the complex characteristics of multiple gene, owing to a lack of crop phenotypic data. Efficient, automatic, and accurate technologies and platforms that can capture phenotypic data that can
be linked to genomics information for crop improvement at all growth stages have become as important as genotyping. Thus,
high-throughput phenotyping has become the major bottleneck restricting crop breeding. Plant phenomics has been defined as the high-throughput, accurate acquisition and analysis of multi-dimensional phenotypes
during crop growing stages at the organism level, including the cell, tissue, organ, individual plant, plot, and field levels. With the rapid development of novel sensors, imaging technology,
and analysis methods, numerous infrastructure platforms have been developed for phenotyping.
DERIVATION OF MODIFIED BERNOULLI EQUATION WITH VISCOUS EFFECTS AND TERMINAL V...Wasswaderrick3
In this book, we use conservation of energy techniques on a fluid element to derive the Modified Bernoulli equation of flow with viscous or friction effects. We derive the general equation of flow/ velocity and then from this we derive the Pouiselle flow equation, the transition flow equation and the turbulent flow equation. In the situations where there are no viscous effects , the equation reduces to the Bernoulli equation. From experimental results, we are able to include other terms in the Bernoulli equation. We also look at cases where pressure gradients exist. We use the Modified Bernoulli equation to derive equations of flow rate for pipes of different cross sectional areas connected together. We also extend our techniques of energy conservation to a sphere falling in a viscous medium under the effect of gravity. We demonstrate Stokes equation of terminal velocity and turbulent flow equation. We look at a way of calculating the time taken for a body to fall in a viscous medium. We also look at the general equation of terminal velocity.
Seminar of U.V. Spectroscopy by SAMIR PANDASAMIR PANDA
Spectroscopy is a branch of science dealing the study of interaction of electromagnetic radiation with matter.
Ultraviolet-visible spectroscopy refers to absorption spectroscopy or reflect spectroscopy in the UV-VIS spectral region.
Ultraviolet-visible spectroscopy is an analytical method that can measure the amount of light received by the analyte.
Fighting with Sparsity of the Synonymy Dictionaries for Automatic Synset Induction
1. Fighting with Sparsity Ustalov D.A. et al.
Fighting with Sparsity of the
Synonymy Dictionaries for
Automatic Synset Induction
Dmitry Ustalov, Mikhail Chernoskutov
Ural Federal University
Chris Biemann, Alexander Panchenko
Universität Hamburg
2. Fighting with Sparsity Ustalov D.A. et al.
Outline
•Introduction
•The Problem
•The Approaches
•Evaluation
•Discussion
•Conclusion
2
3. Fighting with Sparsity Ustalov D.A. et al.
Introduction
•Synset Induction is an unsupervised
task of discovering synsets in a
synonymy graph.
•Notable Methods:
• MaxMax (Hope & Keller, 2013),
• ECO (Gonçalo-Oliveira & Gomes, 2014),
• WATSET (Ustalov et al., 2017) ← SOTA.
•See the survey in our paper.
3
4. Fighting with Sparsity Ustalov D.A. et al.
The Problem
•A synonymy graph contains densely
connected subgraphs.
•These subgraphs correspond to the
synsets.
•The synonymy dictionaries are not
perfect.
•Sometimes they have missing edges.
4
6. Fighting with Sparsity Ustalov D.A. et al.
The Approaches
•We propose two approaches for
reducing graph sparseness by adding
potentially pertinent edges.
• Synonymy Relation Transitivity (A1)
• Similar Synset Merging (A2)
•We also evaluate them on two lexical
semantic resources for Russian:
RuWordNet and YARN.
6
7. Fighting with Sparsity Ustalov D.A. et al.
•Synonymy is an equivalence relation:
• reflexiveness, symmetry, transitivity.
•We assume that if an edge is missing,
the graph still contains several
relatively short paths between the
synonymous words.
•This approach is designed to be
executed before the synset induction.
7
A1: Synonymy Transitivity
8. Fighting with Sparsity Ustalov D.A. et al.
A1: Synonymy Transitivity
•For each vertex, extract its 2nd order
ego network.
• Compute the set of candidate edges by
connecting the disconnected nodes.
• Compute the number of paths between
the nodes in candidate edges.
• Add an edge iff there exist at least k
paths of lengths [i; j].
•Then, the augmented graph is passed
to synset induction.
8
9. Fighting with Sparsity Ustalov D.A. et al.
A2: Synset Merging
•A similarity measure can be computed
between two vectors.
• Think of synset embeddings.
•We assume that if two synsets are
really similar, then they can be
merged.
•This approach is designed to be
executed after the synset induction.
9
10. Fighting with Sparsity Ustalov D.A. et al.
A2: Synset Merging
•Obtain synset embeddings using
SenseGram (Pelevina et al., 2016).
• Just average the word vectors in synsets.
•Identify the closely related synsets
using m-kNN algorithm (Panchenko et
al., 2012).
•Merge the t closely related synsets.
• The smallest are merged first.
10
11. Fighting with Sparsity Ustalov D.A. et al.
Evaluation
•We use WATSET, a
soft clustering
algorithm for
undirected
graphs.
•WATSET shows
SOTA results on
synset induction.
11
Ustalov D., Panchenko A., Biemann C. Watset: Automatic Induction of Synsets
from a Graph of Synonyms. In: Proc. ACL 2017.
12. Fighting with Sparsity Ustalov D.A. et al.
Evaluation: Measure & Data
•Measure: paired precision and recall.
•Gold standard: RuWordNet and YARN.
•The input graph: Wiktionary + Abramov
+ UNLDC.
•Word vectors are from RDT.
12
13. Fighting with Sparsity Ustalov D.A. et al.
RuWordNet YARN
13
Evaluation: Results
Input Graph Synonymy Transitivity Synset Merging
14. Fighting with Sparsity Ustalov D.A. et al.
Evaluation: Results
•Obviously, the transitivity approach
shown virtually no improvement.
•The merging approach substantially
increased the recall.
•Both methods trade off gains in recall
for the drops in precision.
14
15. Fighting with Sparsity Ustalov D.A. et al.
Discussion
•Transitivity. No word is a perfect
synonym of another. The communities
with the new edges become bigger.
•Merging. Distributional semantic
models tend to connect co-hyponyms
instead of synonyms.
•Alternatives. Structural Heuristics?
Hearst Patterns? Anaphora
Resolution? Crowdsourcing?
15
16. Fighting with Sparsity Ustalov D.A. et al.
Conclusion
•We fought with sparsity of the
synonymy dictionaries using two
approaches.
• Only synset merging won.
•Synset embeddings are easy to obtain.
They also show better results on such
a challenging task.
• Just average the word vectors and
compute similarity.
16
17. Fighting with Sparsity Ustalov D.A. et al.
Thank You!
• Dmitry Ustalov
dmitry.ustalov@gmail.com
• nlpub.ru/Watset
• nlpub.ru/RDT
Join SIGSLAV,
an ACL SIG on
Slavic languages!
sigslav.cs.helsinki.fi
We acknowledge the support of the Deutsche Forschungsgemeinschaft (DFG) foundation under
the “JOIN-T” project, the DAAD, the RFBR under the projects no. 16-37-00203 мол_а and no. 16-
37-00354 мол_а, and the RFH under the project no. 16-04-12019. The calculations were carried
out using the supercomputer “Uran” at the Krasovskii Institute of Mathematics and Mechanics.
We also thank four anonymous reviewers for their helpful comments.