This document discusses methods for creating word embeddings from text using distributional semantics and dependency structures. It presents experiments comparing word embeddings generated from various dependency structures to traditional Word2Vec models. The embeddings are evaluated using WordNet and VerbNet lexical databases, as well as in a sentiment analysis task. The results show that embeddings incorporating dependency structures outperform topical models on these evaluations, and that certain verb classes and thematic roles yield better embeddings than others. The document concludes that structure-based embeddings capture semantic information better than topical models.
3. Word Embeddings
• From words to dense vectors
• Capturing semantics in a quantifiable vector
• Uses of these embeddings
• Current embedding methods
• Changing the game of natural language processing
8. Dependency Structure Word
Embeddings
• Levy & Goldberg, 2014.
• Used only the head word and word and dependents
• Evaluated on differences of similarity (function) and
relatedness (topical)
• Evaluated on a dataset called WordSim, which is
mostly nouns
• Did not look at if the embeddings created a better
overall model nor did they try varying structures
11. Experiments
• DEP1: the first order dependencies of W
• DEP1H: the first order dependencies of W and the dependency head of W
• DEP12H: the first and second order dependencies of W and the dependency head of W
• DEP1SIB1: the first order dependencies of W, the rightmost sibling of W and the
leftmost sibling of W
• DEP1ALLSIB: the first order dependencies of W, all siblings of W
• DEP1SRLH: the first order dependencies of W, the semantic head of W
• DEP1SRLARG: the first order dependencies of W, all the semantic arguments of W
21. VerbNet
• Finding the best and worst verb classes
• Finding patterns in the best and worst verb classes
• Thematic role labels
• Semantic restrictions
• Note*: for each verb class VerbNet labels both of the above
• Plotting patterns
Ex. of verb class Adopt-93
Members assume, adopt, take
Thematic Roles agent, theme
Semantic Restriction animate, organization
22. Top and Bottom Verb Classes
by Average Rank Correlation
• Sorted all verb classes by the average rank correlation of
verbs in that class
• some verb classes did better than others but no outliers
• many of the top/bottom verb classes were the same for
Word2Vec embeddings and DEP1 embeddings
• What attributes cause certain verb classes to do consistently better?
Word2Vec DEP1
Top Class consider-29.9 cooperate-73
Rank of Top Class 0.1642 0.299
Bottom Class light_emission-43.1 exhale-40.1.3
Rank of Bottom Class -0.0151 -0.1473
25. Extrinsic Evaluation:
Sentiment Analysis
• Task is to categorizes sentences according to their positive
or negative sentiment
• “I hate this movie” VS “This movie is ridiculously good”
• Using Kaggle Challenge data from Rotten Tomatoes Movie
Reviews
• System is composed of a convolutional neural network that
is feed the word vectors of the words in the sentence
28. Conclusions
• Structure based embeddings are better then topical based
embeddings
• Shown by the evaluations on WordNet, VerbNet and
sentiment analysis task case study
• Different POS capture semantic information from
different sentence structures
• Within verbs embeddings: certain subclasses of verbs do
better than others
• Verbs with certain semantic restricts far outperform other
verb classes