SlideShare a Scribd company logo
1 of 37
Download to read offline
A Study on Compositional Semantics
  of Words in Distributional Spaces
     Pierpaolo Basile, Annalina Caputo,
            Giovanni Semeraro
             Dept. of Computer Science
            University of Bari Aldo Moro
                 {basilepp}@di.uniba.it



               6th IEEE International Conference on Semantic Computing
               September 19-21, 2012 Palermo, Italy
Background…
You shall know a word by
the company it keeps!

Meaning of a word is
determined by its usage




                           2
…Background
Distributional Semantic Models (DSMs) defined
  as: <T, C, R, W, M, d, S>
  – T: target elements (words)
  – C: context
  – R: relation between T and C
  – W: weighting schema
  – M: geometric space TxC
  – d: space reduction M -> M’
  – S: similarity function defined in M’
Motivations
• One definition of context at a time
  – encode syntactic information in DSMs
• Words are represented in isolation
  – syntactic role could be used as a glue to compose
    words
  It’s raining cats and dogs = My cats and dogs are in the rain
Outline
• Simple DSMs and simple operators
• Syntactic dependencies in DSMs
  – Structured DSMs
  – Compositional operators
• Evaluation and results
• Final remarks
SIMPLE DSMS AND SIMPLE
OPERATORS
Simple DSMs…
Term-term co-occurrence matrix (TTM): each
cell contains the co-occurrences between two
terms within a prefixed distance
               dog   cat   computer   animal   mouse
      dog      0     4        0         2        1
      cat      4     0        0         3        5
    computer   0     0        0         0        3
     animal    2     3        0         0        2
     mouse     1     5        3         2        0
…Simple DSMs
Latent Semantic Analysis (LSA): relies on the
Singular Value Decomposition (SVD) of the co-
occurrence matrix
Random Indexing (RI): based on the Random
Projection
Latent Semantic Analysis over Random Indexing
(RILSA)
Random Indexing
• Create and assign a context vector to each
  context element (e.g. document, passage,
  term, …)
• Term vector is the sum of the context vectors
  in which the term occurs
  – sometimes the context vector could be boosted by
    a score (e.g. term frequency, PMI, …)



                                                   9
Context Vector

    0 0 0 0 0 0 0 -1 0 0 0 0 1 0 0 -1 0 1 0 0 0 0 1 0 0 0 0 -1



•   sparse
•   high dimensional
•   ternary {-1, 0, +1}
•   small number of randomly distributed non-
    zero elements
                                                                 10
Random Indexing (formal)


                                    B =A R
                                      n,k       n,m   m,k
                                                             k << m
                                    B nearly preserves the
                                    distance between points
                                     (Johnson-Lindenstrauss lemma)
                                    dr = c× d
RI is a locality-sensitive hashing method which approximate the cosine
distance between vectors
                                                                     11
Random Indexing (example)

                                CVjohn -> (0, 0, 0, 0, 0, 0, 1, 0, -1, 0)
John eats a red apple           CVeat -> (1, 0, 0, 0, -1, 0, 0 ,0 ,0 ,0)
                                CVred -> (0, 0, 0, 1, 0, 0, 0, -1, 0, 0)



TVapple= CVjohn+ CVeat+CVred=(1, 0, 0, 1, -1, 0, 1, -1, -1, 0)




                                                                     12
Latent Semantic Analysis over Random
Indexing
1. Reduce the dimension of the co-occurrences
   matrix using RI
2. Perform LSA over RI (LSARI)
  – reduction of LSA computation time: RI matrix
    contains less dimensions than co-occurrences
    matrix
Simple operators…
Addition (+): pointwise sum of components
Multiplication (∘): pointwise multiplication of
  components
Addition and multiplication are commutative
  – do not take into account word order
Complex structures represented summing or
multiplying words which compose them
…Simple operators
Given two word vectors u and v
  – composition by sum p = u + v
  – composition by multiplication p = u ∘ v
Can be applied to any sequence of words
SYNTACTIC DEPENDENCIES IN DSMS
Syntactic dependencies…

            John eats a red apple.
         subject              object



  John               eats                  apple


                                modifier
                     red

                                                   17
…Syntactic dependencies

               John eats a red apple.
DEPENDENT             HEAD
            subject              object



    John                eats                  apple


                                   modifier
                        red

                                                      18
Representing dependences
Use filler/role binding approach to represent a
dependency dep(u, v)
                   rd⊚u + rh⊚v

rd and rh are vectors which represent
respectively the role of dependent and head
⊚ is a placeholder for a composition operator
Representing dependences (example)

obj(apple, eat)

           rd⊚apple + rh⊚eat


                  role vectors
Structured DSMs
1. Vector permutation in RI (PERM) to encode
   dependencies
2. Circular convolution (CONV) as filler/binding
   operator to represent syntactic dependencies
   in DSMs
3. LSA over PERM and CONV carries out two
   spaces: PERMLSA and CONVLSA
Vector permutation in RI (PERM)
Using permutation of elements in context
vectors to encode dependencies
  – right rotation of n elements to encode
    dependents (permutation)
  – left rotation of n elements to encode heads
    (inverse permutation)




                                                  22
PERM (method)
Create and assign a context vector to each term
Assign a rotation function Π+1 to the dependent and
Π-1 to the head
Each term is represented by a vector which is
  – the sum of the permuted vectors of all the dependent
    terms
  – the sum of the inverse permuted vectors of all the
    head terms
  – the sum of the no-permuted vectors of both
    dependent and head words

                                                       23
PERM (example…)
                                John -> (0, 0, 0, 0, 0, 0, 1, 0, -1, 0)
                                eat -> (1, 0, 0, 0, -1, 0, 0 ,0 ,0 ,0)
John eats a red apple           red -> (0, 0, 0, 1, 0, 0, 0, -1, 0, 0)
                                apple -> (1, 0, 0, 0, 0, 0, 0, -1, 0, 0)



TVapple = Π+1(CVred) + Π-1(CVeat) + CVred + CVeat

     red
                      apple                    eats

                                                                    24
PERM (…example)
                                   John -> (0, 0, 0, 0, 0, 0, 1, 0, -1, 0)
                                   eat -> (1, 0, 0, 0, -1, 0, 0 ,0 ,0 ,0)
 John eats a red apple             red-> (0, 0, 0, 1, 0, 0, 0, -1, 0, 0)
                                   apple -> (1, 0, 0, 0, 0, 0, 0, -1, 0, 0)



TVapple = Π+1(CVred) + Π-1(CVeat) + CVred + CVeat=…
                          right shift                           left shift
…=(0, 0, 0, 0, 1, 0, 0, 0, -1, 0) + (0, 0, 0, -1, 0, 0, 0, 0, 1) +

+ (0, 0, 0, 1, 0, 0, 0, -1, 0, 0) + (1, 0, 0, 0, -1, 0, 0 ,0 ,0 ,0)
                                                                       25
Convolution (CONV)
Create and assign a context vector to each term
Create two context vectors for head and dependent
roles
Each term is represented by a vector which is
  – the sum of the convolution between dependent terms
    and the dependent role vector
  – the sum of the convolution between head terms and
    the head role vector
  – the sum of the vectors of both dependent and head
    words
Circular convolution operator
Circular convolution                                   U=<1, 1, -1, -1, 1>
                                                       V=<1, -1, 1, -1, -1>
        p=u⊛v                                          P=<-1, 3, -1, -1, -1>
defined as:                                               U1    U2    U3       U4   U5
             n
                                                  V1      1     1     -1       -1   1
 p   j
                     u k v( j   k ) ( n 1)   P1   V2      -1    -1    1        1    -1
         k       1                           P2   V3      1     1     -1       -1   1
                                             P3   V4      -1    -1    1        1    -1
                                             P4
                                                  V5      -1    -1    1        1    -1
                                             P5
Circular convolution by FFTs
Circular convolution is computed in O(n2)
   – using FFTs is computed in O(nlogn)
Given f the discrete FFTs and f -1 its inverse
   – u ⊛v = f -1( f(u) ∘ f (v) )
CONV (example)
                                       John -> (0, 0, 0, 0, 0, 0, 1, 0, -1, 0)
                                       eat -> (1, 0, 0, 0, -1, 0, 0 ,0 ,0 ,0)
John eats a red apple                  red-> (0, 0, 0, 1, 0, 0, 0, -1, 0, 0)
                                       apple -> (1, 0, 0, 0, 0, 0, 0, -1, 0, 0)
                                       rd -> (0, 0, 1, 0, -1, 0, 0, 0, 0, 0)
                                       rh -> (0, -1, 1, 0, 0, 0, 0, 0, 0, 0)
apple = eat +red + (rd⊛ red) + (rh ⊛ eat)
   Context vector for dependent role
                           Context vector for head role



                                                                           29
Complex operators
Based on filler/role binding taking into account
syntactic role: rd ⊚ u + rh ⊚ v
  – u and v could be recursive structures
Two vector operators to bind the role:
  – convolution (⊛)
  – tensor (⊗)
  – convolution (⊛+): exploits also the sum of term
  vectors
                 rd ⊛ u + rh ⊛ v + v + u
Complex operators (remarks)
Existing operators
  – t1 ⊚ t2 ⊚ … ⊚ tn: does not take into account
    syntactic role
  – t1 ⊛ t2 is commutative
  – t1⊗t2 ⊗ … ⊗tn: tensor order depends on the
    phrase length
     • two phrases with different length are not comparable
  – t1 ⊗ r1 ⊗t2 ⊗r2 ⊗ … ⊗ tn ⊗rn : also depends on
    the sentence length
System setup
• Corpus
  – WaCkypedia EN based on a 2009 dump of Wikipedia
  – about 800 million tokens
  – dependency parse by MaltParser
• DSMs
  –   500 vector dimension (LSA/RI/RILSA)
  –   1,000 vector dimension (PERM/CONV/PERMLSA/CONVLSA)
  –   50,000 most frequent words
  –   co-occurrence distance: 4


                                                           32
Evaluation
• GEMS 2011 Shared Task for compositional
  semantics
  – list of two pairs of words combination
     (support offer) (help provide) 7
     (old person) (right hand) 1
     • rated by humans
     • 5,833 rates
     • 3 types involved: noun-noun (NN), adjective-noun (AN),
       verb-object (VO)
  – GOAL: compare the system performance against
    humans scores
     • Spearman correlation


                                                                33
Results (simple spaces)…
                    NN                         AN                            VO




                                                                 TTM
        TTM




                                   TTM
                           RILSA




                                                       RILSA




                                                                                     RILSA
              LSA




                                         LSA




                                                                       LSA
                      RI




                                                 RI




                                                                               RI
+       .21   .36    .25   .42     .22   .35     .33   .41       .23   .31     .28   .31

∘       .31   .15    .23   .22     .21   .20     .22   .18       .13   .10     .18   .21

⊛       .21   .38    .26   .35     .20   .33     .31   .44       .15   .31     .24   .34

⊛+      .21   .34    .28   .43     .23   .32     .31   .37       .20   .31     .25   .29

⊗       .21   .38    .25   .39     .22   .38     .33   .43       .15   .34     .26   .32

human               .49                        .52                           .55


                                                               Simple Semantic Spaces
…Results (structured spaces)
                      NN                                  AN                                  VO




                                  PERMLSA




                                                                      PERMLSA




                                                                                                          PERMLSA
                        CONVLSA




                                                            CONVLSA




                                                                                                CONVLSA
               PERM




                                                   PERM




                                                                                       PERM
        CONV




                                            CONV




                                                                                CONV
+       .36    .39     .43        .42       .34    .39      .42       .45       .27    .23      .30       .31

∘       .22    .17     .10        .13       .23    .27      .13       .15       .20    .15      .06       .14

⊛       .31    .36     .37        .35       .39    .39      .45       .44       .28    .23      .27       .28

⊛+      .30    .36     .40        .36       .38    .32      .48       .44       .27    .22      .30       .32

⊗       .34    .37     .37        .40       .36    .40      .45       .45       .27    .24      .31       .32

human                 .49                                 .52                                 .55


                                                                      Structured Semantic Spaces
Final remarks
• Best results are obtained when complex
  operators/spaces (or both) are involved
• No best combination of operator/space exists
   – depend on the type of relation (NN, AN, VO)
• Tensor product and convolution provide good results in
  spite of previous results
   – filler/role binding is effective
• Future work
   – generate several rd and rh vectors for each kind of
     dependency
   – apply this approach to other direct graph-based
     representations
Thank you for your attention!

         Questions?
        <basilepp@di.uniba.it>

More Related Content

What's hot

11.homotopy perturbation and elzaki transform for solving nonlinear partial d...
11.homotopy perturbation and elzaki transform for solving nonlinear partial d...11.homotopy perturbation and elzaki transform for solving nonlinear partial d...
11.homotopy perturbation and elzaki transform for solving nonlinear partial d...Alexander Decker
 
11.solution of linear and nonlinear partial differential equations using mixt...
11.solution of linear and nonlinear partial differential equations using mixt...11.solution of linear and nonlinear partial differential equations using mixt...
11.solution of linear and nonlinear partial differential equations using mixt...Alexander Decker
 
Solution of linear and nonlinear partial differential equations using mixture...
Solution of linear and nonlinear partial differential equations using mixture...Solution of linear and nonlinear partial differential equations using mixture...
Solution of linear and nonlinear partial differential equations using mixture...Alexander Decker
 
NONLINEAR DIFFERENCE EQUATIONS WITH SMALL PARAMETERS OF MULTIPLE SCALES
NONLINEAR DIFFERENCE EQUATIONS WITH SMALL PARAMETERS OF MULTIPLE SCALESNONLINEAR DIFFERENCE EQUATIONS WITH SMALL PARAMETERS OF MULTIPLE SCALES
NONLINEAR DIFFERENCE EQUATIONS WITH SMALL PARAMETERS OF MULTIPLE SCALESTahia ZERIZER
 
Lesson 15: Inverse Functions and Logarithms
Lesson 15: Inverse Functions and LogarithmsLesson 15: Inverse Functions and Logarithms
Lesson 15: Inverse Functions and LogarithmsMatthew Leingang
 
Mesh Processing Course : Differential Calculus
Mesh Processing Course : Differential CalculusMesh Processing Course : Differential Calculus
Mesh Processing Course : Differential CalculusGabriel Peyré
 
11 the inverse trigonometric functions x
11 the inverse trigonometric functions x11 the inverse trigonometric functions x
11 the inverse trigonometric functions xmath266
 
Lesson 15: Inverse Functions and Logarithms
Lesson 15: Inverse Functions and LogarithmsLesson 15: Inverse Functions and Logarithms
Lesson 15: Inverse Functions and LogarithmsMatthew Leingang
 
Steven Duplij, "Developing new supermanifolds by revitalizing old ideas"
Steven Duplij, "Developing new supermanifolds by revitalizing old ideas"Steven Duplij, "Developing new supermanifolds by revitalizing old ideas"
Steven Duplij, "Developing new supermanifolds by revitalizing old ideas"Steven Duplij (Stepan Douplii)
 
Inverse circular function
Inverse circular functionInverse circular function
Inverse circular functionAPEX INSTITUTE
 
D. Ishii, K. Ueda, H. Hosobe, A. Goldsztejn: Interval-based Solving of Hybrid...
D. Ishii, K. Ueda, H. Hosobe, A. Goldsztejn: Interval-based Solving of Hybrid...D. Ishii, K. Ueda, H. Hosobe, A. Goldsztejn: Interval-based Solving of Hybrid...
D. Ishii, K. Ueda, H. Hosobe, A. Goldsztejn: Interval-based Solving of Hybrid...dishii
 
11.[95 103]solution of telegraph equation by modified of double sumudu transf...
11.[95 103]solution of telegraph equation by modified of double sumudu transf...11.[95 103]solution of telegraph equation by modified of double sumudu transf...
11.[95 103]solution of telegraph equation by modified of double sumudu transf...Alexander Decker
 
Recursive Formulation of Gradient in a Dense Feed-Forward Deep Neural Network
Recursive Formulation of Gradient in a Dense Feed-Forward Deep Neural NetworkRecursive Formulation of Gradient in a Dense Feed-Forward Deep Neural Network
Recursive Formulation of Gradient in a Dense Feed-Forward Deep Neural NetworkAshwin Rao
 

What's hot (20)

11.homotopy perturbation and elzaki transform for solving nonlinear partial d...
11.homotopy perturbation and elzaki transform for solving nonlinear partial d...11.homotopy perturbation and elzaki transform for solving nonlinear partial d...
11.homotopy perturbation and elzaki transform for solving nonlinear partial d...
 
11.solution of linear and nonlinear partial differential equations using mixt...
11.solution of linear and nonlinear partial differential equations using mixt...11.solution of linear and nonlinear partial differential equations using mixt...
11.solution of linear and nonlinear partial differential equations using mixt...
 
Confluent hypergeometricfunctions
Confluent hypergeometricfunctionsConfluent hypergeometricfunctions
Confluent hypergeometricfunctions
 
Solution of linear and nonlinear partial differential equations using mixture...
Solution of linear and nonlinear partial differential equations using mixture...Solution of linear and nonlinear partial differential equations using mixture...
Solution of linear and nonlinear partial differential equations using mixture...
 
NONLINEAR DIFFERENCE EQUATIONS WITH SMALL PARAMETERS OF MULTIPLE SCALES
NONLINEAR DIFFERENCE EQUATIONS WITH SMALL PARAMETERS OF MULTIPLE SCALESNONLINEAR DIFFERENCE EQUATIONS WITH SMALL PARAMETERS OF MULTIPLE SCALES
NONLINEAR DIFFERENCE EQUATIONS WITH SMALL PARAMETERS OF MULTIPLE SCALES
 
Lesson 15: Inverse Functions and Logarithms
Lesson 15: Inverse Functions and LogarithmsLesson 15: Inverse Functions and Logarithms
Lesson 15: Inverse Functions and Logarithms
 
Mesh Processing Course : Differential Calculus
Mesh Processing Course : Differential CalculusMesh Processing Course : Differential Calculus
Mesh Processing Course : Differential Calculus
 
12 - Overview
12 - Overview12 - Overview
12 - Overview
 
11 the inverse trigonometric functions x
11 the inverse trigonometric functions x11 the inverse trigonometric functions x
11 the inverse trigonometric functions x
 
Lesson 15: Inverse Functions and Logarithms
Lesson 15: Inverse Functions and LogarithmsLesson 15: Inverse Functions and Logarithms
Lesson 15: Inverse Functions and Logarithms
 
Bernoulli eulerriemannzeta
Bernoulli eulerriemannzetaBernoulli eulerriemannzeta
Bernoulli eulerriemannzeta
 
z transforms
z transformsz transforms
z transforms
 
Steven Duplij, "Developing new supermanifolds by revitalizing old ideas"
Steven Duplij, "Developing new supermanifolds by revitalizing old ideas"Steven Duplij, "Developing new supermanifolds by revitalizing old ideas"
Steven Duplij, "Developing new supermanifolds by revitalizing old ideas"
 
Limits and derivatives
Limits and derivativesLimits and derivatives
Limits and derivatives
 
Inverse circular function
Inverse circular functionInverse circular function
Inverse circular function
 
D. Ishii, K. Ueda, H. Hosobe, A. Goldsztejn: Interval-based Solving of Hybrid...
D. Ishii, K. Ueda, H. Hosobe, A. Goldsztejn: Interval-based Solving of Hybrid...D. Ishii, K. Ueda, H. Hosobe, A. Goldsztejn: Interval-based Solving of Hybrid...
D. Ishii, K. Ueda, H. Hosobe, A. Goldsztejn: Interval-based Solving of Hybrid...
 
DIFFERENTIATION
DIFFERENTIATIONDIFFERENTIATION
DIFFERENTIATION
 
11.[95 103]solution of telegraph equation by modified of double sumudu transf...
11.[95 103]solution of telegraph equation by modified of double sumudu transf...11.[95 103]solution of telegraph equation by modified of double sumudu transf...
11.[95 103]solution of telegraph equation by modified of double sumudu transf...
 
Recursive Formulation of Gradient in a Dense Feed-Forward Deep Neural Network
Recursive Formulation of Gradient in a Dense Feed-Forward Deep Neural NetworkRecursive Formulation of Gradient in a Dense Feed-Forward Deep Neural Network
Recursive Formulation of Gradient in a Dense Feed-Forward Deep Neural Network
 
Fourier series
Fourier seriesFourier series
Fourier series
 

Viewers also liked

How to Determine which Algorithms Really Matter
How to Determine which Algorithms Really MatterHow to Determine which Algorithms Really Matter
How to Determine which Algorithms Really MatterDataWorks Summit
 
An introduction to compositional models in distributional semantics
An introduction to compositional models in distributional semanticsAn introduction to compositional models in distributional semantics
An introduction to compositional models in distributional semanticsAndre Freitas
 
Elements of Text Mining Part - I
Elements of Text Mining Part - IElements of Text Mining Part - I
Elements of Text Mining Part - IJaganadh Gopinadhan
 
Drawing word2vec
Drawing word2vecDrawing word2vec
Drawing word2vecKai Sasaki
 
Word2Vec: Learning of word representations in a vector space - Di Mitri & Her...
Word2Vec: Learning of word representations in a vector space - Di Mitri & Her...Word2Vec: Learning of word representations in a vector space - Di Mitri & Her...
Word2Vec: Learning of word representations in a vector space - Di Mitri & Her...Daniele Di Mitri
 
Neoclásico. Contexto político, social. Las artes en el periodo.
Neoclásico. Contexto político, social. Las artes en el periodo.Neoclásico. Contexto político, social. Las artes en el periodo.
Neoclásico. Contexto político, social. Las artes en el periodo.Claudia Magos
 
Word2vec algorithm
Word2vec algorithmWord2vec algorithm
Word2vec algorithmAndrew Koo
 

Viewers also liked (8)

How to Determine which Algorithms Really Matter
How to Determine which Algorithms Really MatterHow to Determine which Algorithms Really Matter
How to Determine which Algorithms Really Matter
 
An introduction to compositional models in distributional semantics
An introduction to compositional models in distributional semanticsAn introduction to compositional models in distributional semantics
An introduction to compositional models in distributional semantics
 
Elements of Text Mining Part - I
Elements of Text Mining Part - IElements of Text Mining Part - I
Elements of Text Mining Part - I
 
Drawing word2vec
Drawing word2vecDrawing word2vec
Drawing word2vec
 
Word2Vec: Learning of word representations in a vector space - Di Mitri & Her...
Word2Vec: Learning of word representations in a vector space - Di Mitri & Her...Word2Vec: Learning of word representations in a vector space - Di Mitri & Her...
Word2Vec: Learning of word representations in a vector space - Di Mitri & Her...
 
Neoclásico. Contexto político, social. Las artes en el periodo.
Neoclásico. Contexto político, social. Las artes en el periodo.Neoclásico. Contexto político, social. Las artes en el periodo.
Neoclásico. Contexto político, social. Las artes en el periodo.
 
Semantics
SemanticsSemantics
Semantics
 
Word2vec algorithm
Word2vec algorithmWord2vec algorithm
Word2vec algorithm
 

Similar to A Study on Compositional Semantics of Words in Distributional Spaces

Building WordSpaces via Random Indexing from simple to complex spaces
Building WordSpaces via Random Indexing from simple to complex spacesBuilding WordSpaces via Random Indexing from simple to complex spaces
Building WordSpaces via Random Indexing from simple to complex spacesPierpaolo Basile
 
"SSC" - Geometria e Semantica del Linguaggio
"SSC" - Geometria e Semantica del Linguaggio"SSC" - Geometria e Semantica del Linguaggio
"SSC" - Geometria e Semantica del LinguaggioAlumni Mathematica
 
Lecture 2: Stochastic Hydrology
Lecture 2: Stochastic Hydrology Lecture 2: Stochastic Hydrology
Lecture 2: Stochastic Hydrology Amro Elfeki
 
Queues internet src2
Queues internet  src2Queues internet  src2
Queues internet src2Ammulu Amma
 
Constraint Programming in Haskell
Constraint Programming in HaskellConstraint Programming in Haskell
Constraint Programming in HaskellDavid Overton
 
The Wishart and inverse-wishart distribution
 The Wishart and inverse-wishart distribution The Wishart and inverse-wishart distribution
The Wishart and inverse-wishart distributionPankaj Das
 
Generating Chebychev Chaotic Sequence
Generating Chebychev Chaotic SequenceGenerating Chebychev Chaotic Sequence
Generating Chebychev Chaotic SequenceCheng-An Yang
 
Encoding syntactic dependencies by vector permutation
Encoding syntactic dependencies by vector permutationEncoding syntactic dependencies by vector permutation
Encoding syntactic dependencies by vector permutationPierpaolo Basile
 
20121020 semi local-string_comparison_tiskin
20121020 semi local-string_comparison_tiskin20121020 semi local-string_comparison_tiskin
20121020 semi local-string_comparison_tiskinComputer Science Club
 
Chapter 3 REGULAR EXPRESSION.pdf
Chapter 3 REGULAR EXPRESSION.pdfChapter 3 REGULAR EXPRESSION.pdf
Chapter 3 REGULAR EXPRESSION.pdfdawod yimer
 
Linear Regression
Linear Regression Linear Regression
Linear Regression Rupak Roy
 
MVPA with SpaceNet: sparse structured priors
MVPA with SpaceNet: sparse structured priorsMVPA with SpaceNet: sparse structured priors
MVPA with SpaceNet: sparse structured priorsElvis DOHMATOB
 

Similar to A Study on Compositional Semantics of Words in Distributional Spaces (20)

Building WordSpaces via Random Indexing from simple to complex spaces
Building WordSpaces via Random Indexing from simple to complex spacesBuilding WordSpaces via Random Indexing from simple to complex spaces
Building WordSpaces via Random Indexing from simple to complex spaces
 
"SSC" - Geometria e Semantica del Linguaggio
"SSC" - Geometria e Semantica del Linguaggio"SSC" - Geometria e Semantica del Linguaggio
"SSC" - Geometria e Semantica del Linguaggio
 
Lecture 2: Stochastic Hydrology
Lecture 2: Stochastic Hydrology Lecture 2: Stochastic Hydrology
Lecture 2: Stochastic Hydrology
 
Matlab II
Matlab IIMatlab II
Matlab II
 
Hak ontoforum
Hak ontoforumHak ontoforum
Hak ontoforum
 
Queues internet src2
Queues internet  src2Queues internet  src2
Queues internet src2
 
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
 
Constraint Programming in Haskell
Constraint Programming in HaskellConstraint Programming in Haskell
Constraint Programming in Haskell
 
The Wishart and inverse-wishart distribution
 The Wishart and inverse-wishart distribution The Wishart and inverse-wishart distribution
The Wishart and inverse-wishart distribution
 
Generating Chebychev Chaotic Sequence
Generating Chebychev Chaotic SequenceGenerating Chebychev Chaotic Sequence
Generating Chebychev Chaotic Sequence
 
Encoding syntactic dependencies by vector permutation
Encoding syntactic dependencies by vector permutationEncoding syntactic dependencies by vector permutation
Encoding syntactic dependencies by vector permutation
 
20121020 semi local-string_comparison_tiskin
20121020 semi local-string_comparison_tiskin20121020 semi local-string_comparison_tiskin
20121020 semi local-string_comparison_tiskin
 
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
 
Chapter 3 REGULAR EXPRESSION.pdf
Chapter 3 REGULAR EXPRESSION.pdfChapter 3 REGULAR EXPRESSION.pdf
Chapter 3 REGULAR EXPRESSION.pdf
 
It 05104 digsig_1
It 05104 digsig_1It 05104 digsig_1
It 05104 digsig_1
 
Digital signal processing part2
Digital signal processing part2Digital signal processing part2
Digital signal processing part2
 
Linear Regression
Linear Regression Linear Regression
Linear Regression
 
Sobolev spaces
Sobolev spacesSobolev spaces
Sobolev spaces
 
unit 2 part b.docx
unit 2 part b.docxunit 2 part b.docx
unit 2 part b.docx
 
MVPA with SpaceNet: sparse structured priors
MVPA with SpaceNet: sparse structured priorsMVPA with SpaceNet: sparse structured priors
MVPA with SpaceNet: sparse structured priors
 

More from Pierpaolo Basile

Diachronic analysis of entities by exploiting wikipedia page revisions
Diachronic analysis of entities by exploiting wikipedia page revisionsDiachronic analysis of entities by exploiting wikipedia page revisions
Diachronic analysis of entities by exploiting wikipedia page revisionsPierpaolo Basile
 
Come l'industria tecnologica ha cancellato le donne dalla storia
Come l'industria tecnologica ha cancellato le donne dalla storiaCome l'industria tecnologica ha cancellato le donne dalla storia
Come l'industria tecnologica ha cancellato le donne dalla storiaPierpaolo Basile
 
EVALITA 2018 NLP4FUN - Solving language games
EVALITA 2018 NLP4FUN - Solving language gamesEVALITA 2018 NLP4FUN - Solving language games
EVALITA 2018 NLP4FUN - Solving language gamesPierpaolo Basile
 
Buon appetito! Analyzing Happiness in Italian Tweets
Buon appetito! Analyzing Happiness in Italian TweetsBuon appetito! Analyzing Happiness in Italian Tweets
Buon appetito! Analyzing Happiness in Italian TweetsPierpaolo Basile
 
Detecting semantic shift in large corpora by exploiting temporal random indexing
Detecting semantic shift in large corpora by exploiting temporal random indexingDetecting semantic shift in large corpora by exploiting temporal random indexing
Detecting semantic shift in large corpora by exploiting temporal random indexingPierpaolo Basile
 
Bi-directional LSTM-CNNs-CRF for Italian Sequence Labeling
Bi-directional LSTM-CNNs-CRF for Italian Sequence LabelingBi-directional LSTM-CNNs-CRF for Italian Sequence Labeling
Bi-directional LSTM-CNNs-CRF for Italian Sequence LabelingPierpaolo Basile
 
INSERT COIN - Storia dei videogame: da Spacewar a Street Fighter
INSERT COIN - Storia dei videogame: da Spacewar a Street FighterINSERT COIN - Storia dei videogame: da Spacewar a Street Fighter
INSERT COIN - Storia dei videogame: da Spacewar a Street FighterPierpaolo Basile
 
QuestionCube DigithON 2017
QuestionCube DigithON 2017QuestionCube DigithON 2017
QuestionCube DigithON 2017Pierpaolo Basile
 
Diachronic Analysis of the Italian Language exploiting Google Ngram
Diachronic Analysis of the Italian Language exploiting Google NgramDiachronic Analysis of the Italian Language exploiting Google Ngram
Diachronic Analysis of the Italian Language exploiting Google NgramPierpaolo Basile
 
La macchina più geek dell’universo The Turing Machine
La macchina più geek dell’universo The Turing MachineLa macchina più geek dell’universo The Turing Machine
La macchina più geek dell’universo The Turing MachinePierpaolo Basile
 
UNIBA: Exploiting a Distributional Semantic Model for Disambiguating and Link...
UNIBA: Exploiting a Distributional Semantic Model for Disambiguating and Link...UNIBA: Exploiting a Distributional Semantic Model for Disambiguating and Link...
UNIBA: Exploiting a Distributional Semantic Model for Disambiguating and Link...Pierpaolo Basile
 
Analysing Word Meaning over Time by Exploiting Temporal Random Indexing
Analysing Word Meaning over Time by Exploiting Temporal Random IndexingAnalysing Word Meaning over Time by Exploiting Temporal Random Indexing
Analysing Word Meaning over Time by Exploiting Temporal Random IndexingPierpaolo Basile
 
COLING 2014 - An Enhanced Lesk Word Sense Disambiguation Algorithm through a ...
COLING 2014 - An Enhanced Lesk Word Sense Disambiguation Algorithm through a ...COLING 2014 - An Enhanced Lesk Word Sense Disambiguation Algorithm through a ...
COLING 2014 - An Enhanced Lesk Word Sense Disambiguation Algorithm through a ...Pierpaolo Basile
 
Exploiting Distributional Semantic Models in Question Answering
Exploiting Distributional Semantic Models in Question AnsweringExploiting Distributional Semantic Models in Question Answering
Exploiting Distributional Semantic Models in Question AnsweringPierpaolo Basile
 
Sst evalita2011 basile_pierpaolo
Sst evalita2011 basile_pierpaoloSst evalita2011 basile_pierpaolo
Sst evalita2011 basile_pierpaoloPierpaolo Basile
 
AI*IA 2012 PAI Workshop OTTHO
AI*IA 2012 PAI Workshop OTTHOAI*IA 2012 PAI Workshop OTTHO
AI*IA 2012 PAI Workshop OTTHOPierpaolo Basile
 
Word Sense Disambiguation and Intelligent Information Access
Word Sense Disambiguation and Intelligent Information AccessWord Sense Disambiguation and Intelligent Information Access
Word Sense Disambiguation and Intelligent Information AccessPierpaolo Basile
 

More from Pierpaolo Basile (19)

Diachronic analysis of entities by exploiting wikipedia page revisions
Diachronic analysis of entities by exploiting wikipedia page revisionsDiachronic analysis of entities by exploiting wikipedia page revisions
Diachronic analysis of entities by exploiting wikipedia page revisions
 
Come l'industria tecnologica ha cancellato le donne dalla storia
Come l'industria tecnologica ha cancellato le donne dalla storiaCome l'industria tecnologica ha cancellato le donne dalla storia
Come l'industria tecnologica ha cancellato le donne dalla storia
 
EVALITA 2018 NLP4FUN - Solving language games
EVALITA 2018 NLP4FUN - Solving language gamesEVALITA 2018 NLP4FUN - Solving language games
EVALITA 2018 NLP4FUN - Solving language games
 
Buon appetito! Analyzing Happiness in Italian Tweets
Buon appetito! Analyzing Happiness in Italian TweetsBuon appetito! Analyzing Happiness in Italian Tweets
Buon appetito! Analyzing Happiness in Italian Tweets
 
Detecting semantic shift in large corpora by exploiting temporal random indexing
Detecting semantic shift in large corpora by exploiting temporal random indexingDetecting semantic shift in large corpora by exploiting temporal random indexing
Detecting semantic shift in large corpora by exploiting temporal random indexing
 
Bi-directional LSTM-CNNs-CRF for Italian Sequence Labeling
Bi-directional LSTM-CNNs-CRF for Italian Sequence LabelingBi-directional LSTM-CNNs-CRF for Italian Sequence Labeling
Bi-directional LSTM-CNNs-CRF for Italian Sequence Labeling
 
INSERT COIN - Storia dei videogame: da Spacewar a Street Fighter
INSERT COIN - Storia dei videogame: da Spacewar a Street FighterINSERT COIN - Storia dei videogame: da Spacewar a Street Fighter
INSERT COIN - Storia dei videogame: da Spacewar a Street Fighter
 
QuestionCube DigithON 2017
QuestionCube DigithON 2017QuestionCube DigithON 2017
QuestionCube DigithON 2017
 
Diachronic Analysis of the Italian Language exploiting Google Ngram
Diachronic Analysis of the Italian Language exploiting Google NgramDiachronic Analysis of the Italian Language exploiting Google Ngram
Diachronic Analysis of the Italian Language exploiting Google Ngram
 
Diachronic Analysis
Diachronic AnalysisDiachronic Analysis
Diachronic Analysis
 
(Open) data hacking
(Open) data hacking(Open) data hacking
(Open) data hacking
 
La macchina più geek dell’universo The Turing Machine
La macchina più geek dell’universo The Turing MachineLa macchina più geek dell’universo The Turing Machine
La macchina più geek dell’universo The Turing Machine
 
UNIBA: Exploiting a Distributional Semantic Model for Disambiguating and Link...
UNIBA: Exploiting a Distributional Semantic Model for Disambiguating and Link...UNIBA: Exploiting a Distributional Semantic Model for Disambiguating and Link...
UNIBA: Exploiting a Distributional Semantic Model for Disambiguating and Link...
 
Analysing Word Meaning over Time by Exploiting Temporal Random Indexing
Analysing Word Meaning over Time by Exploiting Temporal Random IndexingAnalysing Word Meaning over Time by Exploiting Temporal Random Indexing
Analysing Word Meaning over Time by Exploiting Temporal Random Indexing
 
COLING 2014 - An Enhanced Lesk Word Sense Disambiguation Algorithm through a ...
COLING 2014 - An Enhanced Lesk Word Sense Disambiguation Algorithm through a ...COLING 2014 - An Enhanced Lesk Word Sense Disambiguation Algorithm through a ...
COLING 2014 - An Enhanced Lesk Word Sense Disambiguation Algorithm through a ...
 
Exploiting Distributional Semantic Models in Question Answering
Exploiting Distributional Semantic Models in Question AnsweringExploiting Distributional Semantic Models in Question Answering
Exploiting Distributional Semantic Models in Question Answering
 
Sst evalita2011 basile_pierpaolo
Sst evalita2011 basile_pierpaoloSst evalita2011 basile_pierpaolo
Sst evalita2011 basile_pierpaolo
 
AI*IA 2012 PAI Workshop OTTHO
AI*IA 2012 PAI Workshop OTTHOAI*IA 2012 PAI Workshop OTTHO
AI*IA 2012 PAI Workshop OTTHO
 
Word Sense Disambiguation and Intelligent Information Access
Word Sense Disambiguation and Intelligent Information AccessWord Sense Disambiguation and Intelligent Information Access
Word Sense Disambiguation and Intelligent Information Access
 

Recently uploaded

Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...Drew Madelung
 
Artificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and MythsArtificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and MythsJoaquim Jorge
 
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...apidays
 
GenAI Risks & Security Meetup 01052024.pdf
GenAI Risks & Security Meetup 01052024.pdfGenAI Risks & Security Meetup 01052024.pdf
GenAI Risks & Security Meetup 01052024.pdflior mazor
 
🐬 The future of MySQL is Postgres 🐘
🐬  The future of MySQL is Postgres   🐘🐬  The future of MySQL is Postgres   🐘
🐬 The future of MySQL is Postgres 🐘RTylerCroy
 
Partners Life - Insurer Innovation Award 2024
Partners Life - Insurer Innovation Award 2024Partners Life - Insurer Innovation Award 2024
Partners Life - Insurer Innovation Award 2024The Digital Insurer
 
What Are The Drone Anti-jamming Systems Technology?
What Are The Drone Anti-jamming Systems Technology?What Are The Drone Anti-jamming Systems Technology?
What Are The Drone Anti-jamming Systems Technology?Antenna Manufacturer Coco
 
Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024The Digital Insurer
 
A Year of the Servo Reboot: Where Are We Now?
A Year of the Servo Reboot: Where Are We Now?A Year of the Servo Reboot: Where Are We Now?
A Year of the Servo Reboot: Where Are We Now?Igalia
 
How to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerHow to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerThousandEyes
 
Connector Corner: Accelerate revenue generation using UiPath API-centric busi...
Connector Corner: Accelerate revenue generation using UiPath API-centric busi...Connector Corner: Accelerate revenue generation using UiPath API-centric busi...
Connector Corner: Accelerate revenue generation using UiPath API-centric busi...DianaGray10
 
Automating Google Workspace (GWS) & more with Apps Script
Automating Google Workspace (GWS) & more with Apps ScriptAutomating Google Workspace (GWS) & more with Apps Script
Automating Google Workspace (GWS) & more with Apps Scriptwesley chun
 
Real Time Object Detection Using Open CV
Real Time Object Detection Using Open CVReal Time Object Detection Using Open CV
Real Time Object Detection Using Open CVKhem
 
Scaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organizationScaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organizationRadu Cotescu
 
presentation ICT roal in 21st century education
presentation ICT roal in 21st century educationpresentation ICT roal in 21st century education
presentation ICT roal in 21st century educationjfdjdjcjdnsjd
 
Understanding Discord NSFW Servers A Guide for Responsible Users.pdf
Understanding Discord NSFW Servers A Guide for Responsible Users.pdfUnderstanding Discord NSFW Servers A Guide for Responsible Users.pdf
Understanding Discord NSFW Servers A Guide for Responsible Users.pdfUK Journal
 
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
04-2024-HHUG-Sales-and-Marketing-Alignment.pptxHampshireHUG
 
Apidays New York 2024 - The value of a flexible API Management solution for O...
Apidays New York 2024 - The value of a flexible API Management solution for O...Apidays New York 2024 - The value of a flexible API Management solution for O...
Apidays New York 2024 - The value of a flexible API Management solution for O...apidays
 
Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024The Digital Insurer
 

Recently uploaded (20)

Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
 
Artificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and MythsArtificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and Myths
 
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
 
GenAI Risks & Security Meetup 01052024.pdf
GenAI Risks & Security Meetup 01052024.pdfGenAI Risks & Security Meetup 01052024.pdf
GenAI Risks & Security Meetup 01052024.pdf
 
🐬 The future of MySQL is Postgres 🐘
🐬  The future of MySQL is Postgres   🐘🐬  The future of MySQL is Postgres   🐘
🐬 The future of MySQL is Postgres 🐘
 
Partners Life - Insurer Innovation Award 2024
Partners Life - Insurer Innovation Award 2024Partners Life - Insurer Innovation Award 2024
Partners Life - Insurer Innovation Award 2024
 
What Are The Drone Anti-jamming Systems Technology?
What Are The Drone Anti-jamming Systems Technology?What Are The Drone Anti-jamming Systems Technology?
What Are The Drone Anti-jamming Systems Technology?
 
Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024
 
A Year of the Servo Reboot: Where Are We Now?
A Year of the Servo Reboot: Where Are We Now?A Year of the Servo Reboot: Where Are We Now?
A Year of the Servo Reboot: Where Are We Now?
 
How to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerHow to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected Worker
 
Connector Corner: Accelerate revenue generation using UiPath API-centric busi...
Connector Corner: Accelerate revenue generation using UiPath API-centric busi...Connector Corner: Accelerate revenue generation using UiPath API-centric busi...
Connector Corner: Accelerate revenue generation using UiPath API-centric busi...
 
Automating Google Workspace (GWS) & more with Apps Script
Automating Google Workspace (GWS) & more with Apps ScriptAutomating Google Workspace (GWS) & more with Apps Script
Automating Google Workspace (GWS) & more with Apps Script
 
Real Time Object Detection Using Open CV
Real Time Object Detection Using Open CVReal Time Object Detection Using Open CV
Real Time Object Detection Using Open CV
 
Scaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organizationScaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organization
 
presentation ICT roal in 21st century education
presentation ICT roal in 21st century educationpresentation ICT roal in 21st century education
presentation ICT roal in 21st century education
 
Understanding Discord NSFW Servers A Guide for Responsible Users.pdf
Understanding Discord NSFW Servers A Guide for Responsible Users.pdfUnderstanding Discord NSFW Servers A Guide for Responsible Users.pdf
Understanding Discord NSFW Servers A Guide for Responsible Users.pdf
 
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
 
Apidays New York 2024 - The value of a flexible API Management solution for O...
Apidays New York 2024 - The value of a flexible API Management solution for O...Apidays New York 2024 - The value of a flexible API Management solution for O...
Apidays New York 2024 - The value of a flexible API Management solution for O...
 
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
 
Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024
 

A Study on Compositional Semantics of Words in Distributional Spaces

  • 1. A Study on Compositional Semantics of Words in Distributional Spaces Pierpaolo Basile, Annalina Caputo, Giovanni Semeraro Dept. of Computer Science University of Bari Aldo Moro {basilepp}@di.uniba.it 6th IEEE International Conference on Semantic Computing September 19-21, 2012 Palermo, Italy
  • 2. Background… You shall know a word by the company it keeps! Meaning of a word is determined by its usage 2
  • 3. …Background Distributional Semantic Models (DSMs) defined as: <T, C, R, W, M, d, S> – T: target elements (words) – C: context – R: relation between T and C – W: weighting schema – M: geometric space TxC – d: space reduction M -> M’ – S: similarity function defined in M’
  • 4. Motivations • One definition of context at a time – encode syntactic information in DSMs • Words are represented in isolation – syntactic role could be used as a glue to compose words It’s raining cats and dogs = My cats and dogs are in the rain
  • 5. Outline • Simple DSMs and simple operators • Syntactic dependencies in DSMs – Structured DSMs – Compositional operators • Evaluation and results • Final remarks
  • 6. SIMPLE DSMS AND SIMPLE OPERATORS
  • 7. Simple DSMs… Term-term co-occurrence matrix (TTM): each cell contains the co-occurrences between two terms within a prefixed distance dog cat computer animal mouse dog 0 4 0 2 1 cat 4 0 0 3 5 computer 0 0 0 0 3 animal 2 3 0 0 2 mouse 1 5 3 2 0
  • 8. …Simple DSMs Latent Semantic Analysis (LSA): relies on the Singular Value Decomposition (SVD) of the co- occurrence matrix Random Indexing (RI): based on the Random Projection Latent Semantic Analysis over Random Indexing (RILSA)
  • 9. Random Indexing • Create and assign a context vector to each context element (e.g. document, passage, term, …) • Term vector is the sum of the context vectors in which the term occurs – sometimes the context vector could be boosted by a score (e.g. term frequency, PMI, …) 9
  • 10. Context Vector 0 0 0 0 0 0 0 -1 0 0 0 0 1 0 0 -1 0 1 0 0 0 0 1 0 0 0 0 -1 • sparse • high dimensional • ternary {-1, 0, +1} • small number of randomly distributed non- zero elements 10
  • 11. Random Indexing (formal) B =A R n,k n,m m,k k << m B nearly preserves the distance between points (Johnson-Lindenstrauss lemma) dr = c× d RI is a locality-sensitive hashing method which approximate the cosine distance between vectors 11
  • 12. Random Indexing (example) CVjohn -> (0, 0, 0, 0, 0, 0, 1, 0, -1, 0) John eats a red apple CVeat -> (1, 0, 0, 0, -1, 0, 0 ,0 ,0 ,0) CVred -> (0, 0, 0, 1, 0, 0, 0, -1, 0, 0) TVapple= CVjohn+ CVeat+CVred=(1, 0, 0, 1, -1, 0, 1, -1, -1, 0) 12
  • 13. Latent Semantic Analysis over Random Indexing 1. Reduce the dimension of the co-occurrences matrix using RI 2. Perform LSA over RI (LSARI) – reduction of LSA computation time: RI matrix contains less dimensions than co-occurrences matrix
  • 14. Simple operators… Addition (+): pointwise sum of components Multiplication (∘): pointwise multiplication of components Addition and multiplication are commutative – do not take into account word order Complex structures represented summing or multiplying words which compose them
  • 15. …Simple operators Given two word vectors u and v – composition by sum p = u + v – composition by multiplication p = u ∘ v Can be applied to any sequence of words
  • 17. Syntactic dependencies… John eats a red apple. subject object John eats apple modifier red 17
  • 18. …Syntactic dependencies John eats a red apple. DEPENDENT HEAD subject object John eats apple modifier red 18
  • 19. Representing dependences Use filler/role binding approach to represent a dependency dep(u, v) rd⊚u + rh⊚v rd and rh are vectors which represent respectively the role of dependent and head ⊚ is a placeholder for a composition operator
  • 20. Representing dependences (example) obj(apple, eat) rd⊚apple + rh⊚eat role vectors
  • 21. Structured DSMs 1. Vector permutation in RI (PERM) to encode dependencies 2. Circular convolution (CONV) as filler/binding operator to represent syntactic dependencies in DSMs 3. LSA over PERM and CONV carries out two spaces: PERMLSA and CONVLSA
  • 22. Vector permutation in RI (PERM) Using permutation of elements in context vectors to encode dependencies – right rotation of n elements to encode dependents (permutation) – left rotation of n elements to encode heads (inverse permutation) 22
  • 23. PERM (method) Create and assign a context vector to each term Assign a rotation function Π+1 to the dependent and Π-1 to the head Each term is represented by a vector which is – the sum of the permuted vectors of all the dependent terms – the sum of the inverse permuted vectors of all the head terms – the sum of the no-permuted vectors of both dependent and head words 23
  • 24. PERM (example…) John -> (0, 0, 0, 0, 0, 0, 1, 0, -1, 0) eat -> (1, 0, 0, 0, -1, 0, 0 ,0 ,0 ,0) John eats a red apple red -> (0, 0, 0, 1, 0, 0, 0, -1, 0, 0) apple -> (1, 0, 0, 0, 0, 0, 0, -1, 0, 0) TVapple = Π+1(CVred) + Π-1(CVeat) + CVred + CVeat red apple eats 24
  • 25. PERM (…example) John -> (0, 0, 0, 0, 0, 0, 1, 0, -1, 0) eat -> (1, 0, 0, 0, -1, 0, 0 ,0 ,0 ,0) John eats a red apple red-> (0, 0, 0, 1, 0, 0, 0, -1, 0, 0) apple -> (1, 0, 0, 0, 0, 0, 0, -1, 0, 0) TVapple = Π+1(CVred) + Π-1(CVeat) + CVred + CVeat=… right shift left shift …=(0, 0, 0, 0, 1, 0, 0, 0, -1, 0) + (0, 0, 0, -1, 0, 0, 0, 0, 1) + + (0, 0, 0, 1, 0, 0, 0, -1, 0, 0) + (1, 0, 0, 0, -1, 0, 0 ,0 ,0 ,0) 25
  • 26. Convolution (CONV) Create and assign a context vector to each term Create two context vectors for head and dependent roles Each term is represented by a vector which is – the sum of the convolution between dependent terms and the dependent role vector – the sum of the convolution between head terms and the head role vector – the sum of the vectors of both dependent and head words
  • 27. Circular convolution operator Circular convolution U=<1, 1, -1, -1, 1> V=<1, -1, 1, -1, -1> p=u⊛v P=<-1, 3, -1, -1, -1> defined as: U1 U2 U3 U4 U5 n V1 1 1 -1 -1 1 p j u k v( j k ) ( n 1) P1 V2 -1 -1 1 1 -1 k 1 P2 V3 1 1 -1 -1 1 P3 V4 -1 -1 1 1 -1 P4 V5 -1 -1 1 1 -1 P5
  • 28. Circular convolution by FFTs Circular convolution is computed in O(n2) – using FFTs is computed in O(nlogn) Given f the discrete FFTs and f -1 its inverse – u ⊛v = f -1( f(u) ∘ f (v) )
  • 29. CONV (example) John -> (0, 0, 0, 0, 0, 0, 1, 0, -1, 0) eat -> (1, 0, 0, 0, -1, 0, 0 ,0 ,0 ,0) John eats a red apple red-> (0, 0, 0, 1, 0, 0, 0, -1, 0, 0) apple -> (1, 0, 0, 0, 0, 0, 0, -1, 0, 0) rd -> (0, 0, 1, 0, -1, 0, 0, 0, 0, 0) rh -> (0, -1, 1, 0, 0, 0, 0, 0, 0, 0) apple = eat +red + (rd⊛ red) + (rh ⊛ eat) Context vector for dependent role Context vector for head role 29
  • 30. Complex operators Based on filler/role binding taking into account syntactic role: rd ⊚ u + rh ⊚ v – u and v could be recursive structures Two vector operators to bind the role: – convolution (⊛) – tensor (⊗) – convolution (⊛+): exploits also the sum of term vectors rd ⊛ u + rh ⊛ v + v + u
  • 31. Complex operators (remarks) Existing operators – t1 ⊚ t2 ⊚ … ⊚ tn: does not take into account syntactic role – t1 ⊛ t2 is commutative – t1⊗t2 ⊗ … ⊗tn: tensor order depends on the phrase length • two phrases with different length are not comparable – t1 ⊗ r1 ⊗t2 ⊗r2 ⊗ … ⊗ tn ⊗rn : also depends on the sentence length
  • 32. System setup • Corpus – WaCkypedia EN based on a 2009 dump of Wikipedia – about 800 million tokens – dependency parse by MaltParser • DSMs – 500 vector dimension (LSA/RI/RILSA) – 1,000 vector dimension (PERM/CONV/PERMLSA/CONVLSA) – 50,000 most frequent words – co-occurrence distance: 4 32
  • 33. Evaluation • GEMS 2011 Shared Task for compositional semantics – list of two pairs of words combination (support offer) (help provide) 7 (old person) (right hand) 1 • rated by humans • 5,833 rates • 3 types involved: noun-noun (NN), adjective-noun (AN), verb-object (VO) – GOAL: compare the system performance against humans scores • Spearman correlation 33
  • 34. Results (simple spaces)… NN AN VO TTM TTM TTM RILSA RILSA RILSA LSA LSA LSA RI RI RI + .21 .36 .25 .42 .22 .35 .33 .41 .23 .31 .28 .31 ∘ .31 .15 .23 .22 .21 .20 .22 .18 .13 .10 .18 .21 ⊛ .21 .38 .26 .35 .20 .33 .31 .44 .15 .31 .24 .34 ⊛+ .21 .34 .28 .43 .23 .32 .31 .37 .20 .31 .25 .29 ⊗ .21 .38 .25 .39 .22 .38 .33 .43 .15 .34 .26 .32 human .49 .52 .55 Simple Semantic Spaces
  • 35. …Results (structured spaces) NN AN VO PERMLSA PERMLSA PERMLSA CONVLSA CONVLSA CONVLSA PERM PERM PERM CONV CONV CONV + .36 .39 .43 .42 .34 .39 .42 .45 .27 .23 .30 .31 ∘ .22 .17 .10 .13 .23 .27 .13 .15 .20 .15 .06 .14 ⊛ .31 .36 .37 .35 .39 .39 .45 .44 .28 .23 .27 .28 ⊛+ .30 .36 .40 .36 .38 .32 .48 .44 .27 .22 .30 .32 ⊗ .34 .37 .37 .40 .36 .40 .45 .45 .27 .24 .31 .32 human .49 .52 .55 Structured Semantic Spaces
  • 36. Final remarks • Best results are obtained when complex operators/spaces (or both) are involved • No best combination of operator/space exists – depend on the type of relation (NN, AN, VO) • Tensor product and convolution provide good results in spite of previous results – filler/role binding is effective • Future work – generate several rd and rh vectors for each kind of dependency – apply this approach to other direct graph-based representations
  • 37. Thank you for your attention! Questions? <basilepp@di.uniba.it>