Latent Semantic Analysis (LSA) is a mathematical technique for computationally modeling the meaning of words and larger units of texts. LSA works by applying a mathematical technique called Singular Value Decomposition (SVD) to a term*document matrix containing frequency counts for all words found in the corpus in all of the documents or passages in the corpus. After this SVD application, the meaning of a word is represented as a vector in a multidimensional semantic space, which makes it possible to compare word meanings, for instance by computing the cosine between two word vectors.
LSA has been successfully used in a large variety of language related applications from automatic grading of student essays to predicting click trails in website navigation. In Coh-Metrix (Graesser et al. 2004), a computational tool that produces indices of the linguistic and discourse representations of a text, LSA was used as a measure of text cohesion by assuming that cohesion increases as a function of higher cosine scores between adjacent sentences.
Besides being interesting as a technique for building programs that need to deal with semantics, LSA is also interesting as a model of human cognition. LSA can match human performance on word association tasks and vocabulary test. In this talk, Fridolin will focus on LSA as a tool in modeling language acquisition. After framing the area of the talk with sketching the key concepts learning, information, and competence acquisition, and after outlining presuppositions, an introduction into meaningful interaction analysis (MIA) is given. MIA is a means to inspect learning with the support of language analysis that is geometrical in nature. MIA is a fusion of latent semantic analysis (LSA) combined with network analysis (NA/SNA). LSA, NA/SNA, and MIA are illustrated by several examples.
5. Information (96dpi) Information could be the quality of a certain signal. Information could be a logical abstractor, the release mechanism. Information & Knowledge Knowledge could be the delta at the receiver (a paper, a human, a library).
6. Learning is change Learning is about competence development Competence becomes visible in performance Professional competence is mainly about (re-)constructing and processing information and knowledge from cues Professional competence development is much about learning concepts from language Professional performance is much about demonstrating conceptual knowledge with language Language! What is learning about?
7. Tying shoelaces Douglas Adams’ ‘meaning of liff’: Epping: The futile movements of forefingers and eyebrows used when failing to attract the attention of waiters and barmen. Shoeburyness: The vague uncomfortable feeling you get when sitting on a seat which is still warm from somebody else's bottom I have been convincingly Sapir-Whorfed by this book. Non-textual concepts things we can’t (easily) learn from language
9. Word Choice Educated adult understands ~100,000 word forms An average sentence contains 20 tokens. Thus 100,00020 possible combinationsof words in a sentence maximum of log2 100,00020= 332 bits in word choice alone. 20! = 2.4 x 1018 possible orders of 20 words = maximum of 61 bits from order of the words. 332/(61+ 332) = 84%word choice (Landauer, 2007)
10. Latent Semantic Analysis “Humans learn word meanings and how to combine them into passage meaning through experience with ~paragraph unitized verbal environments.” “They don’t remember all the separate words of a passage; they remember its overall gist or meaning.” “LSA learns by ‘reading’ ~paragraph unitized texts that represent the environment.” “It doesn’t remember all the separate words of a text it; it remembers its overall gist or meaning.” (Landauer, 2007)
11. Latent Semantics latent-semantic space In other words:Assumption: language utterances have a semantic structure Problem: structure is obscured by word usage(noise, synonymy, polysemy, …) Solution: map doc-term matrix using conceptual indices derived statistically (truncated SVD) and make similarity comparisons using angles
12. Input (e.g., documents) term = feature vocabulary = ordered set of features Only the red terms appear in more than one document, so strip the rest. TEXTMATRIX { M } = Deerwester, Dumais, Furnas, Landauer, and Harshman (1990): Indexing by Latent Semantic Analysis, In: Journal of the American Society for Information Science, 41(6):391-407
17. Similarity in a Latent-Semantic Space Query Y dimension Target 1 Angle 1 Angle 2 Target 2 X dimension (Landauer, 2007)
18. doc2doc - similarities Unreduced = pure vector space model - Based on M = TSD’ - Pearson Correlation over document vectors reduced - based on M2 = TS2D’ - Pearson Correlation over document vectors
21. b) SVD is computationally expensive From seconds (lower hundreds of documents, optimised linear algebra libraries, truncated SVD) To minutes (hundreds to thousands of documents) To hours (tens and hundreds of thousands) a) SVD factor stability SVD calculates factorsover a given text base; different texts – different factors Problem: avoid unwanted factor changes Solution: folding-in of instead of recalculating Projecting by Folding-In
22. 2 1 vT Folding-In in Detail (cf. Berry et al., 1995) Mk (2) convert „Dk“-format vector to „Mk“-format Tk Sk Dk (1) convert Original Vector to „Dk“-format
23. The Value of Singular Values Pearson(jahr, wien) Pearson(eu, österreich)
31. Social Network Analysis Existing for a long time (term coined 1954) Basic idea: Actors and Relationships between them (e.g. Interactions) Actors can be people (groups, media, tags, …) Actors and Ties form a Graph (edges and nodes) Within that graph, certain structures can be investigated Betweenness, Degree of Centrality, Density, Cohesion Structural Patterns can be identified (e.g. the Troll)
36. Measuring Techniques (Sample) Closenesshow close to all others Degree Centralitynumber of (in/out) connections to others Betweennesshow often intermediary Componentse.g. kmeans cluster (k=3)
39. Paper Collaboration Prolearn e.g. co-authorships of ~30 deliverables of three work packages (ProLearn NoE) Roles: reviewer (red), editor (green), contributor Size: Prestige() But: type of interaction? Content of interaction? => not possible!
43. Meaningful Interaction Analysis (MIA) Fusion: Combining LSA with SNA Terms and Documents (or anything else represented with column vectors or row vectors) are mapped into same space by LSA Semantic proximity can be measured between them: how close is a term to a document? (S)NA allows to analyse these resulting graph structures By e.g. cluster or component analysis By e.g. identifying central descriptors for these
45. Truncated SVD latent-semantic space … we will get a different matrix (different values, but still of the same format as M).
46. Knowledge Proxy: LSA Part Tk= left-hand sided matrix = ‚term loadings‘ on the singular value Dk= right-hand sided matrix = ‚document loadings‘ on the singular value Multiply them into same space VT = TkSk VD = DkTSk Cosine Distance Matrix over ... = a graph Extension: add author vectors VAthrough cluster centroids or vector addition of their publication vectors latent-semantic space Ofcourse:useexistingspaceandfold inthewholesetsofvectors
47. Knowledge Proxy: SNA Part:Filter the Network Every vectorhas a cosinedistancetoeveryother (maybe negative)! So: filter forthedesiredsimilaritystrength
51. Spot unwanted fragmentation e.g. two authors work on the same topic, but with different collaborator groups and with different literature Intervention Instrument: automatically recommend to hold a flashmeeting Bringing together what belongs together Wild, Ochoa, Heinze, Crespo, Quick (2009, to appear)