4. <4>
what is
Information
Meaning could be the
quality of a certain signal.
Meaning could be a
logical abstractor = a release mechanism.
(96dpi)
meaning
5. <5>
meaning
is social
Network effects make a network of
shared understandings more
valuable with growing size:
allowing e.g. ‘distributed cognition’.
To develop a shared understanding
is a natural and necessary process,
because language underspecifies
meaning: future understanding
builds on it
At same time: linguistic relativity
(Sapir-Whorf hypothesis): our own
language culture restricts our thinking
7. <7>
Concepts & Competence
things we can (not) construct from language
Tying shoelaces
Douglas Adams’
‘meaning of liff’:
Epping: The futile
movements of forefingers
and eyebrows used when
failing to attract the attention
of waiters and barmen.
Shoeburyness: The vague
uncomfortable feeling you
get when sitting on a seat
which is still warm from
somebody else's bottom
I have been
convincingly
Sapir-Whorfed
by this book.
10. <10>
Latent Semantic Analysis Two-Mode factor
analysis of the co-
occurrences in the
terminology
Results in a latent-
semantic vector
space
“Humans learn word meanings and how to
combine them into passage meaning
through experience with ~paragraph
unitized verbal environments.”
“They don’t remember all the separate
words of a passage; they remember
its overall gist or meaning.”
“LSA learns by ‘reading’ ~paragraph unitized
texts that represent the environment.”
“It doesn’t remember all the separate words
of a text it; it remembers
its overall gist or meaning.”
-- Landauer, 2007
14. <14>
Example: Classic Landauer
{ M } =
Deerwester, Dumais, Furnas, Landauer, and Harshman (1990):
Indexing by Latent Semantic Analysis, In: Journal of the American
Society for Information Science, 41(6):391-407
Only the red terms appear in more
than one document, so strip the rest.
term = feature
vocabulary = ordered set of features
TEXTMATRIX
16. <16>
doc2doc - similarities
Unreduced = pure vector
space model
- Based on M = TSD’
- Pearson Correlation
over document vectors
reduced
- based on M2 = TS2D’
- Pearson Correlation
over document vectors
18. <18>
Social Network Analysis
Existing for a long time (term coined 1954)
Basic idea:
Actors and Relationships between them (e.g.
Interactions)
Actors can be people (groups, media, tags, …)
Actors and Ties form a Graph (edges and nodes)
Within that graph, certain structures can be
investigated
• Betweenness, Degree of Centrality, Density, Cohesion
• Structural Patterns can be identified (e.g. the Troll)
21. <21>
Measuring Techniques (Sample)
Degree Centrality
number of (in/out) connections to others
Closeness
how close to all others
Betweenness
how often intermediary
Components
e.g. kmeans cluster (k=3)
25. <25>
Meaningful Interaction
Analysis (MIA)
Combines latent semantics
with the means of network analysis
Allows for investigating associative
closeness structures at the same time
as social relations
In latent-semantic spaces only
or in spaces with additional
and different (!) relations
27. <27>
Contextualised Doc & Term Vectors
Tk = left-hand sided matrix
= ‚term loadings‘ on the singular value
Dk = right-hand sided matrix
= ‚document loadings‘ on the singular value
Multiply them into same space
VT = Tk Sk
VD = Dk
T
Sk
Cosine Closeness Matrix over
... = adjacency matrix
= a graph
More: e.g. add author vectors VA
through cluster centroids or
vector addition of their publication vectors
latent-semantic space
DT VV ∪
ADT VVV ∪∪
Speed: use existing space and
fold in e.g. author vectors
43. <43>
Conclusion
Both LSA and SNA alone are not
sufficient for a modern
representation theory
MIA provides one possible
bridge between them
It is a powerful technique
And it is simple to use (in R)