14. Michael Oakes (UoW) Natural Language Processing for Translation
Upcoming SlideShare
Loading in...5
×
 

14. Michael Oakes (UoW) Natural Language Processing for Translation

on

  • 395 views

 

Statistics

Views

Total Views
395
Views on SlideShare
226
Embed Views
169

Actions

Likes
0
Downloads
6
Comments
0

3 Embeds 169

http://expert-itn.eu 153
http://localhost 15
http://www.expert-itn.eu 1

Accessibility

Categories

Upload Details

Uploaded via as Adobe PDF

Usage Rights

© All Rights Reserved

Report content

Flagged as inappropriate Flag as inappropriate
Flag as inappropriate

Select your reason for flagging this presentation as inappropriate.

Cancel
  • Full Name Full Name Comment goes here.
    Are you sure you want to
    Your message goes here
    Processing…
Post Comment
Edit your comment

14. Michael Oakes (UoW) Natural Language Processing for Translation 14. Michael Oakes (UoW) Natural Language Processing for Translation Presentation Transcript

  • Information Retrieval (from EBMT examples) Michael P. Oakes University of Wolverhampton Birmingham Winter School, 2013
  • Finding Out About (FOA)  “Finding Out About”, by Richard Belew (2000), Cambridge University Press.  Finding Out About (FOA) - research activities that allow a decision maker to draw on others’ knowledge, especially the WWW = “Information Retrieval”.  A library (or WWW) contains many books (“documents”) on many topics. The authors are typically people far from our time or place, using language similar but not identical to our own. We must FOA a topic of special interest, looking only for those things which are relevant to our search. This basic skill is a fundamental part of an academic’s job:
  • FOA has 3 phases:  asking a question  constructing an answer  assessing an answer
  • 1. Asking a question  Users of a search engine may be aware of a specific gap in their knowledge, and are motivated to fill it (meta-cognition about ignorance). They may not be able to articulate their knowledge gap. Forming a clearly posed question is the hardest part of answering it! This common cognitive state is the user’s information need. User’s try to take their ill-defined, internal cognitive state and turn it into an external expression of their question. This external expression is called the query, and the language in which it is constructed the query language.
  • 2. Constructing an answer  A human question-answerer might consider:  can they translate the user’s ill-formed query into a better one?  do they know the answer themselves?  are they able to verbalise this answer in terms the user will understand?  can they provide the necessary background knowledge for the user to understand the answer itself?  Current search engines are slightly more limited in scope. The search engine has available to it only a preexisting set of “canned” texts (although this may be very large), and its response is limited to identifying one or more of these passages and presenting them to the users.
  • 3. Assessing the answer  Assessing the answer  We would normally give feedback to a human answerer, e.g. “That isn’t what I meant”, “Let me ask it another way”, “That helps, but I still have this problem” or “What does that have to do with anything?”. So we “close the loop” when the user provides an assessment of how relevant the found the answer provided. In an automatic system this is relevance feedback - the user reacts to each retrieved document as “relevant”, “not relevant” or “neutral”.  See fig 1.4. The three steps in a computerised, algorithmic context, information retrieval.
  • Keywords  The fundamental operation performed by a search engine is a match between descriptive features mentioned by users in their queries, and documents sharing those features. By far the most important kind of features are keywords.  Keywords are linguistic atoms - typically words, pieces of words, or phrases - used to characterise the subject or content of the document.  They are pivotal because they must bridge the gap between the users’ characterisation of their information need (queries) and the characterisation of the documents’ topical focus against which these will be matched.  Contrast natural language queries with bag-of-words queries.
  • Keywords as document descriptors  Keywords are also used as document descriptors.  Indexing is the process of associating one or more keywords with each document.  The vocabulary used can either be controlled or uncontrolled (also known as closed or open). If we organise a conference, and ask the authors of each paper to index it manually using only terms on a fixed list of potential keywords, this is a closed vocabulary.
  • Query syntax  Query syntax. A typical search engine query consists of 2 to 3 words.  Queries defined only as sets of keywords are simple queries - most search engines use this “bag of words” approach.  Other possibilities exist e.g. Boolean operators and / or / not e.g. “neural networks AND speech recognition”.  Verb(subject, object) triples e.g. “aspirin treats blood_clotting”
  • Document length  Document length. Longer documents can discuss more topics, and hence be associated with more keywords, and thus are more likely to be retrieved.  This means we must normalise documents’ indices in some way to compensate for differing lengths.  We also assume that the smallest unit of text with appreciable “aboutness” is the paragraph, and larger documents are constructed of a number of paragraphs.
  • Stemming  Stemming aims to remove surface markings (such as number) to reveal a root form  Using a token’s root form as an index term can give robust retrieval even when the query contains the plural CARS while the document contains the singular CAR  Linguists distinguish inflectional morphology (plurals, third person singular, past tense, -ing) from derivational morphology (e.g. teach (verb), teacher (noun)). Weak vs. strong stemming.
  • Example stemming rules  (.*)SSES  /1SS: PERL-like syntax to say that strings ending in –SSES should be transformed by taking the stem (characters before –SSES) and adding only the two characters –SS.  (.*)IES  /1Y  A complete stemmer contains many such rules (60 in Lovins’ set), and a regime for handling conflicts when multiple rules match the same token, e.g. longest match, rule order.
  • Pros and Cons of Stemming  Reduces the size of the keyword vocabulary, allowing compression of the index files of 10 – 50%.  Increases recall – a query on FOOTBALL now also finds documents on FOOTBALLER(S), FOOTBALLING.  Reduces precision – stripping away morphological features may obscure differences in word meanings. For example, GRAVITY has two senses (earth’s pull, seriousness). GRAVITATION can only refer to earth’s pull – but if we stem it to GRAVITY, it could mean either.
  • Calculating TF-IDF weighting 
  • The Vector Space Model
  • The cosine similarity measure 
  • How well are we doing (1)?  Evaluation of search engines is notoriously difficult. However, we have two measures of objective assessment. The first step is to focus on a particular query.  We identify the set of documents Rel that are determined to be relevant to it (subjective!).  A perfect search engine would retrieve all and only the documents in Rel.  See fig. 1.10
  • Recall  Clearly, the number of documents that were designated both relevant and retrieved, Retr ∩ Rel will be a key measure of success.  But we must compare the size of the set |Retr ∩ Rel| to something.  If we were very concerned that the search engine retrieve every relevant document (e.g. every prior ruling relevant to a judicial case) , we should compare the intersection to the number of documents marked as relevant |Rel|.  This measure is known as recall = |Retr ∩ Rel| / |Rel| :
  • Precision  However, we might instead be worried about how much of what we see is relevant (search engine users want a lot of relevant hits on the first page), so an equally reasonable standard of comparison is precision, the proportion of retrieved documents which are in fact relevant:  P = |Retr ∩ Rel| |Retr|
  • Retrieved versus Relevant Docs High Recall Retrieval Retrieved Relevant High Precision Retrieval
  • Conclusion: IR for short documents like translation examples  Prior annotation of the corpus of past translations (Clifton and Teahan, 2004, QA systems)  Similarity measures for short segments of text: stemming to increase recall, then document expansion, Kullback-Leibler Divergence as a similarity measure (Metzler et al., 2007).  Tao Tao et al (2006). Need for short text matching: query/image caption, sponsored search: query/ad keyword; query reformulation: query / query similarity  Document expansion. Rohit Gupta expands the documents with all entries in the PPDB Paraphrase database: lexical, phrasal and syntactic.  Relevance feedback? (“more like this”).