SEMI-SUPERVISED CLASSIFICATION FOR
NATURAL LANGUAGE PROCESSING
PRESENTATION AT A GLANCE
•
–
–
–
–

•
–
–
–
–

•
2
SEMI-SUPERVISED LEARNING
•
–

•
•

3
SEMI-SUPERVISED LEARNING PROBLEMS

(1)
Learn from labeled data

Inductive
Learning

(2)
Apply learning on
unlabeled data to label
them
Transductive
Learning

(4)
Apply learning on
unseen unlabeled data

(3)
If confident in labeling,
then learn from
(1) and (2)

4
SEMI-SUPERVISED LEARNING PROBLEMS
•
–

•
–

5
SCOPES OF SEMI-SUPERVISED LEARNING
•
–
–
–

6
HOW DOES SEMI-SUPERVISED CLASSIFICATION WORK?

7
TYPES OF SEMI-SUPERVISED LEARNING
•
•
•
•
•

8
GENERATIVE VS DISCRIMINATIVE MODELS
(x,y)

Discriminative Models

Generative Models

9
GENERATIVE VS DISCRIMINATIVE MODELS
•
•

•

10
GENERATIVE VS DISCRIMINATIVE MODELS
•

•
•

11
GENERATIVE VS DISCRIMINATIVE MODELS
Conditional Probability,
to determine class
boundaries

Transductive SVM,
Graph-based
methods

Joint Probability P(x,y),
for any given y, we can
generate its x

EM Algorithm,
Self-learning

Cannot be used without considering P(x)
Difficult because P(x|y) are inadequate

12
GENERATIVE VS DISCRIMINATIVE MODELS

•
•

•

•
•

•

13
IS THERE A FREE LUNCH?
•
–

•

14
IS THERE A FREE LUNCH?
•
•

•

15
IS THERE A FREE LUNCH?

•
–

•

16
SELF-TRAINING

17
CO-TRAINING
•
•

18
CO-TRAINING

19
CO-TRAINING
•
•

•
•
•
20
CO-TRAINING
•
•
•

•

21
CO-TRAINING: COVEATS

22
ACTIVE LEARNING

23
WHICH METHOD SHOULD I USE?
•
–

•
–

•
–

•
–
24
WHICH METHOD SHOULD I USE?
•
–

•
–

25
SEMI-SUPERVISED CLASSIFICATION FOR NLP
•
•
•
•

26
EFFECTIVE SELF-TRAINING
FOR PARSING

27
INTRODUCTION
•
•

–

28
METHODS
•
•
•
–

29
DATASETS
•
–

•
•
•
–

30
RESULTS
•

–

•
–
–

31
LIMITATIONS
•
•
•
–
32
SEMI-SUPERVISED SPAM FILTERING:
DOES IT WORK?

33
INTRODUCTION
•

•

34
BACKGROUND
•
–
•
•
•
•
•

–
•
•
•
35
BACKGROUND
•
•
•
•

36
BACKGROUND
•
–
–

37
METHODS AND MATERIALS
•
–
•
•
•

–
•
•

38
RESULTS: DELAYED FEEDBACK VS CROSS-USER

Delayed Feedback

Cross-User

39
RESULTS: CROSS-CORPUS
•
•

40
EXTRACTIVE SUMMARIZATION USING
SUPERVISED AND SEMI-SUPERVISED
LEARNING

41
INTRODUCTION
•
•

42
METHOD
•

•
–

•
–
–
43
DATASETS
•
•
–

•
•
–

•
–
44
RESULTS: FEATURE SELECTION
•

Human Summary ROUGE I
Score was 0.422

45
RESULTS: EFFECT OF UNLABELED DATA

More labeled data
produced better Fscore

46
RESULTS: SUPERVISED VS SEMI-SUPERVISED

47
RESULTS: EFFECT OF SUMMARY LENGTH

48
LIMITATIONS
•
–

•
–

49
SEMI-SUPERVISED CLASSIFICATION FOR
EXTRACTING PROTEIN INTERACTION SENTENCES
USING DEPENDENCY PARSING

50
INTRODUCTION
•
•
•

•
•

51
INTRODUCTION
•
•

52
METHOD
•

•

53
DATASETS
•

–
–

54
RESULTS: AIMED DATASET

55
RESULTS: CB DATASET

56
RESULTS: EFFECT OF TRAINING DATA SIZE
(AIMED)
•

•

57
RESULTS: EFFECT OF TRAINING DATA SIZE
(CB)
•
•

58
LIMITATIONS
•
–

•
–

59
HOW MUCH UNLABELED DATA IS USED?

60
CONCLUSIONS
•
•
–
–
–

61
CONCLUSIONS
•
–
–
–

–

62
CONCLUSIONS

63

Semi-supervised classification for natural language processing