SlideShare a Scribd company logo
1 of 27
Download to read offline
Stumping Along a Summary
Exploration & Exploitation Challenge 2011
Christophe Salperwyck1
Tanguy Urvoy2
1
Orange Labs / Lille University
2
Orange Labs
ICML 2011
Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 1 / 27
Outline
1 Problem Statement
2 Approach
Online Summary
Base Predictors
Online Combination of Predictors
3 Experiments and Results
Adobe Dataset (Challenge)
Orange Portal Dataset
4 Conclusion and Discussion
Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 2 / 27
Problem Statement Online content selection
Online content selection problem
General problem statement
a web site offers different items (news, ads, books. . . )
in order to maximize the number of clicks:
select for each visitor the item she is most likely to click on
instance described by visitor profile features + item features + mixed features
gps position, time, navigation history, query. . .
item id, topic, creation time, content keywords. . .
cosine similarity, gps distance. . .
prediction/learning/selection must be done online:
limited time
limited memory
rare and noisy events
Our formalization: only order matters
online (bipartite) ranking with partial feedback
ranking + scarcity → optimize AUC rather than accuracy
partial feedback → risk of being stuck in suboptimal predictive models
Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 3 / 27
Problem Statement Online content selection
Off-line evaluation of online content selection systems
online evaluation: costly
rejection sampling from logs: unbiased estimator, but too many rejections
This challenge simulation setting
choose both profile and item from a batch of 6 consecutive log instances
less realistic than rejection sampling ⇒ no rejection
maybe enough for comparison of algorithms
From our own experiments with Orange portal logs
some profile features are very discriminant
e.g. pageid: orange home page click rate = 5 × other pages
only learn to select home page vs. other pages
what we did to reduce this bias: take logs from a single page (web-mail)
maybe better: select batch of instances with “similar” visitor profiles
same time, same visited pages. . .
close gps positions etc.
Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 4 / 27
Approach Online Summary
Outline
1 Problem Statement
2 Approach
Online Summary
Base Predictors
Online Combination of Predictors
3 Experiments and Results
Adobe Dataset (Challenge)
Orange Portal Dataset
4 Conclusion and Discussion
Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 5 / 27
Approach Online Summary
Motivation for a Summary
Challenge constraints
limited time: 100ms/round for predicting and learning
limited memory: 1.7GB of memory
keep Garbage Collector quiet
We wanted a summary with small memory footprint
fast updates
fast queries
Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 6 / 27
Approach Online Summary
Numerical features summary
Assuming a static distribution, we want to answer questions like:
what are the click/¬click counts for feature xi ∈]a, b]?
GK merge-prune algorithm: online equal-freq histogram
Greenwald M., Khanna S. Space-efficient online computation of quantile summaries. SIGMOD’01
this algorithm maintains an array of tuples < vi , gi , ∆i > where:
vi is a value seen in the stream
gi is the number of values seen between vi−1 and vi
∆i is the maximum possible error on gi
properties:
equal-freq, online, bounded memory
insensitive to data order and/or data distribution
strong error guarantees ( -approximate quantile summary)
Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 7 / 27
Approach Online Summary
Nominal features summary
Problem: the number of nominal values is not supposed to be bounded
for instance: client ids, cookies or city names
we do not need to keep rare values: e.g. outdated cookies or small cities
Solution: hashing
hash nominal values into a fixed number of buckets
count click/¬click for each bucket
may be done with several hashing functions to reduce error:
Cormode G., Muthukrishnan S. An improved data stream summary: the count-min sketch
and its applications. Journal of Algorithms. 2005.
Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 8 / 27
Approach Base Predictors
Outline
1 Problem Statement
2 Approach
Online Summary
Base Predictors
Online Combination of Predictors
3 Experiments and Results
Adobe Dataset (Challenge)
Orange Portal Dataset
4 Conclusion and Discussion
Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 9 / 27
Approach Base Predictors
Base Predictors: probability estimation trees
Output: click-rate estimate
Split criterion: Gini impurity
Numerical features:
simple decision stumps (depth=1)
estimate =
left estimate if xi ≤ θ
right estimate else
deeper univariate trees
gradual unfolding: add new leaves with data arrival
Nominal features:
hash-stump: hash and treat as continuous
best vs. all: best value/others
Univariate trees: work with a single feature
global summary → full-rebuild from summary at each step
Multivariate trees: combine several features
one summary per leaf → unfolding is a definitive action
Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 10 / 27
Approach Online Combination of Predictors
Outline
1 Problem Statement
2 Approach
Online Summary
Base Predictors
Online Combination of Predictors
3 Experiments and Results
Adobe Dataset (Challenge)
Orange Portal Dataset
4 Conclusion and Discussion
Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 11 / 27
Approach Online Combination of Predictors
Online combination of predictors
Output linear combination of predictions
Intuition: whack bad predictors
exponential weighting
online linear ranker
set weight to optimize bipartite ranking loss (i.e. AUC)
Simple averaging: optimizes nothing
Online bagging
Oza N.C.,Russell S. Online Bagging and Boosting. ICONS 2001.
Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 12 / 27
Experiments and Results
Experiments
Adobe dataset (phase 1 of challenge)
3M instances (i.e. 500K batches of 6)
120 features: 99 continuous / 21 nominals (last nominal feature is ”option id”)
click rate: ∼ 0.0024
Orange portal web-mail dataset
6M instances (i.e. 1M batches of 6)
33 nominal features
click rate: ∼ 0.0005
Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 13 / 27
Experiments and Results Adobe Dataset (Challenge)
Outline
1 Problem Statement
2 Approach
Online Summary
Base Predictors
Online Combination of Predictors
3 Experiments and Results
Adobe Dataset (Challenge)
Orange Portal Dataset
4 Conclusion and Discussion
Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 14 / 27
Experiments and Results Adobe Dataset (Challenge)
Adobe dataset with perfect player
Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 15 / 27
Experiments and Results Adobe Dataset (Challenge)
Adobe dataset without perfect player
Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 16 / 27
Experiments and Results Adobe Dataset (Challenge)
Deviation from static model
Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 17 / 27
Experiments and Results Adobe Dataset (Challenge)
Utree vs. full dataset discretization (MODL)
Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 18 / 27
Experiments and Results Orange Portal Dataset
Outline
1 Problem Statement
2 Approach
Online Summary
Base Predictors
Online Combination of Predictors
3 Experiments and Results
Adobe Dataset (Challenge)
Orange Portal Dataset
4 Conclusion and Discussion
Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 19 / 27
Experiments and Results Orange Portal Dataset
Orange portal dataset with perfect player
Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 20 / 27
Experiments and Results Orange Portal Dataset
Orange dataset with perfect player
Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 21 / 27
Conclusion and Discussion
Conclusion and discussion
Thank you for this challenge!
Experiment summary
interesting simulation scheme: how many items/batches ?
combination of GK merge-prune and hashing provides a good stream summary to
rely on for click prediction
surprisingly (unfortunately?) simple decision stumps seem to be be competitive
base predictors
Why are decision stumps so good?
maybe because of their stability?
maybe stump-induced partial order lets some room for exploration?
Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 22 / 27
Appendix
Experiment summary (for Adobe dataset)
GK-100: Greenwald & Khanna (100 tuples), CMS-1: Count min-sketch (1 hash)
1R: single rules xi ∈]a, b] ? 1 : 0
HStump: hash then split like stump
RF: random forests, UTree: univariate tree
AVG: averaging, OZA: Oza bagging (instance weighting)
multivariate predictors typed in red
name summary base-predictors combi. score
Random 1127 ± 41
LinRank GK-4/CMS-1 1R (proba. output) LRank-120 ∼ 1600
LinRank GK-4/CMS-1 1R (0/1 output) LRank-120 ∼ 1800
RF GK-100/CMS-1 20 trees (10 feat.) AVG-20 ∼ 1900
Stumps1VA GK-100/CMS-1 Stump/1 vs. all AVG-120 ∼ 2000
UTree GK-100/CMS-1 UTree/HStump AVG-120 ∼ 2030
Cnrs ∼ 2040
Simple stumps GK-100/CMS-1 Stump/HStump AVG-120 ∼ 2050
Best GK-100/CMS-1 800 stumps + 200 RF OZA-1K ∼ 2080
Inria ∼ 2200
Perfect 6678
Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 23 / 27
Appendix
Adobe dataset (start zoom)
Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 24 / 27
Appendix
Adobe dataset (end zoom)
Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 25 / 27
Appendix
Adobe dataset (final zoom)
Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 26 / 27
Appendix
Utree learning
Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 27 / 27

More Related Content

Similar to explorationexploitation2011_salperwyck_urvoy_contr_01

Sarah Callaghan Research Data Overview
Sarah Callaghan Research Data OverviewSarah Callaghan Research Data Overview
Sarah Callaghan Research Data Overview
OpenAIRE
 
Jos van Sas - Testimonial Alcatel-Lucent Bell Labs
Jos van Sas - Testimonial Alcatel-Lucent Bell LabsJos van Sas - Testimonial Alcatel-Lucent Bell Labs
Jos van Sas - Testimonial Alcatel-Lucent Bell Labs
imec.archive
 
UNEP-Live prototype 20110821
UNEP-Live prototype 20110821UNEP-Live prototype 20110821
UNEP-Live prototype 20110821
Mick Wilson
 

Similar to explorationexploitation2011_salperwyck_urvoy_contr_01 (20)

The Other HPC: High Productivity Computing
The Other HPC: High Productivity ComputingThe Other HPC: High Productivity Computing
The Other HPC: High Productivity Computing
 
Presentation on "Recommenders in Social Learning Platforms" at #iknow2014 & #...
Presentation on "Recommenders in Social Learning Platforms" at #iknow2014 & #...Presentation on "Recommenders in Social Learning Platforms" at #iknow2014 & #...
Presentation on "Recommenders in Social Learning Platforms" at #iknow2014 & #...
 
Building a Graph Database in Neo4j with Spark & Spark SQL to gain new insight...
Building a Graph Database in Neo4j with Spark & Spark SQL to gain new insight...Building a Graph Database in Neo4j with Spark & Spark SQL to gain new insight...
Building a Graph Database in Neo4j with Spark & Spark SQL to gain new insight...
 
Barcamp Bangalore 2 - On User Experience and Usability Testing
Barcamp Bangalore 2 - On User Experience and Usability TestingBarcamp Bangalore 2 - On User Experience and Usability Testing
Barcamp Bangalore 2 - On User Experience and Usability Testing
 
Troubleshooting Deep Neural Networks - Full Stack Deep Learning
Troubleshooting Deep Neural Networks - Full Stack Deep LearningTroubleshooting Deep Neural Networks - Full Stack Deep Learning
Troubleshooting Deep Neural Networks - Full Stack Deep Learning
 
On Specifying and Sharing Scientific Workflow Optimization Results Using Rese...
On Specifying and Sharing Scientific Workflow Optimization Results Using Rese...On Specifying and Sharing Scientific Workflow Optimization Results Using Rese...
On Specifying and Sharing Scientific Workflow Optimization Results Using Rese...
 
BOM DMAIC TEMPLATE
BOM DMAIC TEMPLATEBOM DMAIC TEMPLATE
BOM DMAIC TEMPLATE
 
2_ResearchDataOverview_SarahCallaghan
2_ResearchDataOverview_SarahCallaghan2_ResearchDataOverview_SarahCallaghan
2_ResearchDataOverview_SarahCallaghan
 
Sarah Callaghan Research Data Overview
Sarah Callaghan Research Data OverviewSarah Callaghan Research Data Overview
Sarah Callaghan Research Data Overview
 
Jos van Sas - Testimonial Alcatel-Lucent Bell Labs
Jos van Sas - Testimonial Alcatel-Lucent Bell LabsJos van Sas - Testimonial Alcatel-Lucent Bell Labs
Jos van Sas - Testimonial Alcatel-Lucent Bell Labs
 
Exploration & Exploitation Challenge 2011
Exploration & Exploitation Challenge 2011Exploration & Exploitation Challenge 2011
Exploration & Exploitation Challenge 2011
 
EOSC-hub and OpenAIRE Advance webinar - introduction
EOSC-hub and OpenAIRE Advance webinar - introductionEOSC-hub and OpenAIRE Advance webinar - introduction
EOSC-hub and OpenAIRE Advance webinar - introduction
 
Mark Hughes Annual Seminar Presentation on Open Source
Mark Hughes Annual Seminar Presentation on Open Source Mark Hughes Annual Seminar Presentation on Open Source
Mark Hughes Annual Seminar Presentation on Open Source
 
Orfeo ToolBox workshop at FOSS4G Europe 2015
Orfeo ToolBox workshop at FOSS4G Europe 2015Orfeo ToolBox workshop at FOSS4G Europe 2015
Orfeo ToolBox workshop at FOSS4G Europe 2015
 
Enterprise class deployment for GeoServer and GeoWebcache Optimizing perform...
Enterprise class deployment  for GeoServer and GeoWebcache Optimizing perform...Enterprise class deployment  for GeoServer and GeoWebcache Optimizing perform...
Enterprise class deployment for GeoServer and GeoWebcache Optimizing perform...
 
Detecting Food and Activities in Lifelogging Images
Detecting Food and Activities in Lifelogging ImagesDetecting Food and Activities in Lifelogging Images
Detecting Food and Activities in Lifelogging Images
 
[3.6] Beyond Data Sharing - Pieter van Gorp [3TU.Datacentrum Symposium 2014, ...
[3.6] Beyond Data Sharing - Pieter van Gorp [3TU.Datacentrum Symposium 2014, ...[3.6] Beyond Data Sharing - Pieter van Gorp [3TU.Datacentrum Symposium 2014, ...
[3.6] Beyond Data Sharing - Pieter van Gorp [3TU.Datacentrum Symposium 2014, ...
 
Woop - Workflow Optimizer
Woop - Workflow OptimizerWoop - Workflow Optimizer
Woop - Workflow Optimizer
 
UNEP-Live prototype 20110821
UNEP-Live prototype 20110821UNEP-Live prototype 20110821
UNEP-Live prototype 20110821
 
Rv11
Rv11Rv11
Rv11
 

explorationexploitation2011_salperwyck_urvoy_contr_01

  • 1. Stumping Along a Summary Exploration & Exploitation Challenge 2011 Christophe Salperwyck1 Tanguy Urvoy2 1 Orange Labs / Lille University 2 Orange Labs ICML 2011 Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 1 / 27
  • 2. Outline 1 Problem Statement 2 Approach Online Summary Base Predictors Online Combination of Predictors 3 Experiments and Results Adobe Dataset (Challenge) Orange Portal Dataset 4 Conclusion and Discussion Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 2 / 27
  • 3. Problem Statement Online content selection Online content selection problem General problem statement a web site offers different items (news, ads, books. . . ) in order to maximize the number of clicks: select for each visitor the item she is most likely to click on instance described by visitor profile features + item features + mixed features gps position, time, navigation history, query. . . item id, topic, creation time, content keywords. . . cosine similarity, gps distance. . . prediction/learning/selection must be done online: limited time limited memory rare and noisy events Our formalization: only order matters online (bipartite) ranking with partial feedback ranking + scarcity → optimize AUC rather than accuracy partial feedback → risk of being stuck in suboptimal predictive models Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 3 / 27
  • 4. Problem Statement Online content selection Off-line evaluation of online content selection systems online evaluation: costly rejection sampling from logs: unbiased estimator, but too many rejections This challenge simulation setting choose both profile and item from a batch of 6 consecutive log instances less realistic than rejection sampling ⇒ no rejection maybe enough for comparison of algorithms From our own experiments with Orange portal logs some profile features are very discriminant e.g. pageid: orange home page click rate = 5 × other pages only learn to select home page vs. other pages what we did to reduce this bias: take logs from a single page (web-mail) maybe better: select batch of instances with “similar” visitor profiles same time, same visited pages. . . close gps positions etc. Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 4 / 27
  • 5. Approach Online Summary Outline 1 Problem Statement 2 Approach Online Summary Base Predictors Online Combination of Predictors 3 Experiments and Results Adobe Dataset (Challenge) Orange Portal Dataset 4 Conclusion and Discussion Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 5 / 27
  • 6. Approach Online Summary Motivation for a Summary Challenge constraints limited time: 100ms/round for predicting and learning limited memory: 1.7GB of memory keep Garbage Collector quiet We wanted a summary with small memory footprint fast updates fast queries Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 6 / 27
  • 7. Approach Online Summary Numerical features summary Assuming a static distribution, we want to answer questions like: what are the click/¬click counts for feature xi ∈]a, b]? GK merge-prune algorithm: online equal-freq histogram Greenwald M., Khanna S. Space-efficient online computation of quantile summaries. SIGMOD’01 this algorithm maintains an array of tuples < vi , gi , ∆i > where: vi is a value seen in the stream gi is the number of values seen between vi−1 and vi ∆i is the maximum possible error on gi properties: equal-freq, online, bounded memory insensitive to data order and/or data distribution strong error guarantees ( -approximate quantile summary) Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 7 / 27
  • 8. Approach Online Summary Nominal features summary Problem: the number of nominal values is not supposed to be bounded for instance: client ids, cookies or city names we do not need to keep rare values: e.g. outdated cookies or small cities Solution: hashing hash nominal values into a fixed number of buckets count click/¬click for each bucket may be done with several hashing functions to reduce error: Cormode G., Muthukrishnan S. An improved data stream summary: the count-min sketch and its applications. Journal of Algorithms. 2005. Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 8 / 27
  • 9. Approach Base Predictors Outline 1 Problem Statement 2 Approach Online Summary Base Predictors Online Combination of Predictors 3 Experiments and Results Adobe Dataset (Challenge) Orange Portal Dataset 4 Conclusion and Discussion Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 9 / 27
  • 10. Approach Base Predictors Base Predictors: probability estimation trees Output: click-rate estimate Split criterion: Gini impurity Numerical features: simple decision stumps (depth=1) estimate = left estimate if xi ≤ θ right estimate else deeper univariate trees gradual unfolding: add new leaves with data arrival Nominal features: hash-stump: hash and treat as continuous best vs. all: best value/others Univariate trees: work with a single feature global summary → full-rebuild from summary at each step Multivariate trees: combine several features one summary per leaf → unfolding is a definitive action Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 10 / 27
  • 11. Approach Online Combination of Predictors Outline 1 Problem Statement 2 Approach Online Summary Base Predictors Online Combination of Predictors 3 Experiments and Results Adobe Dataset (Challenge) Orange Portal Dataset 4 Conclusion and Discussion Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 11 / 27
  • 12. Approach Online Combination of Predictors Online combination of predictors Output linear combination of predictions Intuition: whack bad predictors exponential weighting online linear ranker set weight to optimize bipartite ranking loss (i.e. AUC) Simple averaging: optimizes nothing Online bagging Oza N.C.,Russell S. Online Bagging and Boosting. ICONS 2001. Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 12 / 27
  • 13. Experiments and Results Experiments Adobe dataset (phase 1 of challenge) 3M instances (i.e. 500K batches of 6) 120 features: 99 continuous / 21 nominals (last nominal feature is ”option id”) click rate: ∼ 0.0024 Orange portal web-mail dataset 6M instances (i.e. 1M batches of 6) 33 nominal features click rate: ∼ 0.0005 Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 13 / 27
  • 14. Experiments and Results Adobe Dataset (Challenge) Outline 1 Problem Statement 2 Approach Online Summary Base Predictors Online Combination of Predictors 3 Experiments and Results Adobe Dataset (Challenge) Orange Portal Dataset 4 Conclusion and Discussion Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 14 / 27
  • 15. Experiments and Results Adobe Dataset (Challenge) Adobe dataset with perfect player Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 15 / 27
  • 16. Experiments and Results Adobe Dataset (Challenge) Adobe dataset without perfect player Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 16 / 27
  • 17. Experiments and Results Adobe Dataset (Challenge) Deviation from static model Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 17 / 27
  • 18. Experiments and Results Adobe Dataset (Challenge) Utree vs. full dataset discretization (MODL) Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 18 / 27
  • 19. Experiments and Results Orange Portal Dataset Outline 1 Problem Statement 2 Approach Online Summary Base Predictors Online Combination of Predictors 3 Experiments and Results Adobe Dataset (Challenge) Orange Portal Dataset 4 Conclusion and Discussion Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 19 / 27
  • 20. Experiments and Results Orange Portal Dataset Orange portal dataset with perfect player Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 20 / 27
  • 21. Experiments and Results Orange Portal Dataset Orange dataset with perfect player Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 21 / 27
  • 22. Conclusion and Discussion Conclusion and discussion Thank you for this challenge! Experiment summary interesting simulation scheme: how many items/batches ? combination of GK merge-prune and hashing provides a good stream summary to rely on for click prediction surprisingly (unfortunately?) simple decision stumps seem to be be competitive base predictors Why are decision stumps so good? maybe because of their stability? maybe stump-induced partial order lets some room for exploration? Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 22 / 27
  • 23. Appendix Experiment summary (for Adobe dataset) GK-100: Greenwald & Khanna (100 tuples), CMS-1: Count min-sketch (1 hash) 1R: single rules xi ∈]a, b] ? 1 : 0 HStump: hash then split like stump RF: random forests, UTree: univariate tree AVG: averaging, OZA: Oza bagging (instance weighting) multivariate predictors typed in red name summary base-predictors combi. score Random 1127 ± 41 LinRank GK-4/CMS-1 1R (proba. output) LRank-120 ∼ 1600 LinRank GK-4/CMS-1 1R (0/1 output) LRank-120 ∼ 1800 RF GK-100/CMS-1 20 trees (10 feat.) AVG-20 ∼ 1900 Stumps1VA GK-100/CMS-1 Stump/1 vs. all AVG-120 ∼ 2000 UTree GK-100/CMS-1 UTree/HStump AVG-120 ∼ 2030 Cnrs ∼ 2040 Simple stumps GK-100/CMS-1 Stump/HStump AVG-120 ∼ 2050 Best GK-100/CMS-1 800 stumps + 200 RF OZA-1K ∼ 2080 Inria ∼ 2200 Perfect 6678 Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 23 / 27
  • 24. Appendix Adobe dataset (start zoom) Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 24 / 27
  • 25. Appendix Adobe dataset (end zoom) Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 25 / 27
  • 26. Appendix Adobe dataset (final zoom) Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 26 / 27
  • 27. Appendix Utree learning Christophe Salperwyck, Tanguy Urvoy (Orange labs) Stumping along a summary July 2011 27 / 27