SlideShare a Scribd company logo
1 of 75
Probability Sampling
  and Alternative
   Methodologies
  “The Future of Survey Research”
    National Science Foundation
            Oct. 4, 2012
             Gary Langer
      Langer Research Associates
      glanger@langerresearch.com
           @Langer Research
Good Data

•   Are powerful and compelling
•   Rise above anecdote
•   Sustain precision
•   Expand our knowledge, enrich our
    understanding and inform our judgment
Other Data
• May be manufactured to promote a product,
  client or point of view
• Are easily manipulated
• Are increasingly prevalent (cheaply produced
  via Internet, e-mail)
• Leave the house of inferential statistics
• Lack the validity and reliability we seek
• Often mis-ask and misanalyze
• Misinform, even disinform our judgment
The Challenge
• Subject established methods to rigorous and
  ongoing evaluation
• Subject new methods to the same standards
• Value theoreticism as much as empiricism
• Consider fitness for purpose
• Go with the facts, not with the fashion
Survey Sampling
    A brief history…
Sample Techniques
                           pre-1936

• Full population (census survey)
  • Pros: High level of accuracy
  • Cons: Prohibitively expensive and highly challenging
    (unless the population is small)

• “Availability” sampling (e.g., straw polls)
  • Pros: Quick and inexpensive
  • Cons: No-scientific basis, no theoretical justification for
    generalizing beyond the sample
  • Justification: “Good enough” – Literary Digest correctly
    predicted presidential election winners from1916 to 1932
The 1936 Election
                  Roosevelt (D) vs. Landon (R)
• Literary Digest poll
  • Technique: An availability sample
  • Sample frame: Combination of magazine subscribers, phone books,
    and car registration lists
  • Contacted 10 million, 2.4 million responded

• Found an easy win for Landon. Roosevelt won in a landslide.
• What happened?
  • Coverage bias: The sampling frame differed in important ways from
    the population of interest. Republicans were more likely to afford
    magazine subscriptions, telephones and cars during the Great
    Depression.
  • Systematic nonresponse: Republicans were disproportionately likely
    to return their ballot (perhaps to express discontent with the New
    Deal).
On to quota sampling
• After 1936, availability sampling largely was replaced with
  quota sampling (which had correctly predicted FDR‟s win).

• Quota sampling essentially attempts to build a “miniature”
  version of the target population.
  • Identifies supposed key demographic characteristics in the
    population and attempts to mirror them in the sample.
  • Matching the sample to the population on variables such as
    gender, race, age, income and others was said to produce a
    “representative” sample.
How‟d that work out?
What went wrong?
1. Timing: Pollsters stopped collecting data in mid-October,
   assuming that there would be no late-breaking changes.

2. Faulty assumptions: Pollsters erroneously assumed that
   undecided voters would vote in proportion with those who
   had made up their minds. Likely voters also may have
   been misidentified.

3. Sampling problems: Quota sampling may have allowed
   interviewers to select more-educated and better-off
   respondents within their assigned quotas – pro-Dewey
   groups. This is a key potential flaw in purposive sampling.
Quota Sampling Issues
•       Even if quota sampling were not entirely to blame, the
        Dewey-Truman fiasco highlighted inherent limitations in
        quota sampling
    •       It‟s impossible to anticipate, let alone match, every
            potentially important demographic:
        •    e.g., if you match on sex (2), ethnicity (2), race (5), education (4), age (5),
             region (4) and income (6) that yields 9,600 cells. If you want to interview
             more than one person per cell…
    •       Respondent selection is left up to the interviewers, allowing
            human bias to sneak in (e.g., picking easy-to-reach resps).
    •       As with availability sampling, there is no theoretical
            justification for drawing inferences about the greater
            population.
Probability Sampling
Not a new concept
A “criterion that a sample design should meet (at least if
one is to make important decisions on the basis of the
sample results) is that the reliability of the sample
results should be susceptible of measurement. An
essential feature of such sampling methods is that each
element of the population being sampled … has a
chance of being included in the sample and, moreover,
that that chance or probability is known.”


Hansen and Hauser, POQ, 1945
This is really not a new concept
“Diagoras, surnamed the Atheist, once paid a visit to
Samothrace, and a friend of his addressed him thus: „You
believe that the gods have no interest in human welfare.
Please observe these countless painted tablets; they show
how many persons have withstood the rage of the tempest
and safely reached the haven because they made vows to
the gods.‟
“„Quite so,‟ Diagoras answered, „but where are the tablets of
those who suffered shipwreck and perished in the deep?‟”
“On the Nature of the Gods,” Marcus Tullius Cicero, 45 B.C.
(cited by Kruskal and Mosteller, International Statistical Review, 1979)
A shared view
•   “…the stratified random sample is recognized by mathematical statisticians
    as the only practical device now available for securing a representative
    sample from human populations…” Snedecor, Journal of Farm Economics,
    1939

•   “It is obvious that the sample can be representative of the population only if
    all parts of the population have a chance of being sampled.” Tippett, The
    Methods of Statistics, 1952

•   “If the sample is to be representative of the population from which it is
    drawn, the elements of the sample must have been chosen at random.”
    Johnson and Jackson, Modern Statistical Methods, 1959

•   “Probability sampling is important for three reasons: (1) Its measurability
    leads to objective statistical inference, in contrast to the subjective
    inference from judgment sampling. (2) Like any scientific method, it permits
    cumulative improvement through the separation and objective appraisal of
    its sources of errors. (3) When simple methods fail, researchers turn to
    probability sampling…” Kish, Survey Sampling, 1965
Today: Internet Opt-Ins
Copyright 2003 Times Newspapers Limited
               Sunday Times (London)

              January 26, 2003, Sunday

SECTION: Features; News; 12

LENGTH: 312 words

HEADLINE: Blair fails to make a case for action that
convinces public

BODY:
Tony Blair has failed to convince people in Britain of
the need for war with Iraq, a poll for The Sunday
Times shows. Even among Labour supporters, the
prime minister has not yet made the case.

The poll of nearly 2,000 people by YouGov shows that
only 26% say Blair has convinced them that Saddam
Hussein is sufficiently dangerous to justify military
action, against 68% who say he has not done so.
-----Original Message-----
From: Students SM Team [mailto:alumteam@teams.com]
Sent: Wednesday, October 04, 2006 11:27 AM
Subject: New Job opening

Hi,
Going to school requires a serious commitment, but most students still need extra money for rent, food,
    gas, books, tuition, clothes, pleasure and a whole list of other things.
So what do you do? "Find some sort of work", but the problem is that many jobs are boring, have low pay
    and rigid/inflexible schedules. So you are in the middle of mid-terms and you need to study but you
    have to be at work, so your grades and education suffer at the expense of your "College Job".
Now you can do flexible work that fits your schedule! Our company and several nationwide companies
   want your help. We are looking to expand, by using independent workers we can do so without buying
   additional buildings and equipment. You can START IMMEDIATELY!
This type of work is Great for College and University Students who are seriously looking for extra income!
We have compiled and researched hundreds of research companies that are willing to pay you
   between $5 and $75 per hour simply to answer an online survey in the peacefulness of your
   own home. That's all there is to it, and there's no catch or gimmicks! We've put together the most
   reliable and reputable companies in the industry. Our list of research companies will allow you to earn
   $5 to $75 filling out surveys on the internet from home. One hour focus groups will earn you $50 to
   $150. It's as simple as that.
Our companies just want you to give them your opinion so that they can empower their own market
    research. Since your time is valuable, they are willing to pay you for it.
If you want to apply for the job position, please email at:
job2@alum.com Students SM Team
-----Original Message-----
From: Ipsos News Alerts [mailto:newsalerts@ipsos-na.com]
Sent: Friday, March 27, 2009 5:12 PM
To: Langer, Gary
Subject: McLeansville Mother Wins a Car By Taking Surveys

McLeansville Mother Wins a Car By Taking Surveys

Toronto, ON- McLeansville, NC native, Jennifer Gattis beats the odds and
wins a car by answering online surveys. Gattis was one of over 105 300
North Americans eligible to win. Representatives from Ipsos i-Say, a
leading online market research panel will be in Greensboro on Tuesday,
March 31, 2009 to present Gattis with a 2009 Toyota Prius.

Access the full press release at:
http://www.ipsos-na.com/news/pressrelease.cfm?id=4331
Opt-in online panelist
32-year-old Spanish-speaking female
     African-American physician
       residing in Billings, MT
Professional Respondents?

Among 10 largest opt-in panels:

•10% of panel participants account for 81% of survey
responses;

•   1% of participants account for 34% of responses.


Gian Fulgoni, chairman, comScore, Council of American Survey
Research Organizations annual conference, Los Angeles, October 2006.
Questions
   Who joins the club, how and why?

   What verification and validation of respondent identities are
    undertaken?

   What logical and QC checks (duration, patterning, data quality)
    are applied?

   What weights are applied, and how? On what theoretical basis
    and with what effect?

   What level of disclosure is provided?

   What claims are made about the data, and how are they
    justified?
One Claim:
    Google Consumer Surveys
• “Produces results that are as accurate as probability-based
  panels.” http://www.google.com/insights/consumersurveys/how
• 1 or 2-question pop-ups delivered to search engine users of
  “premium content.”
• Demographic data are imputed through analysis of users‟ IP
  addresses and previous page views
• Among Langer Research staff, one woman was identified as
  a 55-year-old male with an interest in beauty and fitness;
  another as a 65+ woman (double her actual age) with an
  interest in motorcycles and one man as a senior citizen in
  the Pacific Northwest with an interest in space technology.
  https://www.google.com/settings/ads/onweb/
Another claim:
“Bayesian Credibility Interval”
• Ipsos opt-in online panel – presdential tracking
  poll

• Computation of “Credibility Internval” appears to
  match how you‟d compute SRS MoE.

• Its basis is “the opinion of Ipsos experts.”

• (Headdesk)
Further claims:
        Convenience Sample MoE
• Zogby Interactive: "The margin of error is +/- 0.6 percentage points.”
• Ipsos/Reuters: “The margin of error is plus or minus 3.1 percentage
  points."
• Kelton Research: “The survey results indicate a margin of error of +/-
  3.1 percent at a 95 percent confidence level.”
• Economist/YouGov/Polimetrix: “Margin of error: +/- 4%.”
• PNC/HNW/Harris Interactive: “Findings are significant at the 95 percent
  confidence level with a margin of error of +/- 2.5 percent.”
• Radio One/Yankelovich: “Margin of error: +/-2 percentage points.”
• Citi Credit-ED/Synovate: “The margin of error is +/- 3.0 percentage
  points.”
• Spectrem: “The data have a margin of error of plus or minus 6.2
  percentage points.”
• Luntz: “+3.5% margin of error”
Online Survey

Traditional phone-based survey techniques suffer from deteriorating response rates and
escalating costs. YouGovPolimetrix combines technology infrastructure for data collection,
integration with large scale databases, and novel instrumentation to deliver new
capabilities for polling and survey research.
(read more)
Response Rates

• Surveys based in probability sampling cannot achieve pure
  probability (e.g., 100% RR)

• Pew (5/15/12) reports decline in its RRs from 36% in 1997
  to 9% now. (ABC/Post: 19%)

• Does this trend poison the well?
A Look at the Lit
2006: "…little to suggest that unit nonresponse within the range of
response rates obtained seriously threatens the quality of survey
estimates."
Keeter et al., POQ, 2006 – comp. RR 25 vs. 50%; see also Keeter et al., POQ, 2000, RR 36 vs. 61

2008: “In general population RDD telephone surveys, lower response
rates do not notably reduce the quality of survey demographic estimates.
… This evidence challenges the assumption that response rates are a
key indicator of survey data quality…”
Holbrook et al.: Demographic comparison, 81 RDD surveys, 1996-2005; AAPOR3 RRs from 5 to 54%

2012: “Overall, there are only modest differences in responses between
the standard and high-effort surveys. Similar to 1997 and 2003, the
additional time and effort to encourage cooperation in the high-effort
survey does not lead to significantly different estimates on most
questions.”
Pew 5/15/12 - Comp. RR 9 vs. 22%
See also O‟Neil, 1979; Smith, 1984; Merkle et al., 1993;
Curtin et al., 2000; Groves et al., 2004; Curtin et al., 2005.

And Groves, 2006:

“Hence, there is little empirical support for the notion that low
response rate surveys de facto produce estimates with high
nonresponse bias.”
Next step
Empirical testing
Yeager, Krosnick, et al., 2011
                 (See also Malhotra and Krosnick, 2006)


• Paper compares seven opt-in online convenience-sample
  surveys with two probability sample surveys
• Probability-sample surveys were “consistently highly accurate”
• Opt-in online surveys were “always less accurate… and less
  consistent in their level of accuracy.”
• Weighted probability samples sig. diff. from benchmarks 31 or
  46 percent of the time for probability samples, 62 to 77 percent
  of the time for convenience samples.
• Opt-in data‟s highest single error vs. benchmarks, 19 points;
  average error, 9.9.
Average Absolute Errors
                           primary (weighting) demographics, unweighted



       20%
       18%
       16%
       14%                                                                        12.0%
       12%
       10%
        8%                                            6.4%   6.4%
        6%                                     5.0%                 5.3%   4.7%
                                        4.1%
        4%        3.3%
                                 2.0%
        2%
        0%




Yeager, Krosnick, et al., 2011
Average Absolute Errors
                                 secondary and non-demographic results


       10%                              Unweighted     Weighted
         9%
         8%
         7%
                                                                5.9%
         6%                                                              5.2%
         5%
                                 3.8%
         4%
                                         3.2%
         3%
         2%
         1%
         0%
                  Probability samples (combined)         Internet samples (combined)



Yeager, Krosnick, et al., 2011
Largest Absolute Error
                                        all benchmarks, unweighted


       40%
                                                                                        35.5%
       35%
       30%
       25%
       20%                              18.0%
                                                        15.6%   15.3%           16.0%
       15%                                      13.2%                   13.7%
                  11.7%
                                 9.6%
       10%
         5%
         0%




Yeager, Krosnick, et al., 2011
Largest Absolute Error
                                 secondary and non-demographic results


       20%                              Unweighted     Weighted
       18%
       16%                                                    14.9%
       14%                                                               13.5%
       12%                   10.7%
       10%                              8.7%
         8%
         6%
         4%
         2%
         0%
                  Probability samples (combined)         Internet samples (combined)

Yeager, Krosnick, et al., 2011
% significantly different from benchmarks
                                 secondary and non-demographic results



    100%                                 Unweighted       Weighted
      90%
      80%
                                                                70.6%    71.4%
      70%
      60%                    57.5%
      50%
                                      38.5%
      40%
      30%
      20%
      10%
        0%
                  Probability samples (combined)         Internet samples (combined)

Yeager, Krosnick, et al., 2011
Additional conclusions
                        Yeager, Krosnick, et al., 2011


• Little support for the claim that some non-probability internet
  surveys are consistently more accurate than others.
  • Weighting did not always improve the accuracy of the opt-in samples.
• No support for the idea that higher completion = greater
  accuracy:
  • There was a significant negative correlation between a survey‟s average
    absolute error and its response rate.
• Probability samples were not only more accurate overall, but
  also more consistently accurate.
  • True both across surveys (using pre-existing probability samples as
    comparison points) as well as within surveys.
  • Meaning it‟s virtually impossible to anticipate whether an opt-in survey
    will be just somewhat less accurate than a probability sample or
    substantially less accurate; and knowing an opt-in is accurate on one
    benchmark does not predict its accuracy on other measures.
Additional conclusions
                    Yeager, Krosnick, et al., 2011



• The fact that internet samples are less accurate, and less
  consistent, on average, than probability samples should “come
  as no surprise, because no theory provides a rationale whereby
  samples generated by non-probability methods would yield
  accurate results”

• It is “possible to cherry-pick such results to claim that non-
  probability sampling can yield veridical measurements. But a
  systematic look at a wide array of benchmarks documented that
  such results are the exception rather than the rule.”
ARF FoQ via Reg Baker, 2009

• Reported data on estimates of smoking prevalence: similar
  across three probability methods, but with as many as 14
  points of variation across 17 opt-in online panels.

• “In the end, the results we get for any given study are
  highly dependent (and mostly unpredictable) on the panel
  we use. This is not good news.”
Callegaro et al., 2012
• Reviewed 45 studies comparing the quality of data
  collected via opt-in panels vs. another mode and/or
  benchmarks. In summary:
  • Panel estimates substantially deviated from benchmarks, and
    to a far greater degree than probability samples
  • High variability in data from different opt-in panels
  • High levels of multiple-panel membership (between 19-45% of
    respondents belong to 5+ panels)
  • Substantial differences among low- and higher-membership
    respondents
  • Weighting did not correct variations in the data
Traugott, 2012
• Tested the accuracy of four low cost data collection
  (LCDC) methods (Mechanical Turk, Pulse Opinion
  Research, Qualtrics and Zoomerang)
  • LCDC unweighted demographics substantially deviated from
    ACS benchmarks to a far greater degree than RDD polls or
    the ANES
    • MTurk was substantially younger (45% under 30), female
      (60%), and more educated (47% with a college degree) than
      benchmarks.
    • Pulse was disproportionately older (43% over 65), female
      (61%), white (90%) and educated (46% with a college degree).
    • Qualtrics and Zoomerang had better unweighted age, sex, and
      race distributions, but still had far too few respondents without a
      high school degree and too many with some college or more.
Traugott, 2012 (cont‟d)
100%                     Unweighted Party ID
90%                  Republican       Democrat      Independent
80%
70%
60%                       57%
                                                    49%
50%                                                                     44%
                                            41%           42%
40%      35%                          37%                         36%
               31%
       27%             28%
30%
20%                                           13%                         13%
                                11%
10%
                                                            1%
 0%
        ANES             Mturk        Zoomerang       Pulse       Qualtrics
Traugott, 2012 (cont‟d)
100%                         Weighted Party ID
90%                  Republican       Democrat      Independent
80%
70%
                          58%
60%
50%                                         43%     44% 45%             46%

40%      37%
                                      34%                         34%
               31%
30%    26%             26%

20%                             12%           13%                             13%
10%                                                       2%
 0%
        ANES             Mturk        Zoomerang       Pulse       Qualtrics
AAPOR‟s “Report on Online Panels,” April 2010

•   “Researchers should avoid nonprobability online panels when one of the
    research objectives is to accurately estimate population values.”
•   “The nonprobability character of volunteer online panels … violates the
    underlying principles of probability theory.”
•   “Empirical evaluations of online panels abroad and in the U.S. leave no
    doubt that those who choose to join online panels differ in important and
    nonignorable ways from those who do not.”
•   “In sum, the existing body of evidence shows that online surveys with
    nonprobability panels elicit systematically different results than probability
    sample surveys in a wide variety of attitudes and behaviors.”
•   “The reporting of a margin of sampling error associated with an opt-in
    sample is misleading.”
AAPOR‟s conclusion

“There currently is no generally accepted theoretical basis
from which to claim that survey results using samples from
nonprobability online panels are projectable to the general
population. Thus, claims of „representativeness‟ should be
avoided when using these sample sources.”
AAPOR Report on Online Panels, April 2010
OK, but apart from population values, we
  can still use convenience samples to
 evaluate relationships among variables
           and trends over time.

                Right?
Pasek & Krosnick, 2010
• Comparison of opt-in online and RDD surveys sponsored by the U.S.
  Census Bureau assessing intent to fill out the Census.
• “The telephone samples were more demographically representative of
  the nation‟s population than were the Internet samples, even after
  post-stratification.”
• “The distributions of opinions and behaviors were often significantly
  and substantially different across the two data streams.
  Thus, research conclusions would often be different.”
• Instances “where the two data streams told very different stories about
  change over time … over-time trends in one line did not meaningfully
  covary with over-time trends in the other line.”
Top Predictors of Intent to
            Complete the Census
   In the probability RDD sample             In the internet opt-in sample
Age: 18-24                       -1.76   Age: 18-24                         -1.81
Education: <H.S.                 -1.16   Age: 25-44                         -1.29
Education: H.S. graduate         -.97    Think Census can help them          1.19
Spanish speaking household       .97     Think Census can harm them          -.98
Counting everyone is important   .92     Participation doesn‟t matter         .95
Age: 25-44                       -.76    (disagree)

Don‟t have time (disagree)       .70     Counting everyone is important       .71
Education: Some college          -.67    Age: 45-64                          -.62

Participation doesn‟t matter     .67     Don‟t have time (disagree)           .61
(disagree)                               Race: Hispanic                       .56
Counting everyone is             -.62    Education: <H.S.                    -.47
important (disagree)                                         Pasek & Krosnick, 2010
Biggest Differences in Predictors
of Intent to Complete the Census
                               RDD       Opt-In         Diff.
 Spanish speaking household    .97***     -.38        -1.34***
 Importance of counting       -.62***      .14          .76**
 everyone (disagree)
 Education: <H.S.             -1.16***    -.47*         .69**
 Education: H.S. graduate     -.97***     -.38          .59**
 Think Census can help them    .61***    1.19***       .57***
 Age: 25-44                   -.76***    -1.29***      -.54**
 Age: 45-64                    -.21*     -.62***        -.42*
 Region: Northeast             -.06       .27*          .33*
                                           Pasek & Krosnick, 2010
Differences in Time Trend




                 Pasek & Krosnick, 2010
Pasek/Krosnick conclusion

“This investigation revealed systematic and often sizable
differences between probability sample telephone data
and non-probability Internet data in terms of demographic
representativeness of the samples, the proportion of
respondents reporting various opinions and behaviors, the
predictors of intent to complete the Census form and
actual completion of the form, changes over time in
responses, and relations between variables.”
Pasek and Krosnick, 2010
Can non-probability samples
           be „fixed‟?
• Bayesian analysis?
  • What variables? How derived? Are we weighting to the DV
    or its correlates?

• Sample balancing? (i.e., quota sampling revisited)
  • “The microcosm idea will rarely work in a complicated
    social problem because we always have additional
    variables that may have important consequences for the
    outcome.”
  • Gilbert, Light and Mosteller, Statistics and Public Policy, 1977
Some say you can

“The survey was administered by YouGovPolimetrix during
July 16-July 26, 2008. YouGovPolimetrix employs sample
matching techniques to build representative web samples
through its pool of opt-in respondents (see Rivers 2008).
Studies that use representative samples yielded this way
find that their quality meets, and sometimes exceeds, the
quality of samples yielded through more traditional survey
techniques.”

Perez, Political Behavior, 2010
AAPOR‟s task force disagrees

“There currently is no generally accepted theoretical
basis from which to claim that survey results using
samples from nonprobability online panels are
projectable to the general population. Thus, claims of
„representativeness‟ should be avoided when using these
sample sources.”
AAPOR Report on Online Panels, 2010
It has company
• “Unfortunately, convenience samples are often used inappropriately
  as the basis for inference to some larger population.”
• “…unlike random samples, purposive samples contain no
  information on how close the sample estimate is to the true value of
  the population parameter.”
• “Quota sampling suffers from essentially the same limitations as
  convenience, judgment, and purposive sampling (i.e., it has no
  probabilistic basis for statistical inference).”
• “Some variations of quota sampling contain elements of random
  sampling but are still not statistically valid methods.”

Biemer and Lyberg, Introduction to Survey Quality, 2003
Social Media
Sampling Challenges
• Users are not limited to one tweet/post per day;
  some may post incessantly; others rarely
• Users can have multiple Twitter accounts
• Accounts often do not belong to individuals, but
  rather companies, organizations or associations
• Posts can be driven by a public relations campaign
  (i.e., through volunteers, paid agents or bots) rather
  than as an expression of individual attitudes
Sampling Challenges (cont.)
• Users‟ location is often not accurately captured (if
  at all)
• Users are self-selected. There is no theoretical
  basis on which to assume that their views are
  representative of the views of non-users.
  • 15 percent of online adults use Twitter, 8 percent daily (Pew)
  • Fewer than 30 percent of Twitter users are Americans
    (Semiocast)
  • In 2011, 45% of U.S. Facebook users were 25 or younger
    (Smith, 2012)
Sentiment Analysis
• Determining the meaning of posts or tweets requires
  content analysis:
  • Traditional approaches (using independent coders and a
    codeback) typically is unrealistic given the volume of data
  • Computerized sentiment analysis programs, while lower-cost
    and faster, are fraught with problems:
    • Parsing
      slang, irony, sarcasm, abbreviations, acronyms, emoticons,
       contextual meaning and hashtags is highly complex
    • Ambiguity in determining target words (e.g., Obama vs.
      prez)
• Regardless, contextual data (demographic, other attitudes) are
  sparse or absent, severely limiting the research questions that
  can be answered
Manual vs. Automated Coding
                                         Manual Coding
                              Positive      Neutral      Negative   Uncodable
                              (n=100)       (n=285)       (n=81)     (n=34)
                   Positive     8%            5%           2%          0%
  radian6
                   Neutral     86%           90%          83%         94%
    55%
                   Negative     0%            5%          15%          6%
                   Positive    42%           25%           6%         21%
    STAS
                   Neutral     45%           57%          71%         68%
    45%
                   Negative    13%           18%          23%         12%
                   Positive    30%           20%           2%          9%
clarabridge
                   Neutral     60%           61%          85%         79%
    43%
                   Negative    10%           19%          12%         12%
Kim et al., 2012
Data Mining
• Rather than attempting to parse the content of online
  posts, some researchers instead focus on the sheer quantity
  of posts about a certain topic.
  • Tumasjan et al. (2010) analyzed the German national election and
    found that “the mere number of tweets mentioning a political
    party” was quite good at reflecting their vote share and that its
    predictive power was similar to that of traditional polls.
  • At the same time, they found that only 4% of users accounted for
    more than 40% of the messages.
    • When such a small number of people are accounting for a large
      portion of the results, it opens the door for politically motivated
      actors to distort results intentionally – especially as the use of such
      methods becomes more popular.
Data Mining
• Similar analyses in the U.S. have differed
  • Lui et al. (2011) showed that “Google trends” was worse at
    predicting 2008 and 2010 election outcomes than the New
    York Times (probability-sample) polls and even chance.
  • Gayo-Avello et al. (2011) looked at the 2010 election and
    found the mean average error for Twitter volume was 17.1%
    (it was 7.6% for sentiment analysis), far higher than the 2-
    3% that is typical of RDD polls.
  • Olson & Bunnett (2011) found that Facebook “likes” in 2010
    explained 13% of the variance in Senate races, but were
    very weak (or slightly negative) predictors of gubernatorial
    and house races
Twitter Index?
                           The Twitter Political Index




Aug. 6 – Obama‟s index # is three times greater than Romney‟s (“61” v. “24”)
Aug. 7 – “35” v. “19”
Facebook sampling?
• Bhutta, 2012: Snowball sampling to reach Catholics
  • faster and cheaper than traditional methods, but:

                           Facebook    GSS
          70%                         67%          66%
                53%


                           33%
                                         24%             27%


                      6%


         % female     % Latino   % college grad. % attend mass
                                                    weekly+
“…the Facebook respondents cannot possibly
serve as a representative sample of the general
Catholic population. (Pollsters should view
Facebook findings with extreme caution.)”

Bhutta, 2012
Twitter bots

•Recent estimate that as many as 30% of Obama‟s
Twitter followers and 22% of Romney‟s were
fabricated. (Calzolari, 2012)

•News report: Individuals can purchase 25,000
fabricated Twitter followers for $247; these “bots” will
automatically tweet snippets of text. (Feifer, 2012)
And the future?
In probability sampling:
• Continued study of response-rate effects
• Concerted efforts to maintain response rates
• Renewed focus on other data-quality issues in
  probability samples, e.g. coverage
• Development of probability-based
  alternatives, e.g. mixed-mode, ABS
The Future, cont.
In convenience sampling:
• Further study of appropriate uses (as well as
  inappropriate misuses) of convenience-sample
  data
• Further evaluation of well-disclosed, emerging
  techniques in convenience sampling
• Enhanced disclosure!
• The quest for an online sampling frame
The future, cont.
In social media:
• Vast and growing source of data
• Further evaluation of its appropriate use is needed
  • It‟s important to study how social media influences the formation
    and change of public opinion. But that does not mean it can
    supplant probability-sample polling.
  • Information gleaned from social media may be useful as a
    compliment to public opinion polls (akin to focus groups); for
    trend-spotting, qualitative analysis, perhaps modeling

• The key: establish the research question; evaluate
  the limitations of the tools available to answer it.
Thank you!

        Gary Langer
 Langer Research Associates
glanger@langerresearch.com

More Related Content

What's hot

Modeling Social Data, Lecture 2: Introduction to Counting
Modeling Social Data, Lecture 2: Introduction to CountingModeling Social Data, Lecture 2: Introduction to Counting
Modeling Social Data, Lecture 2: Introduction to Countingjakehofman
 
475 2015 media effects methods up
475 2015 media effects methods up475 2015 media effects methods up
475 2015 media effects methods upmpeffl
 
Cyberbullying and Hate Speech
Cyberbullying and Hate SpeechCyberbullying and Hate Speech
Cyberbullying and Hate SpeechBrandwatch
 
Increasing Voter Knowledge with Pre-Election Interventions on Facebook
Increasing Voter Knowledge with Pre-Election Interventions on FacebookIncreasing Voter Knowledge with Pre-Election Interventions on Facebook
Increasing Voter Knowledge with Pre-Election Interventions on FacebookMIT GOV/LAB
 
Guide to survey_poll
Guide to survey_pollGuide to survey_poll
Guide to survey_pollKeiko Ono
 
How social media used by politicians? 2016
How social media used by politicians? 2016How social media used by politicians? 2016
How social media used by politicians? 2016Susana Gallardo
 
Impact of social media on voting behaviour
Impact of social media on voting behaviourImpact of social media on voting behaviour
Impact of social media on voting behaviourShoeb Khan
 
Generation Apathy- A study of youth interaction with the Irish democratic pro...
Generation Apathy- A study of youth interaction with the Irish democratic pro...Generation Apathy- A study of youth interaction with the Irish democratic pro...
Generation Apathy- A study of youth interaction with the Irish democratic pro...Lucy O' Shea
 
Capstone Thesis Paper
Capstone Thesis PaperCapstone Thesis Paper
Capstone Thesis PaperEmily Greene
 
Modeling Social Data, Lecture 10: Networks
Modeling Social Data, Lecture 10: NetworksModeling Social Data, Lecture 10: Networks
Modeling Social Data, Lecture 10: Networksjakehofman
 
Slayter - SWK 855 - Election Teach-In 2016
Slayter - SWK 855 - Election Teach-In 2016Slayter - SWK 855 - Election Teach-In 2016
Slayter - SWK 855 - Election Teach-In 2016Elspeth Slayter
 
Election, particiption and voting behaviour
Election, particiption and voting behaviourElection, particiption and voting behaviour
Election, particiption and voting behaviourShyam Budhwar
 

What's hot (20)

Modeling Social Data, Lecture 2: Introduction to Counting
Modeling Social Data, Lecture 2: Introduction to CountingModeling Social Data, Lecture 2: Introduction to Counting
Modeling Social Data, Lecture 2: Introduction to Counting
 
475 2015 media effects methods up
475 2015 media effects methods up475 2015 media effects methods up
475 2015 media effects methods up
 
Cyberbullying and Hate Speech
Cyberbullying and Hate SpeechCyberbullying and Hate Speech
Cyberbullying and Hate Speech
 
Top 14 booklet 2019
Top 14 booklet 2019Top 14 booklet 2019
Top 14 booklet 2019
 
Top 14 booklet 2019
Top 14 booklet 2019 Top 14 booklet 2019
Top 14 booklet 2019
 
Top 14 booklet 2019
Top 14 booklet 2019Top 14 booklet 2019
Top 14 booklet 2019
 
Intentional exposure to pornography online: Is everybody doing it?
Intentional exposure to pornography online: Is everybody doing it?Intentional exposure to pornography online: Is everybody doing it?
Intentional exposure to pornography online: Is everybody doing it?
 
Politics Have Failed Us
Politics Have Failed UsPolitics Have Failed Us
Politics Have Failed Us
 
Youth perpetration of online sexual solicitation: Internet use and psychosoci...
Youth perpetration of online sexual solicitation: Internet use and psychosoci...Youth perpetration of online sexual solicitation: Internet use and psychosoci...
Youth perpetration of online sexual solicitation: Internet use and psychosoci...
 
Increasing Voter Knowledge with Pre-Election Interventions on Facebook
Increasing Voter Knowledge with Pre-Election Interventions on FacebookIncreasing Voter Knowledge with Pre-Election Interventions on Facebook
Increasing Voter Knowledge with Pre-Election Interventions on Facebook
 
Guide to survey_poll
Guide to survey_pollGuide to survey_poll
Guide to survey_poll
 
Splc the trump_effect
Splc the trump_effectSplc the trump_effect
Splc the trump_effect
 
How social media used by politicians? 2016
How social media used by politicians? 2016How social media used by politicians? 2016
How social media used by politicians? 2016
 
Impact of social media on voting behaviour
Impact of social media on voting behaviourImpact of social media on voting behaviour
Impact of social media on voting behaviour
 
Generation Apathy- A study of youth interaction with the Irish democratic pro...
Generation Apathy- A study of youth interaction with the Irish democratic pro...Generation Apathy- A study of youth interaction with the Irish democratic pro...
Generation Apathy- A study of youth interaction with the Irish democratic pro...
 
Capstone Thesis Paper
Capstone Thesis PaperCapstone Thesis Paper
Capstone Thesis Paper
 
Modeling Social Data, Lecture 10: Networks
Modeling Social Data, Lecture 10: NetworksModeling Social Data, Lecture 10: Networks
Modeling Social Data, Lecture 10: Networks
 
Slayter - SWK 855 - Election Teach-In 2016
Slayter - SWK 855 - Election Teach-In 2016Slayter - SWK 855 - Election Teach-In 2016
Slayter - SWK 855 - Election Teach-In 2016
 
Youth exposure to pornography and violent web sites
Youth exposure to pornography and violent web sitesYouth exposure to pornography and violent web sites
Youth exposure to pornography and violent web sites
 
Election, particiption and voting behaviour
Election, particiption and voting behaviourElection, particiption and voting behaviour
Election, particiption and voting behaviour
 

Viewers also liked

Non probability sampling
Non probability samplingNon probability sampling
Non probability samplingsafwanthayath
 
PROBABILITY SAMPLING TECHNIQUES
PROBABILITY SAMPLING TECHNIQUESPROBABILITY SAMPLING TECHNIQUES
PROBABILITY SAMPLING TECHNIQUESAzam Ghaffar
 
Non Probability Sampling
Non Probability SamplingNon Probability Sampling
Non Probability SamplingMuhammad Farooq
 
probability sampling
probability samplingprobability sampling
probability samplingRoshni Kapoor
 
Probability sampling
Probability samplingProbability sampling
Probability samplingBhanu Teja
 
Sampling Design
Sampling DesignSampling Design
Sampling DesignJale Nonan
 
Non probability sampling
Non  probability samplingNon  probability sampling
Non probability samplingcorayu13
 
Sampling and Sample Types
Sampling  and Sample TypesSampling  and Sample Types
Sampling and Sample TypesDr. Sunil Kumar
 
SAMPLING AND SAMPLING ERRORS
SAMPLING AND SAMPLING ERRORSSAMPLING AND SAMPLING ERRORS
SAMPLING AND SAMPLING ERRORSrambhu21
 

Viewers also liked (11)

Non probability sampling
Non probability samplingNon probability sampling
Non probability sampling
 
PROBABILITY SAMPLING TECHNIQUES
PROBABILITY SAMPLING TECHNIQUESPROBABILITY SAMPLING TECHNIQUES
PROBABILITY SAMPLING TECHNIQUES
 
Sampling
SamplingSampling
Sampling
 
Non Probability Sampling
Non Probability SamplingNon Probability Sampling
Non Probability Sampling
 
probability sampling
probability samplingprobability sampling
probability sampling
 
Probability sampling
Probability samplingProbability sampling
Probability sampling
 
Sampling Design
Sampling DesignSampling Design
Sampling Design
 
Non probability sampling
Non  probability samplingNon  probability sampling
Non probability sampling
 
Sampling and Sample Types
Sampling  and Sample TypesSampling  and Sample Types
Sampling and Sample Types
 
Quota Sampling
Quota SamplingQuota Sampling
Quota Sampling
 
SAMPLING AND SAMPLING ERRORS
SAMPLING AND SAMPLING ERRORSSAMPLING AND SAMPLING ERRORS
SAMPLING AND SAMPLING ERRORS
 

Similar to Probability Sampling and Alternative Methodologies

Survey (Primer on Questions, Sampling + Case Study)
Survey (Primer on Questions, Sampling + Case Study)Survey (Primer on Questions, Sampling + Case Study)
Survey (Primer on Questions, Sampling + Case Study)Dada Veloso-Beltran
 
475 media effects methods 2012 up
475 media effects methods 2012 up475 media effects methods 2012 up
475 media effects methods 2012 upmpeffl
 
PS 101 Public Opinion & Polling Fall 2008
PS 101 Public Opinion & Polling Fall 2008PS 101 Public Opinion & Polling Fall 2008
PS 101 Public Opinion & Polling Fall 2008Christopher Rice
 
Classroom Obsevation- 4 SAMPLING METHODS.pptx
Classroom Obsevation- 4 SAMPLING METHODS.pptxClassroom Obsevation- 4 SAMPLING METHODS.pptx
Classroom Obsevation- 4 SAMPLING METHODS.pptxMarkBryanLoterte1
 
Developing A Final Draft Of A Research Paper ENGL 1
Developing A Final Draft Of A Research Paper  ENGL 1Developing A Final Draft Of A Research Paper  ENGL 1
Developing A Final Draft Of A Research Paper ENGL 1Kelley Hunter
 
Beyond the Bubble: A Critical Review of the Evidence for Echo Chambers and Fi...
Beyond the Bubble: A Critical Review of the Evidence for Echo Chambers and Fi...Beyond the Bubble: A Critical Review of the Evidence for Echo Chambers and Fi...
Beyond the Bubble: A Critical Review of the Evidence for Echo Chambers and Fi...Axel Bruns
 
final-Sampling-techniques.ppt
final-Sampling-techniques.pptfinal-Sampling-techniques.ppt
final-Sampling-techniques.pptsaifurrahmanSapai
 
Lab Write-Up Rubric Lab Write-Ups are worth 10 points tota.docx
Lab Write-Up Rubric Lab Write-Ups are worth 10 points tota.docxLab Write-Up Rubric Lab Write-Ups are worth 10 points tota.docx
Lab Write-Up Rubric Lab Write-Ups are worth 10 points tota.docxsmile790243
 
Are Filter Bubbles Real?
Are Filter Bubbles Real?Are Filter Bubbles Real?
Are Filter Bubbles Real?Axel Bruns
 
Fake News, Real Concerns: Developing Information Literate Students (December ...
Fake News, Real Concerns: Developing Information Literate Students (December ...Fake News, Real Concerns: Developing Information Literate Students (December ...
Fake News, Real Concerns: Developing Information Literate Students (December ...ALAeLearningSolutions
 
Student Name CATEGORY AND CRITERIA(A+)15(A).docx
Student Name CATEGORY AND CRITERIA(A+)15(A).docxStudent Name CATEGORY AND CRITERIA(A+)15(A).docx
Student Name CATEGORY AND CRITERIA(A+)15(A).docxcpatriciarpatricia
 
The Spectrum of Community-Based Action Research
The Spectrum of Community-Based Action ResearchThe Spectrum of Community-Based Action Research
The Spectrum of Community-Based Action ResearchEverettProgram
 
Student Name CATEGORY AND CRITERIA(A+)15(A).docx
Student Name CATEGORY AND CRITERIA(A+)15(A).docxStudent Name CATEGORY AND CRITERIA(A+)15(A).docx
Student Name CATEGORY AND CRITERIA(A+)15(A).docxdeanmtaylor1545
 
Week 12 13 march 3rd 2012
Week 12 13 march 3rd 2012Week 12 13 march 3rd 2012
Week 12 13 march 3rd 2012icleir_
 
POL SOC 360 Sampling Generalizability
POL SOC 360 Sampling Generalizability POL SOC 360 Sampling Generalizability
POL SOC 360 Sampling Generalizability atrantham
 
Methodology 2.pptx
Methodology 2.pptxMethodology 2.pptx
Methodology 2.pptxMarcCollazo1
 
Teaching Students How (Not) to Lie with Statistics
Teaching Students How (Not) to Lie with StatisticsTeaching Students How (Not) to Lie with Statistics
Teaching Students How (Not) to Lie with StatisticsLynette Hoelter
 
Survey question and questionnaire design slideshare 022113 dmf
Survey question and questionnaire design slideshare 022113 dmfSurvey question and questionnaire design slideshare 022113 dmf
Survey question and questionnaire design slideshare 022113 dmfDavid Filiberto
 

Similar to Probability Sampling and Alternative Methodologies (20)

Survey (Primer on Questions, Sampling + Case Study)
Survey (Primer on Questions, Sampling + Case Study)Survey (Primer on Questions, Sampling + Case Study)
Survey (Primer on Questions, Sampling + Case Study)
 
475 media effects methods 2012 up
475 media effects methods 2012 up475 media effects methods 2012 up
475 media effects methods 2012 up
 
PS 101 Public Opinion & Polling Fall 2008
PS 101 Public Opinion & Polling Fall 2008PS 101 Public Opinion & Polling Fall 2008
PS 101 Public Opinion & Polling Fall 2008
 
Classroom Obsevation- 4 SAMPLING METHODS.pptx
Classroom Obsevation- 4 SAMPLING METHODS.pptxClassroom Obsevation- 4 SAMPLING METHODS.pptx
Classroom Obsevation- 4 SAMPLING METHODS.pptx
 
Developing A Final Draft Of A Research Paper ENGL 1
Developing A Final Draft Of A Research Paper  ENGL 1Developing A Final Draft Of A Research Paper  ENGL 1
Developing A Final Draft Of A Research Paper ENGL 1
 
Beyond the Bubble: A Critical Review of the Evidence for Echo Chambers and Fi...
Beyond the Bubble: A Critical Review of the Evidence for Echo Chambers and Fi...Beyond the Bubble: A Critical Review of the Evidence for Echo Chambers and Fi...
Beyond the Bubble: A Critical Review of the Evidence for Echo Chambers and Fi...
 
final-Sampling-techniques.ppt
final-Sampling-techniques.pptfinal-Sampling-techniques.ppt
final-Sampling-techniques.ppt
 
Lab Write-Up Rubric Lab Write-Ups are worth 10 points tota.docx
Lab Write-Up Rubric Lab Write-Ups are worth 10 points tota.docxLab Write-Up Rubric Lab Write-Ups are worth 10 points tota.docx
Lab Write-Up Rubric Lab Write-Ups are worth 10 points tota.docx
 
Are Filter Bubbles Real?
Are Filter Bubbles Real?Are Filter Bubbles Real?
Are Filter Bubbles Real?
 
Fake News, Real Concerns: Developing Information Literate Students (December ...
Fake News, Real Concerns: Developing Information Literate Students (December ...Fake News, Real Concerns: Developing Information Literate Students (December ...
Fake News, Real Concerns: Developing Information Literate Students (December ...
 
Student Name CATEGORY AND CRITERIA(A+)15(A).docx
Student Name CATEGORY AND CRITERIA(A+)15(A).docxStudent Name CATEGORY AND CRITERIA(A+)15(A).docx
Student Name CATEGORY AND CRITERIA(A+)15(A).docx
 
The Spectrum of Community-Based Action Research
The Spectrum of Community-Based Action ResearchThe Spectrum of Community-Based Action Research
The Spectrum of Community-Based Action Research
 
Student Name CATEGORY AND CRITERIA(A+)15(A).docx
Student Name CATEGORY AND CRITERIA(A+)15(A).docxStudent Name CATEGORY AND CRITERIA(A+)15(A).docx
Student Name CATEGORY AND CRITERIA(A+)15(A).docx
 
Week 12 13 march 3rd 2012
Week 12 13 march 3rd 2012Week 12 13 march 3rd 2012
Week 12 13 march 3rd 2012
 
Public opinion sp2020
Public opinion sp2020Public opinion sp2020
Public opinion sp2020
 
POL SOC 360 Sampling Generalizability
POL SOC 360 Sampling Generalizability POL SOC 360 Sampling Generalizability
POL SOC 360 Sampling Generalizability
 
Methodology 2.pptx
Methodology 2.pptxMethodology 2.pptx
Methodology 2.pptx
 
Teaching Students How (Not) to Lie with Statistics
Teaching Students How (Not) to Lie with StatisticsTeaching Students How (Not) to Lie with Statistics
Teaching Students How (Not) to Lie with Statistics
 
SociologyExchange.co.uk Shared Resource
SociologyExchange.co.uk Shared ResourceSociologyExchange.co.uk Shared Resource
SociologyExchange.co.uk Shared Resource
 
Survey question and questionnaire design slideshare 022113 dmf
Survey question and questionnaire design slideshare 022113 dmfSurvey question and questionnaire design slideshare 022113 dmf
Survey question and questionnaire design slideshare 022113 dmf
 

More from LangerResearch

Correlations: Bloomberg CCI and Other Indicators
Correlations: Bloomberg CCI and Other IndicatorsCorrelations: Bloomberg CCI and Other Indicators
Correlations: Bloomberg CCI and Other IndicatorsLangerResearch
 
Predicting State-Level 2016 Vote using MRP and a National Tracking Poll
Predicting State-Level 2016 Vote using MRP and a National Tracking PollPredicting State-Level 2016 Vote using MRP and a National Tracking Poll
Predicting State-Level 2016 Vote using MRP and a National Tracking PollLangerResearch
 
The 2016 Election - How and why it's President Trump
The 2016 Election - How and why it's President TrumpThe 2016 Election - How and why it's President Trump
The 2016 Election - How and why it's President TrumpLangerResearch
 
A 2016 Election Post-Mortem: The ABC News/Washington Post Tracking Poll
A 2016 Election Post-Mortem: The ABC News/Washington Post Tracking PollA 2016 Election Post-Mortem: The ABC News/Washington Post Tracking Poll
A 2016 Election Post-Mortem: The ABC News/Washington Post Tracking PollLangerResearch
 
Impact Assessment: Bangladesh Leadership Development Program
Impact Assessment: Bangladesh Leadership Development ProgramImpact Assessment: Bangladesh Leadership Development Program
Impact Assessment: Bangladesh Leadership Development ProgramLangerResearch
 
The 2016 Elections: Exit Polls and Trumpismo
The 2016 Elections: Exit Polls and TrumpismoThe 2016 Elections: Exit Polls and Trumpismo
The 2016 Elections: Exit Polls and TrumpismoLangerResearch
 
Attitudes on Climate Change: Expressed Belief and Preference x Priority
Attitudes on Climate Change: Expressed Belief and Preference x PriorityAttitudes on Climate Change: Expressed Belief and Preference x Priority
Attitudes on Climate Change: Expressed Belief and Preference x PriorityLangerResearch
 
The Sleep Health Index
The Sleep Health IndexThe Sleep Health Index
The Sleep Health IndexLangerResearch
 
Correlations: Bloomberg CCI and Other Indicators
Correlations: Bloomberg CCI and Other IndicatorsCorrelations: Bloomberg CCI and Other Indicators
Correlations: Bloomberg CCI and Other IndicatorsLangerResearch
 
connectedness and continuity: a prescription for patient engagement
connectedness and continuity: a prescription for patient engagementconnectedness and continuity: a prescription for patient engagement
connectedness and continuity: a prescription for patient engagementLangerResearch
 
Cli presentation 6 19-15-web
Cli presentation 6 19-15-webCli presentation 6 19-15-web
Cli presentation 6 19-15-webLangerResearch
 
Polling the 2014 Election: A Comparison of the ABC/Post Poll and the NEP Exit...
Polling the 2014 Election: A Comparison of the ABC/Post Poll and the NEP Exit...Polling the 2014 Election: A Comparison of the ABC/Post Poll and the NEP Exit...
Polling the 2014 Election: A Comparison of the ABC/Post Poll and the NEP Exit...LangerResearch
 
AAPOR 2015 afghan elections
AAPOR 2015 afghan electionsAAPOR 2015 afghan elections
AAPOR 2015 afghan electionsLangerResearch
 
AAPOR 2015 election presentation
AAPOR 2015 election presentationAAPOR 2015 election presentation
AAPOR 2015 election presentationLangerResearch
 
AAPOR_2015_OpportunitySurvey
AAPOR_2015_OpportunitySurveyAAPOR_2015_OpportunitySurvey
AAPOR_2015_OpportunitySurveyLangerResearch
 
Building Better Healthcare for Low-income Californians: Results on Shared Dec...
Building Better Healthcare for Low-income Californians: Results on Shared Dec...Building Better Healthcare for Low-income Californians: Results on Shared Dec...
Building Better Healthcare for Low-income Californians: Results on Shared Dec...LangerResearch
 
Afghan Election Fed Forum-June 6 2014
Afghan Election Fed Forum-June 6 2014Afghan Election Fed Forum-June 6 2014
Afghan Election Fed Forum-June 6 2014LangerResearch
 
Exploring Consumer Sentiment in Predictive Modeling
Exploring Consumer Sentiment in Predictive ModelingExploring Consumer Sentiment in Predictive Modeling
Exploring Consumer Sentiment in Predictive ModelingLangerResearch
 
Afghan Election Aapor 2014_final
Afghan Election Aapor 2014_finalAfghan Election Aapor 2014_final
Afghan Election Aapor 2014_finalLangerResearch
 
NYAAPOR: Question Wording and Questionnaire Design
NYAAPOR: Question Wording and Questionnaire DesignNYAAPOR: Question Wording and Questionnaire Design
NYAAPOR: Question Wording and Questionnaire DesignLangerResearch
 

More from LangerResearch (20)

Correlations: Bloomberg CCI and Other Indicators
Correlations: Bloomberg CCI and Other IndicatorsCorrelations: Bloomberg CCI and Other Indicators
Correlations: Bloomberg CCI and Other Indicators
 
Predicting State-Level 2016 Vote using MRP and a National Tracking Poll
Predicting State-Level 2016 Vote using MRP and a National Tracking PollPredicting State-Level 2016 Vote using MRP and a National Tracking Poll
Predicting State-Level 2016 Vote using MRP and a National Tracking Poll
 
The 2016 Election - How and why it's President Trump
The 2016 Election - How and why it's President TrumpThe 2016 Election - How and why it's President Trump
The 2016 Election - How and why it's President Trump
 
A 2016 Election Post-Mortem: The ABC News/Washington Post Tracking Poll
A 2016 Election Post-Mortem: The ABC News/Washington Post Tracking PollA 2016 Election Post-Mortem: The ABC News/Washington Post Tracking Poll
A 2016 Election Post-Mortem: The ABC News/Washington Post Tracking Poll
 
Impact Assessment: Bangladesh Leadership Development Program
Impact Assessment: Bangladesh Leadership Development ProgramImpact Assessment: Bangladesh Leadership Development Program
Impact Assessment: Bangladesh Leadership Development Program
 
The 2016 Elections: Exit Polls and Trumpismo
The 2016 Elections: Exit Polls and TrumpismoThe 2016 Elections: Exit Polls and Trumpismo
The 2016 Elections: Exit Polls and Trumpismo
 
Attitudes on Climate Change: Expressed Belief and Preference x Priority
Attitudes on Climate Change: Expressed Belief and Preference x PriorityAttitudes on Climate Change: Expressed Belief and Preference x Priority
Attitudes on Climate Change: Expressed Belief and Preference x Priority
 
The Sleep Health Index
The Sleep Health IndexThe Sleep Health Index
The Sleep Health Index
 
Correlations: Bloomberg CCI and Other Indicators
Correlations: Bloomberg CCI and Other IndicatorsCorrelations: Bloomberg CCI and Other Indicators
Correlations: Bloomberg CCI and Other Indicators
 
connectedness and continuity: a prescription for patient engagement
connectedness and continuity: a prescription for patient engagementconnectedness and continuity: a prescription for patient engagement
connectedness and continuity: a prescription for patient engagement
 
Cli presentation 6 19-15-web
Cli presentation 6 19-15-webCli presentation 6 19-15-web
Cli presentation 6 19-15-web
 
Polling the 2014 Election: A Comparison of the ABC/Post Poll and the NEP Exit...
Polling the 2014 Election: A Comparison of the ABC/Post Poll and the NEP Exit...Polling the 2014 Election: A Comparison of the ABC/Post Poll and the NEP Exit...
Polling the 2014 Election: A Comparison of the ABC/Post Poll and the NEP Exit...
 
AAPOR 2015 afghan elections
AAPOR 2015 afghan electionsAAPOR 2015 afghan elections
AAPOR 2015 afghan elections
 
AAPOR 2015 election presentation
AAPOR 2015 election presentationAAPOR 2015 election presentation
AAPOR 2015 election presentation
 
AAPOR_2015_OpportunitySurvey
AAPOR_2015_OpportunitySurveyAAPOR_2015_OpportunitySurvey
AAPOR_2015_OpportunitySurvey
 
Building Better Healthcare for Low-income Californians: Results on Shared Dec...
Building Better Healthcare for Low-income Californians: Results on Shared Dec...Building Better Healthcare for Low-income Californians: Results on Shared Dec...
Building Better Healthcare for Low-income Californians: Results on Shared Dec...
 
Afghan Election Fed Forum-June 6 2014
Afghan Election Fed Forum-June 6 2014Afghan Election Fed Forum-June 6 2014
Afghan Election Fed Forum-June 6 2014
 
Exploring Consumer Sentiment in Predictive Modeling
Exploring Consumer Sentiment in Predictive ModelingExploring Consumer Sentiment in Predictive Modeling
Exploring Consumer Sentiment in Predictive Modeling
 
Afghan Election Aapor 2014_final
Afghan Election Aapor 2014_finalAfghan Election Aapor 2014_final
Afghan Election Aapor 2014_final
 
NYAAPOR: Question Wording and Questionnaire Design
NYAAPOR: Question Wording and Questionnaire DesignNYAAPOR: Question Wording and Questionnaire Design
NYAAPOR: Question Wording and Questionnaire Design
 

Recently uploaded

Top 10 Wealthiest People In The World.pdf
Top 10 Wealthiest People In The World.pdfTop 10 Wealthiest People In The World.pdf
Top 10 Wealthiest People In The World.pdfauroraaudrey4826
 
Brief biography of Julius Robert Oppenheimer
Brief biography of Julius Robert OppenheimerBrief biography of Julius Robert Oppenheimer
Brief biography of Julius Robert OppenheimerOmarCabrera39
 
57 Bidens Annihilation Nation Policy.pdf
57 Bidens Annihilation Nation Policy.pdf57 Bidens Annihilation Nation Policy.pdf
57 Bidens Annihilation Nation Policy.pdfGerald Furnkranz
 
Global Terrorism and its types and prevention ppt.
Global Terrorism and its types and prevention ppt.Global Terrorism and its types and prevention ppt.
Global Terrorism and its types and prevention ppt.NaveedKhaskheli1
 
HARNESSING AI FOR ENHANCED MEDIA ANALYSIS A CASE STUDY ON CHATGPT AT DRONE EM...
HARNESSING AI FOR ENHANCED MEDIA ANALYSIS A CASE STUDY ON CHATGPT AT DRONE EM...HARNESSING AI FOR ENHANCED MEDIA ANALYSIS A CASE STUDY ON CHATGPT AT DRONE EM...
HARNESSING AI FOR ENHANCED MEDIA ANALYSIS A CASE STUDY ON CHATGPT AT DRONE EM...Ismail Fahmi
 
Chandrayaan 3 Successful Moon Landing Mission.pdf
Chandrayaan 3 Successful Moon Landing Mission.pdfChandrayaan 3 Successful Moon Landing Mission.pdf
Chandrayaan 3 Successful Moon Landing Mission.pdfauroraaudrey4826
 
Dynamics of Destructive Polarisation in Mainstream and Social Media: The Case...
Dynamics of Destructive Polarisation in Mainstream and Social Media: The Case...Dynamics of Destructive Polarisation in Mainstream and Social Media: The Case...
Dynamics of Destructive Polarisation in Mainstream and Social Media: The Case...Axel Bruns
 
Opportunities, challenges, and power of media and information
Opportunities, challenges, and power of media and informationOpportunities, challenges, and power of media and information
Opportunities, challenges, and power of media and informationReyMonsales
 
Manipur-Book-Final-2-compressed.pdfsal'rpk
Manipur-Book-Final-2-compressed.pdfsal'rpkManipur-Book-Final-2-compressed.pdfsal'rpk
Manipur-Book-Final-2-compressed.pdfsal'rpkbhavenpr
 
complaint-ECI-PM-media-1-Chandru.pdfra;;prfk
complaint-ECI-PM-media-1-Chandru.pdfra;;prfkcomplaint-ECI-PM-media-1-Chandru.pdfra;;prfk
complaint-ECI-PM-media-1-Chandru.pdfra;;prfkbhavenpr
 
AP Election Survey 2024: TDP-Janasena-BJP Alliance Set To Sweep Victory
AP Election Survey 2024: TDP-Janasena-BJP Alliance Set To Sweep VictoryAP Election Survey 2024: TDP-Janasena-BJP Alliance Set To Sweep Victory
AP Election Survey 2024: TDP-Janasena-BJP Alliance Set To Sweep Victoryanjanibaddipudi1
 
Quiz for Heritage Indian including all the rounds
Quiz for Heritage Indian including all the roundsQuiz for Heritage Indian including all the rounds
Quiz for Heritage Indian including all the roundsnaxymaxyy
 
Referendum Party 2024 Election Manifesto
Referendum Party 2024 Election ManifestoReferendum Party 2024 Election Manifesto
Referendum Party 2024 Election ManifestoSABC News
 
N Chandrababu Naidu Launches 'Praja Galam' As Part of TDP’s Election Campaign
N Chandrababu Naidu Launches 'Praja Galam' As Part of TDP’s Election CampaignN Chandrababu Naidu Launches 'Praja Galam' As Part of TDP’s Election Campaign
N Chandrababu Naidu Launches 'Praja Galam' As Part of TDP’s Election Campaignanjanibaddipudi1
 
Different Frontiers of Social Media War in Indonesia Elections 2024
Different Frontiers of Social Media War in Indonesia Elections 2024Different Frontiers of Social Media War in Indonesia Elections 2024
Different Frontiers of Social Media War in Indonesia Elections 2024Ismail Fahmi
 
VIP Girls Available Call or WhatsApp 9711199012
VIP Girls Available Call or WhatsApp 9711199012VIP Girls Available Call or WhatsApp 9711199012
VIP Girls Available Call or WhatsApp 9711199012ankitnayak356677
 

Recently uploaded (16)

Top 10 Wealthiest People In The World.pdf
Top 10 Wealthiest People In The World.pdfTop 10 Wealthiest People In The World.pdf
Top 10 Wealthiest People In The World.pdf
 
Brief biography of Julius Robert Oppenheimer
Brief biography of Julius Robert OppenheimerBrief biography of Julius Robert Oppenheimer
Brief biography of Julius Robert Oppenheimer
 
57 Bidens Annihilation Nation Policy.pdf
57 Bidens Annihilation Nation Policy.pdf57 Bidens Annihilation Nation Policy.pdf
57 Bidens Annihilation Nation Policy.pdf
 
Global Terrorism and its types and prevention ppt.
Global Terrorism and its types and prevention ppt.Global Terrorism and its types and prevention ppt.
Global Terrorism and its types and prevention ppt.
 
HARNESSING AI FOR ENHANCED MEDIA ANALYSIS A CASE STUDY ON CHATGPT AT DRONE EM...
HARNESSING AI FOR ENHANCED MEDIA ANALYSIS A CASE STUDY ON CHATGPT AT DRONE EM...HARNESSING AI FOR ENHANCED MEDIA ANALYSIS A CASE STUDY ON CHATGPT AT DRONE EM...
HARNESSING AI FOR ENHANCED MEDIA ANALYSIS A CASE STUDY ON CHATGPT AT DRONE EM...
 
Chandrayaan 3 Successful Moon Landing Mission.pdf
Chandrayaan 3 Successful Moon Landing Mission.pdfChandrayaan 3 Successful Moon Landing Mission.pdf
Chandrayaan 3 Successful Moon Landing Mission.pdf
 
Dynamics of Destructive Polarisation in Mainstream and Social Media: The Case...
Dynamics of Destructive Polarisation in Mainstream and Social Media: The Case...Dynamics of Destructive Polarisation in Mainstream and Social Media: The Case...
Dynamics of Destructive Polarisation in Mainstream and Social Media: The Case...
 
Opportunities, challenges, and power of media and information
Opportunities, challenges, and power of media and informationOpportunities, challenges, and power of media and information
Opportunities, challenges, and power of media and information
 
Manipur-Book-Final-2-compressed.pdfsal'rpk
Manipur-Book-Final-2-compressed.pdfsal'rpkManipur-Book-Final-2-compressed.pdfsal'rpk
Manipur-Book-Final-2-compressed.pdfsal'rpk
 
complaint-ECI-PM-media-1-Chandru.pdfra;;prfk
complaint-ECI-PM-media-1-Chandru.pdfra;;prfkcomplaint-ECI-PM-media-1-Chandru.pdfra;;prfk
complaint-ECI-PM-media-1-Chandru.pdfra;;prfk
 
AP Election Survey 2024: TDP-Janasena-BJP Alliance Set To Sweep Victory
AP Election Survey 2024: TDP-Janasena-BJP Alliance Set To Sweep VictoryAP Election Survey 2024: TDP-Janasena-BJP Alliance Set To Sweep Victory
AP Election Survey 2024: TDP-Janasena-BJP Alliance Set To Sweep Victory
 
Quiz for Heritage Indian including all the rounds
Quiz for Heritage Indian including all the roundsQuiz for Heritage Indian including all the rounds
Quiz for Heritage Indian including all the rounds
 
Referendum Party 2024 Election Manifesto
Referendum Party 2024 Election ManifestoReferendum Party 2024 Election Manifesto
Referendum Party 2024 Election Manifesto
 
N Chandrababu Naidu Launches 'Praja Galam' As Part of TDP’s Election Campaign
N Chandrababu Naidu Launches 'Praja Galam' As Part of TDP’s Election CampaignN Chandrababu Naidu Launches 'Praja Galam' As Part of TDP’s Election Campaign
N Chandrababu Naidu Launches 'Praja Galam' As Part of TDP’s Election Campaign
 
Different Frontiers of Social Media War in Indonesia Elections 2024
Different Frontiers of Social Media War in Indonesia Elections 2024Different Frontiers of Social Media War in Indonesia Elections 2024
Different Frontiers of Social Media War in Indonesia Elections 2024
 
VIP Girls Available Call or WhatsApp 9711199012
VIP Girls Available Call or WhatsApp 9711199012VIP Girls Available Call or WhatsApp 9711199012
VIP Girls Available Call or WhatsApp 9711199012
 

Probability Sampling and Alternative Methodologies

  • 1. Probability Sampling and Alternative Methodologies “The Future of Survey Research” National Science Foundation Oct. 4, 2012 Gary Langer Langer Research Associates glanger@langerresearch.com @Langer Research
  • 2.
  • 3. Good Data • Are powerful and compelling • Rise above anecdote • Sustain precision • Expand our knowledge, enrich our understanding and inform our judgment
  • 4. Other Data • May be manufactured to promote a product, client or point of view • Are easily manipulated • Are increasingly prevalent (cheaply produced via Internet, e-mail) • Leave the house of inferential statistics • Lack the validity and reliability we seek • Often mis-ask and misanalyze • Misinform, even disinform our judgment
  • 5. The Challenge • Subject established methods to rigorous and ongoing evaluation • Subject new methods to the same standards • Value theoreticism as much as empiricism • Consider fitness for purpose • Go with the facts, not with the fashion
  • 6. Survey Sampling A brief history…
  • 7. Sample Techniques pre-1936 • Full population (census survey) • Pros: High level of accuracy • Cons: Prohibitively expensive and highly challenging (unless the population is small) • “Availability” sampling (e.g., straw polls) • Pros: Quick and inexpensive • Cons: No-scientific basis, no theoretical justification for generalizing beyond the sample • Justification: “Good enough” – Literary Digest correctly predicted presidential election winners from1916 to 1932
  • 8. The 1936 Election Roosevelt (D) vs. Landon (R) • Literary Digest poll • Technique: An availability sample • Sample frame: Combination of magazine subscribers, phone books, and car registration lists • Contacted 10 million, 2.4 million responded • Found an easy win for Landon. Roosevelt won in a landslide. • What happened? • Coverage bias: The sampling frame differed in important ways from the population of interest. Republicans were more likely to afford magazine subscriptions, telephones and cars during the Great Depression. • Systematic nonresponse: Republicans were disproportionately likely to return their ballot (perhaps to express discontent with the New Deal).
  • 9. On to quota sampling • After 1936, availability sampling largely was replaced with quota sampling (which had correctly predicted FDR‟s win). • Quota sampling essentially attempts to build a “miniature” version of the target population. • Identifies supposed key demographic characteristics in the population and attempts to mirror them in the sample. • Matching the sample to the population on variables such as gender, race, age, income and others was said to produce a “representative” sample.
  • 11. What went wrong? 1. Timing: Pollsters stopped collecting data in mid-October, assuming that there would be no late-breaking changes. 2. Faulty assumptions: Pollsters erroneously assumed that undecided voters would vote in proportion with those who had made up their minds. Likely voters also may have been misidentified. 3. Sampling problems: Quota sampling may have allowed interviewers to select more-educated and better-off respondents within their assigned quotas – pro-Dewey groups. This is a key potential flaw in purposive sampling.
  • 12. Quota Sampling Issues • Even if quota sampling were not entirely to blame, the Dewey-Truman fiasco highlighted inherent limitations in quota sampling • It‟s impossible to anticipate, let alone match, every potentially important demographic: • e.g., if you match on sex (2), ethnicity (2), race (5), education (4), age (5), region (4) and income (6) that yields 9,600 cells. If you want to interview more than one person per cell… • Respondent selection is left up to the interviewers, allowing human bias to sneak in (e.g., picking easy-to-reach resps). • As with availability sampling, there is no theoretical justification for drawing inferences about the greater population.
  • 14. Not a new concept A “criterion that a sample design should meet (at least if one is to make important decisions on the basis of the sample results) is that the reliability of the sample results should be susceptible of measurement. An essential feature of such sampling methods is that each element of the population being sampled … has a chance of being included in the sample and, moreover, that that chance or probability is known.” Hansen and Hauser, POQ, 1945
  • 15. This is really not a new concept “Diagoras, surnamed the Atheist, once paid a visit to Samothrace, and a friend of his addressed him thus: „You believe that the gods have no interest in human welfare. Please observe these countless painted tablets; they show how many persons have withstood the rage of the tempest and safely reached the haven because they made vows to the gods.‟ “„Quite so,‟ Diagoras answered, „but where are the tablets of those who suffered shipwreck and perished in the deep?‟” “On the Nature of the Gods,” Marcus Tullius Cicero, 45 B.C. (cited by Kruskal and Mosteller, International Statistical Review, 1979)
  • 16. A shared view • “…the stratified random sample is recognized by mathematical statisticians as the only practical device now available for securing a representative sample from human populations…” Snedecor, Journal of Farm Economics, 1939 • “It is obvious that the sample can be representative of the population only if all parts of the population have a chance of being sampled.” Tippett, The Methods of Statistics, 1952 • “If the sample is to be representative of the population from which it is drawn, the elements of the sample must have been chosen at random.” Johnson and Jackson, Modern Statistical Methods, 1959 • “Probability sampling is important for three reasons: (1) Its measurability leads to objective statistical inference, in contrast to the subjective inference from judgment sampling. (2) Like any scientific method, it permits cumulative improvement through the separation and objective appraisal of its sources of errors. (3) When simple methods fail, researchers turn to probability sampling…” Kish, Survey Sampling, 1965
  • 18. Copyright 2003 Times Newspapers Limited Sunday Times (London) January 26, 2003, Sunday SECTION: Features; News; 12 LENGTH: 312 words HEADLINE: Blair fails to make a case for action that convinces public BODY: Tony Blair has failed to convince people in Britain of the need for war with Iraq, a poll for The Sunday Times shows. Even among Labour supporters, the prime minister has not yet made the case. The poll of nearly 2,000 people by YouGov shows that only 26% say Blair has convinced them that Saddam Hussein is sufficiently dangerous to justify military action, against 68% who say he has not done so.
  • 19.
  • 20.
  • 21. -----Original Message----- From: Students SM Team [mailto:alumteam@teams.com] Sent: Wednesday, October 04, 2006 11:27 AM Subject: New Job opening Hi, Going to school requires a serious commitment, but most students still need extra money for rent, food, gas, books, tuition, clothes, pleasure and a whole list of other things. So what do you do? "Find some sort of work", but the problem is that many jobs are boring, have low pay and rigid/inflexible schedules. So you are in the middle of mid-terms and you need to study but you have to be at work, so your grades and education suffer at the expense of your "College Job". Now you can do flexible work that fits your schedule! Our company and several nationwide companies want your help. We are looking to expand, by using independent workers we can do so without buying additional buildings and equipment. You can START IMMEDIATELY! This type of work is Great for College and University Students who are seriously looking for extra income! We have compiled and researched hundreds of research companies that are willing to pay you between $5 and $75 per hour simply to answer an online survey in the peacefulness of your own home. That's all there is to it, and there's no catch or gimmicks! We've put together the most reliable and reputable companies in the industry. Our list of research companies will allow you to earn $5 to $75 filling out surveys on the internet from home. One hour focus groups will earn you $50 to $150. It's as simple as that. Our companies just want you to give them your opinion so that they can empower their own market research. Since your time is valuable, they are willing to pay you for it. If you want to apply for the job position, please email at: job2@alum.com Students SM Team
  • 22. -----Original Message----- From: Ipsos News Alerts [mailto:newsalerts@ipsos-na.com] Sent: Friday, March 27, 2009 5:12 PM To: Langer, Gary Subject: McLeansville Mother Wins a Car By Taking Surveys McLeansville Mother Wins a Car By Taking Surveys Toronto, ON- McLeansville, NC native, Jennifer Gattis beats the odds and wins a car by answering online surveys. Gattis was one of over 105 300 North Americans eligible to win. Representatives from Ipsos i-Say, a leading online market research panel will be in Greensboro on Tuesday, March 31, 2009 to present Gattis with a 2009 Toyota Prius. Access the full press release at: http://www.ipsos-na.com/news/pressrelease.cfm?id=4331
  • 23. Opt-in online panelist 32-year-old Spanish-speaking female African-American physician residing in Billings, MT
  • 24.
  • 25. Professional Respondents? Among 10 largest opt-in panels: •10% of panel participants account for 81% of survey responses; • 1% of participants account for 34% of responses. Gian Fulgoni, chairman, comScore, Council of American Survey Research Organizations annual conference, Los Angeles, October 2006.
  • 26. Questions  Who joins the club, how and why?  What verification and validation of respondent identities are undertaken?  What logical and QC checks (duration, patterning, data quality) are applied?  What weights are applied, and how? On what theoretical basis and with what effect?  What level of disclosure is provided?  What claims are made about the data, and how are they justified?
  • 27. One Claim: Google Consumer Surveys • “Produces results that are as accurate as probability-based panels.” http://www.google.com/insights/consumersurveys/how • 1 or 2-question pop-ups delivered to search engine users of “premium content.” • Demographic data are imputed through analysis of users‟ IP addresses and previous page views • Among Langer Research staff, one woman was identified as a 55-year-old male with an interest in beauty and fitness; another as a 65+ woman (double her actual age) with an interest in motorcycles and one man as a senior citizen in the Pacific Northwest with an interest in space technology. https://www.google.com/settings/ads/onweb/
  • 28. Another claim: “Bayesian Credibility Interval” • Ipsos opt-in online panel – presdential tracking poll • Computation of “Credibility Internval” appears to match how you‟d compute SRS MoE. • Its basis is “the opinion of Ipsos experts.” • (Headdesk)
  • 29. Further claims: Convenience Sample MoE • Zogby Interactive: "The margin of error is +/- 0.6 percentage points.” • Ipsos/Reuters: “The margin of error is plus or minus 3.1 percentage points." • Kelton Research: “The survey results indicate a margin of error of +/- 3.1 percent at a 95 percent confidence level.” • Economist/YouGov/Polimetrix: “Margin of error: +/- 4%.” • PNC/HNW/Harris Interactive: “Findings are significant at the 95 percent confidence level with a margin of error of +/- 2.5 percent.” • Radio One/Yankelovich: “Margin of error: +/-2 percentage points.” • Citi Credit-ED/Synovate: “The margin of error is +/- 3.0 percentage points.” • Spectrem: “The data have a margin of error of plus or minus 6.2 percentage points.” • Luntz: “+3.5% margin of error”
  • 30. Online Survey Traditional phone-based survey techniques suffer from deteriorating response rates and escalating costs. YouGovPolimetrix combines technology infrastructure for data collection, integration with large scale databases, and novel instrumentation to deliver new capabilities for polling and survey research. (read more)
  • 31. Response Rates • Surveys based in probability sampling cannot achieve pure probability (e.g., 100% RR) • Pew (5/15/12) reports decline in its RRs from 36% in 1997 to 9% now. (ABC/Post: 19%) • Does this trend poison the well?
  • 32. A Look at the Lit 2006: "…little to suggest that unit nonresponse within the range of response rates obtained seriously threatens the quality of survey estimates." Keeter et al., POQ, 2006 – comp. RR 25 vs. 50%; see also Keeter et al., POQ, 2000, RR 36 vs. 61 2008: “In general population RDD telephone surveys, lower response rates do not notably reduce the quality of survey demographic estimates. … This evidence challenges the assumption that response rates are a key indicator of survey data quality…” Holbrook et al.: Demographic comparison, 81 RDD surveys, 1996-2005; AAPOR3 RRs from 5 to 54% 2012: “Overall, there are only modest differences in responses between the standard and high-effort surveys. Similar to 1997 and 2003, the additional time and effort to encourage cooperation in the high-effort survey does not lead to significantly different estimates on most questions.” Pew 5/15/12 - Comp. RR 9 vs. 22%
  • 33. See also O‟Neil, 1979; Smith, 1984; Merkle et al., 1993; Curtin et al., 2000; Groves et al., 2004; Curtin et al., 2005. And Groves, 2006: “Hence, there is little empirical support for the notion that low response rate surveys de facto produce estimates with high nonresponse bias.”
  • 34.
  • 36. Yeager, Krosnick, et al., 2011 (See also Malhotra and Krosnick, 2006) • Paper compares seven opt-in online convenience-sample surveys with two probability sample surveys • Probability-sample surveys were “consistently highly accurate” • Opt-in online surveys were “always less accurate… and less consistent in their level of accuracy.” • Weighted probability samples sig. diff. from benchmarks 31 or 46 percent of the time for probability samples, 62 to 77 percent of the time for convenience samples. • Opt-in data‟s highest single error vs. benchmarks, 19 points; average error, 9.9.
  • 37. Average Absolute Errors primary (weighting) demographics, unweighted 20% 18% 16% 14% 12.0% 12% 10% 8% 6.4% 6.4% 6% 5.0% 5.3% 4.7% 4.1% 4% 3.3% 2.0% 2% 0% Yeager, Krosnick, et al., 2011
  • 38. Average Absolute Errors secondary and non-demographic results 10% Unweighted Weighted 9% 8% 7% 5.9% 6% 5.2% 5% 3.8% 4% 3.2% 3% 2% 1% 0% Probability samples (combined) Internet samples (combined) Yeager, Krosnick, et al., 2011
  • 39. Largest Absolute Error all benchmarks, unweighted 40% 35.5% 35% 30% 25% 20% 18.0% 15.6% 15.3% 16.0% 15% 13.2% 13.7% 11.7% 9.6% 10% 5% 0% Yeager, Krosnick, et al., 2011
  • 40. Largest Absolute Error secondary and non-demographic results 20% Unweighted Weighted 18% 16% 14.9% 14% 13.5% 12% 10.7% 10% 8.7% 8% 6% 4% 2% 0% Probability samples (combined) Internet samples (combined) Yeager, Krosnick, et al., 2011
  • 41. % significantly different from benchmarks secondary and non-demographic results 100% Unweighted Weighted 90% 80% 70.6% 71.4% 70% 60% 57.5% 50% 38.5% 40% 30% 20% 10% 0% Probability samples (combined) Internet samples (combined) Yeager, Krosnick, et al., 2011
  • 42. Additional conclusions Yeager, Krosnick, et al., 2011 • Little support for the claim that some non-probability internet surveys are consistently more accurate than others. • Weighting did not always improve the accuracy of the opt-in samples. • No support for the idea that higher completion = greater accuracy: • There was a significant negative correlation between a survey‟s average absolute error and its response rate. • Probability samples were not only more accurate overall, but also more consistently accurate. • True both across surveys (using pre-existing probability samples as comparison points) as well as within surveys. • Meaning it‟s virtually impossible to anticipate whether an opt-in survey will be just somewhat less accurate than a probability sample or substantially less accurate; and knowing an opt-in is accurate on one benchmark does not predict its accuracy on other measures.
  • 43. Additional conclusions Yeager, Krosnick, et al., 2011 • The fact that internet samples are less accurate, and less consistent, on average, than probability samples should “come as no surprise, because no theory provides a rationale whereby samples generated by non-probability methods would yield accurate results” • It is “possible to cherry-pick such results to claim that non- probability sampling can yield veridical measurements. But a systematic look at a wide array of benchmarks documented that such results are the exception rather than the rule.”
  • 44. ARF FoQ via Reg Baker, 2009 • Reported data on estimates of smoking prevalence: similar across three probability methods, but with as many as 14 points of variation across 17 opt-in online panels. • “In the end, the results we get for any given study are highly dependent (and mostly unpredictable) on the panel we use. This is not good news.”
  • 45. Callegaro et al., 2012 • Reviewed 45 studies comparing the quality of data collected via opt-in panels vs. another mode and/or benchmarks. In summary: • Panel estimates substantially deviated from benchmarks, and to a far greater degree than probability samples • High variability in data from different opt-in panels • High levels of multiple-panel membership (between 19-45% of respondents belong to 5+ panels) • Substantial differences among low- and higher-membership respondents • Weighting did not correct variations in the data
  • 46. Traugott, 2012 • Tested the accuracy of four low cost data collection (LCDC) methods (Mechanical Turk, Pulse Opinion Research, Qualtrics and Zoomerang) • LCDC unweighted demographics substantially deviated from ACS benchmarks to a far greater degree than RDD polls or the ANES • MTurk was substantially younger (45% under 30), female (60%), and more educated (47% with a college degree) than benchmarks. • Pulse was disproportionately older (43% over 65), female (61%), white (90%) and educated (46% with a college degree). • Qualtrics and Zoomerang had better unweighted age, sex, and race distributions, but still had far too few respondents without a high school degree and too many with some college or more.
  • 47. Traugott, 2012 (cont‟d) 100% Unweighted Party ID 90% Republican Democrat Independent 80% 70% 60% 57% 49% 50% 44% 41% 42% 40% 35% 37% 36% 31% 27% 28% 30% 20% 13% 13% 11% 10% 1% 0% ANES Mturk Zoomerang Pulse Qualtrics
  • 48. Traugott, 2012 (cont‟d) 100% Weighted Party ID 90% Republican Democrat Independent 80% 70% 58% 60% 50% 43% 44% 45% 46% 40% 37% 34% 34% 31% 30% 26% 26% 20% 12% 13% 13% 10% 2% 0% ANES Mturk Zoomerang Pulse Qualtrics
  • 49. AAPOR‟s “Report on Online Panels,” April 2010 • “Researchers should avoid nonprobability online panels when one of the research objectives is to accurately estimate population values.” • “The nonprobability character of volunteer online panels … violates the underlying principles of probability theory.” • “Empirical evaluations of online panels abroad and in the U.S. leave no doubt that those who choose to join online panels differ in important and nonignorable ways from those who do not.” • “In sum, the existing body of evidence shows that online surveys with nonprobability panels elicit systematically different results than probability sample surveys in a wide variety of attitudes and behaviors.” • “The reporting of a margin of sampling error associated with an opt-in sample is misleading.”
  • 50. AAPOR‟s conclusion “There currently is no generally accepted theoretical basis from which to claim that survey results using samples from nonprobability online panels are projectable to the general population. Thus, claims of „representativeness‟ should be avoided when using these sample sources.” AAPOR Report on Online Panels, April 2010
  • 51. OK, but apart from population values, we can still use convenience samples to evaluate relationships among variables and trends over time. Right?
  • 52. Pasek & Krosnick, 2010 • Comparison of opt-in online and RDD surveys sponsored by the U.S. Census Bureau assessing intent to fill out the Census. • “The telephone samples were more demographically representative of the nation‟s population than were the Internet samples, even after post-stratification.” • “The distributions of opinions and behaviors were often significantly and substantially different across the two data streams. Thus, research conclusions would often be different.” • Instances “where the two data streams told very different stories about change over time … over-time trends in one line did not meaningfully covary with over-time trends in the other line.”
  • 53. Top Predictors of Intent to Complete the Census In the probability RDD sample In the internet opt-in sample Age: 18-24 -1.76 Age: 18-24 -1.81 Education: <H.S. -1.16 Age: 25-44 -1.29 Education: H.S. graduate -.97 Think Census can help them 1.19 Spanish speaking household .97 Think Census can harm them -.98 Counting everyone is important .92 Participation doesn‟t matter .95 Age: 25-44 -.76 (disagree) Don‟t have time (disagree) .70 Counting everyone is important .71 Education: Some college -.67 Age: 45-64 -.62 Participation doesn‟t matter .67 Don‟t have time (disagree) .61 (disagree) Race: Hispanic .56 Counting everyone is -.62 Education: <H.S. -.47 important (disagree) Pasek & Krosnick, 2010
  • 54. Biggest Differences in Predictors of Intent to Complete the Census RDD Opt-In Diff. Spanish speaking household .97*** -.38 -1.34*** Importance of counting -.62*** .14 .76** everyone (disagree) Education: <H.S. -1.16*** -.47* .69** Education: H.S. graduate -.97*** -.38 .59** Think Census can help them .61*** 1.19*** .57*** Age: 25-44 -.76*** -1.29*** -.54** Age: 45-64 -.21* -.62*** -.42* Region: Northeast -.06 .27* .33* Pasek & Krosnick, 2010
  • 55. Differences in Time Trend Pasek & Krosnick, 2010
  • 56. Pasek/Krosnick conclusion “This investigation revealed systematic and often sizable differences between probability sample telephone data and non-probability Internet data in terms of demographic representativeness of the samples, the proportion of respondents reporting various opinions and behaviors, the predictors of intent to complete the Census form and actual completion of the form, changes over time in responses, and relations between variables.” Pasek and Krosnick, 2010
  • 57. Can non-probability samples be „fixed‟? • Bayesian analysis? • What variables? How derived? Are we weighting to the DV or its correlates? • Sample balancing? (i.e., quota sampling revisited) • “The microcosm idea will rarely work in a complicated social problem because we always have additional variables that may have important consequences for the outcome.” • Gilbert, Light and Mosteller, Statistics and Public Policy, 1977
  • 58. Some say you can “The survey was administered by YouGovPolimetrix during July 16-July 26, 2008. YouGovPolimetrix employs sample matching techniques to build representative web samples through its pool of opt-in respondents (see Rivers 2008). Studies that use representative samples yielded this way find that their quality meets, and sometimes exceeds, the quality of samples yielded through more traditional survey techniques.” Perez, Political Behavior, 2010
  • 59. AAPOR‟s task force disagrees “There currently is no generally accepted theoretical basis from which to claim that survey results using samples from nonprobability online panels are projectable to the general population. Thus, claims of „representativeness‟ should be avoided when using these sample sources.” AAPOR Report on Online Panels, 2010
  • 60. It has company • “Unfortunately, convenience samples are often used inappropriately as the basis for inference to some larger population.” • “…unlike random samples, purposive samples contain no information on how close the sample estimate is to the true value of the population parameter.” • “Quota sampling suffers from essentially the same limitations as convenience, judgment, and purposive sampling (i.e., it has no probabilistic basis for statistical inference).” • “Some variations of quota sampling contain elements of random sampling but are still not statistically valid methods.” Biemer and Lyberg, Introduction to Survey Quality, 2003
  • 62. Sampling Challenges • Users are not limited to one tweet/post per day; some may post incessantly; others rarely • Users can have multiple Twitter accounts • Accounts often do not belong to individuals, but rather companies, organizations or associations • Posts can be driven by a public relations campaign (i.e., through volunteers, paid agents or bots) rather than as an expression of individual attitudes
  • 63. Sampling Challenges (cont.) • Users‟ location is often not accurately captured (if at all) • Users are self-selected. There is no theoretical basis on which to assume that their views are representative of the views of non-users. • 15 percent of online adults use Twitter, 8 percent daily (Pew) • Fewer than 30 percent of Twitter users are Americans (Semiocast) • In 2011, 45% of U.S. Facebook users were 25 or younger (Smith, 2012)
  • 64. Sentiment Analysis • Determining the meaning of posts or tweets requires content analysis: • Traditional approaches (using independent coders and a codeback) typically is unrealistic given the volume of data • Computerized sentiment analysis programs, while lower-cost and faster, are fraught with problems: • Parsing slang, irony, sarcasm, abbreviations, acronyms, emoticons, contextual meaning and hashtags is highly complex • Ambiguity in determining target words (e.g., Obama vs. prez) • Regardless, contextual data (demographic, other attitudes) are sparse or absent, severely limiting the research questions that can be answered
  • 65. Manual vs. Automated Coding Manual Coding Positive Neutral Negative Uncodable (n=100) (n=285) (n=81) (n=34) Positive 8% 5% 2% 0% radian6 Neutral 86% 90% 83% 94% 55% Negative 0% 5% 15% 6% Positive 42% 25% 6% 21% STAS Neutral 45% 57% 71% 68% 45% Negative 13% 18% 23% 12% Positive 30% 20% 2% 9% clarabridge Neutral 60% 61% 85% 79% 43% Negative 10% 19% 12% 12% Kim et al., 2012
  • 66. Data Mining • Rather than attempting to parse the content of online posts, some researchers instead focus on the sheer quantity of posts about a certain topic. • Tumasjan et al. (2010) analyzed the German national election and found that “the mere number of tweets mentioning a political party” was quite good at reflecting their vote share and that its predictive power was similar to that of traditional polls. • At the same time, they found that only 4% of users accounted for more than 40% of the messages. • When such a small number of people are accounting for a large portion of the results, it opens the door for politically motivated actors to distort results intentionally – especially as the use of such methods becomes more popular.
  • 67. Data Mining • Similar analyses in the U.S. have differed • Lui et al. (2011) showed that “Google trends” was worse at predicting 2008 and 2010 election outcomes than the New York Times (probability-sample) polls and even chance. • Gayo-Avello et al. (2011) looked at the 2010 election and found the mean average error for Twitter volume was 17.1% (it was 7.6% for sentiment analysis), far higher than the 2- 3% that is typical of RDD polls. • Olson & Bunnett (2011) found that Facebook “likes” in 2010 explained 13% of the variance in Senate races, but were very weak (or slightly negative) predictors of gubernatorial and house races
  • 68. Twitter Index? The Twitter Political Index Aug. 6 – Obama‟s index # is three times greater than Romney‟s (“61” v. “24”) Aug. 7 – “35” v. “19”
  • 69. Facebook sampling? • Bhutta, 2012: Snowball sampling to reach Catholics • faster and cheaper than traditional methods, but: Facebook GSS 70% 67% 66% 53% 33% 24% 27% 6% % female % Latino % college grad. % attend mass weekly+
  • 70. “…the Facebook respondents cannot possibly serve as a representative sample of the general Catholic population. (Pollsters should view Facebook findings with extreme caution.)” Bhutta, 2012
  • 71. Twitter bots •Recent estimate that as many as 30% of Obama‟s Twitter followers and 22% of Romney‟s were fabricated. (Calzolari, 2012) •News report: Individuals can purchase 25,000 fabricated Twitter followers for $247; these “bots” will automatically tweet snippets of text. (Feifer, 2012)
  • 72. And the future? In probability sampling: • Continued study of response-rate effects • Concerted efforts to maintain response rates • Renewed focus on other data-quality issues in probability samples, e.g. coverage • Development of probability-based alternatives, e.g. mixed-mode, ABS
  • 73. The Future, cont. In convenience sampling: • Further study of appropriate uses (as well as inappropriate misuses) of convenience-sample data • Further evaluation of well-disclosed, emerging techniques in convenience sampling • Enhanced disclosure! • The quest for an online sampling frame
  • 74. The future, cont. In social media: • Vast and growing source of data • Further evaluation of its appropriate use is needed • It‟s important to study how social media influences the formation and change of public opinion. But that does not mean it can supplant probability-sample polling. • Information gleaned from social media may be useful as a compliment to public opinion polls (akin to focus groups); for trend-spotting, qualitative analysis, perhaps modeling • The key: establish the research question; evaluate the limitations of the tools available to answer it.
  • 75. Thank you! Gary Langer Langer Research Associates glanger@langerresearch.com

Editor's Notes

  1. After 1948, government, academics and pollsters basically all came to a consensus to adopt probability sampling…
  2. Note that neither availability or quota sampling meet these criteria
  3. prob can consolidate the RR issue a bit since Krosnick says someone else is deal with it. Will do.
  4. Average absolute error (unweighted) 3.3 and 3.5 points in probability samples vs. 4.9. to 9.9 points in convenience samplesHighest single error (weighted) 9 points in probability samples, 18 points in convenience samples
  5. primary demographics = sex, age, race/ethnicity, education, and regionaveraging across, it’s 2.6% for probability samples v. 6.3% for internet samples
  6. secondary demographics = marital status, total number of people living in the household, employment status, number of bedrooms in the home, number of vehicles owned, home ownership and household incomenon-demographic questions = frequency of smoking cigarettes, whether the have ever had 12 drinks of alcohol during their lifetimes, the average number of drinks of alcohol they have on days when they drink, ratings of quality of their health and possession of a US password and a driver’s license.
  7. primary demographics = sex, age, race/ethnicity, education, and region
  8. secondary demographics = marital status, total number of people living in the household, employment status, number of bedrooms in the home, number of vehicles owned, home ownership and household incomenon-demographic questions = frequency of smoking cigarettes, whether the have ever had 12 drinks of alcohol during their lifetimes, the average number of drinks of alcohol they have on days when they drink, ratings of quality of their health and possession of a US password and a driver’s license.
  9. These are betas from a multiple logistic regression. Bold indicates that the betas are significantly different across samples. So this shows that depending on what sample you use, you get a different list of top predictors. Even though not all predictors are significantly different, they appear in a different order so tell a different story.
  10. This shows it a different way. Here are the predictors that significantly differ by sample type. Again, these are betas from a logistic regression.
  11. Instances where the two samples told VERY different stories about change over time. The top three grafs show significant NEGATIVE correlations between the two over-time lines.However, there were cases where the data did correlate positively – so it’s important to mention that this discordance wasn’t always the case. Even still, the divergence is crazy!
  12. can discuss Google’s attempt to impute demographic information on the last point. I can make a full slide about it if you’d like.
  13. topic: Salviathis shows that the computer programs were pretty good at matching neutral tweets (well, at least the first one was) – but miserable at matching positive and negative tweets, which, presumably, are the more interesting ones to know about.
  14. Google trends computes how many searches on a particular topic (i.e. for selected search terms) have been done relative to the total number of google searches during that time period
  15. potentially promising for trend-spotting, modeling or qualitative analysis