Successfully reported this slideshow.
We use your LinkedIn profile and activity data to personalize ads and to show you more relevant ads. You can change your ad preferences anytime.

Using Surveys to Improve Your Library - Part 1

756 views

Published on

Using Surveys to Improve Your Library - Part 1

Published in: Education
  • I can advise you this service - ⇒ www.HelpWriting.net ⇐ Bought essay here. No problem.
       Reply 
    Are you sure you want to  Yes  No
    Your message goes here
  • ⇒ www.WritePaper.info ⇐ is a good website if you’re looking to get your essay written for you. You can also request things like research papers or dissertations. It’s really convenient and helpful.
       Reply 
    Are you sure you want to  Yes  No
    Your message goes here
  • My brother found Custom Writing Service ⇒ www.WritePaper.info ⇐ and ordered a couple of works. Their customer service is outstanding, never left a query unanswered.
       Reply 
    Are you sure you want to  Yes  No
    Your message goes here
  • Most surveys don't pay cash, =>> https://dwz1.cc/EWG1lhe4
       Reply 
    Are you sure you want to  Yes  No
    Your message goes here
  • I can advise you this service - ⇒ HelpWriting.net ⇐ Bought essay here. No problem.
       Reply 
    Are you sure you want to  Yes  No
    Your message goes here

Using Surveys to Improve Your Library - Part 1

  1. 1. Using surveys to improve your library Emily Daly & Joyce Chapman Duke University Libraries ALA eLearning Workshop, February 2017
  2. 2. Introductions Emily Daly, Head of Assessment & User Experience Joyce Chapman, Assessment Analyst & Consultant
  3. 3. Agenda • The assessment lifecycle • When to survey • Sampling populations and methods • Survey planning • Survey validation and piloting
  4. 4. Time to hear from you! 1. What type of library do you work in? 2. Which statement best describes assessment in your library?
  5. 5. THE ASSESSMENT LIFECYCLE
  6. 6. What is assessment? Assessment is a continuous and cyclical process by which we evaluate and improve services, products, workflows, and learning.
  7. 7. The assessment lifecycle
  8. 8. The assessment lifecycle: Plan • Determine your objectives • Define the questions that need to be answered and map questions to data • Design a method to answer the questions (set up a study, collect new data, extract existing data from a system)
  9. 9. The assessment lifecycle
  10. 10. The assessment lifecycle: Implement • We frequently measure everything that’s easy to measure, without a good reason • For data collection to foster assessment, we must first determine what it is we really care about, then initiate data collection that will inform meaningful analysis and outcomes
  11. 11. The assessment lifecycle
  12. 12. The assessment lifecycle: Analyze and report • Analyze our data and report them to stakeholders • Unfortunately, the buck often stops here!
  13. 13. The assessment lifecycle
  14. 14. The assessment lifecycle: React and refine • The most frequent piece of the assessment cycle that is ignored is the last: making change based on the findings of data analysis • It is often inaction that causes the assessment loop to remain incomplete
  15. 15. The assessment lifecycle
  16. 16. Questions or comments?
  17. 17. WHEN TO SURVEY
  18. 18. What is a survey, and when do we use it?
  19. 19. Pros and cons of a survey Pros • Inexpensive, quick, ability to reach large numbers of people, can collect both qualitative and quantitative data, no observer subjectivity Cons • Subject to misinterpretation, inflexible design, inability to follow up or probe deeper
  20. 20. Surveys are best when… • You want data on attitudes, beliefs, experiences, needs, demographics, perceived behavior, etc. • You can’t acquire from a machine source • You want info from a large number of people and do not need to follow up on questions or probe deeply (at least not yet!)
  21. 21. …but you may need more • It’s often necessary to follow up on what you learn in a survey. • We use a technique we call triangulation.
  22. 22. Triangulating data • Methods to consider when triangulating data: – thoughtfully planned focus groups – semi-structured interviews – observational studies – targeted, more focused surveys – usage statistics or other numerical data (e.g., gate counts, circulation stats, web metrics)
  23. 23. Time to hear from you! Which of the following have you conducted or been involved in at your library?
  24. 24. Questions or comments?
  25. 25. SAMPLING POPULATIONS AND METHODS
  26. 26. Clearly define your target population
  27. 27. Census survey • While most surveys rely on sampling, you can also gather information from every single person in a target population. • This is called a census. • Example: your academic library provides carrels to select graduate students and in return, each student is required to complete a survey. You have information from 100% of the population.
  28. 28. Sampling • Sampling uses a representative group of a given population to determine characteristics of the entire population. • If you can’t talk to everyone, you get a sample.
  29. 29. Random sampling Ideally we use random sampling: • We invite a smaller group of people (the sample) from a larger group (the population) to answer our survey. • Each person is chosen randomly and each member of the population has an equal chance of being included in the sample. • In this way, we hope various groups within the population end up equally represented in the results.
  30. 30. Convenience sampling • A convenience sample is composed of people who are easy to reach • Unlikely that a convenience sample will accurately represent your target population • Example: placing a satisfaction survey at your library’s reference desk
  31. 31. Who are we missing? • Sometimes you want feedback from people who are not easily accessible – Patrons who only use online resources – Community members who do not use the library – People without email • Plan accordingly!
  32. 32. Types of survey distribution • Mail-out surveys • Web-based surveys • In-library paper-based surveys • Telephone surveys • In-person interviews • Intercept surveys
  33. 33. Web-based surveys Pros • Convenience, rapid data collection, cost- effective, ample time, confidentiality and security Cons • Limited respondent base, self-selection*, lack of interviewer involvement
  34. 34. Questions or comments?
  35. 35. SURVEY PLANNING
  36. 36. Before drafting your survey • What problems are you trying to solve? • What questions are you trying to answer? • What data will help you answer the questions, and does it already exist?
  37. 37. Getting permission and buy-in • Which staff, administrators, users, or other stakeholders need to be involved or kept informed? • Do you need IRB approval (colleges and universities)? • Who will ultimately receive your survey results? It’s wise to include them from the start!
  38. 38. Academic libraries: Working with IR • Do you have an Institutional Research office? • Develop a good working relationship with IR • Collaborate on surveys (when appropriate) months ahead of time – You may need their approval or assistance – Often they can provide a random sample – Can they provide demographic data with a sample, or add demographics to returned data?
  39. 39. Academic libraries: IRB • Institutional Review Board (IRB) • Reviews and approves research involving human subjects to ensure that it is conducted in accordance with all federal, institutional, and ethical guidelines • IRB is concerned with protecting the welfare, rights, and privacy of human subjects • Established in 1974 after major human rights abuses in research of the 20th century
  40. 40. IRB rules of thumb 1. Does it involve human subjects? –Yes 2. Is it “research”? – Quite possibly not! – Systematic investigation designed to develop or contribute to generalizable knowledge through public dissemination such as published articles, presentations, and poster displays. • Anonymous surveys are often exempt (you must still follow all ethical guidelines but may not be subject to IRB oversight)
  41. 41. Do I need to go through IRB? • Ask others at your library if there is a library- wide policy – sometimes libraries have blanket policies with their local IRBs for most surveys • Contact the IRB – if you have medical facilities there may be two, contact the non-medical – Describe your project – They will tell you whether you do not need to submit, or if you can submit an exempt proposal
  42. 42. Confidential v. anonymous
  43. 43. Timing is key! • Consider your primary audience when thinking about timing for survey release • Consider the circumstances of different target populations
  44. 44. Distribution details • What type of distribution is required? – Rolling: Is the survey ongoing, without a closing date? – One time/periodic: Is the survey distributed one time, or once a year, etc.? – Program-dependent: Is the survey distribution linked to a particular program? • Beware survey fatigue!
  45. 45. Incentives • Will you provide an incentive? – Raffle? Each participant? – $$? Service? Goods? • Anonymity may pose a problem for incentives
  46. 46. Recruitment strategy • Target audience: Program participants, users of a particular service you want to know more about • Random sample if you are interested in a cross section of users • Recruitment methods: Direct email, links on homepage or pertinent webpages, email blasts/listservs, bathroom fliers
  47. 47. Web-based distribution • For open web links – Will you collect an identifier? – Will you prevent “ballot box stuffing”? • For direct invites – Leave open 1-3 weeks – Consider invitation email carefully – Send at least one follow-up reminder (ideally only to those who have not yet completed the survey)
  48. 48. Questions or comments?
  49. 49. SURVEY VALIDATION AND PILOTING
  50. 50. What is survey validation? • The process of assessing the survey questions for their dependability • Have two parties review the survey if possible: 1. People familiar with the topic 2. Expert in survey question design
  51. 51. Validating the survey • Start with a simple text document • Have validators go through the survey and make notes – Do all questions and answer choices make sense, are they unbiased, etc.? – Will the resulting data help you answer your questions? Is all topical content accurate? • Make changes based on validation!
  52. 52. What is survey piloting? • Select a small subset of your target population to take your survey • Even on pilot tester is better than none! • Try to get a range of different people who represent your target group
  53. 53. Pilot testing your survey • Enter survey into online tool or final paper doc, and then test with several respondents • Revise your survey and re-test • Time respondents
  54. 54. Questions or comments?
  55. 55. Plan for next week’s workshop • Questions or comments from last week? • Structure of the survey • Writing unbiased, actionable questions • Survey tools • Acting on survey data • Tips and lessons learned
  56. 56. Thank you! ALA eLearning will send participants a link to the recorded workshop and slide deck. With questions about our content, contact – Emily Daly: emily.daly@duke.edu – Joyce Chapman: joyce.chapman@duke.edu With questions for ALA eLearning, contact – editionscoursehelp@ala.org

×