Advertisement
Advertisement

More Related Content

Similar to See Potential for crowdsourcing and mobile phones Nov 10 2014(20)

Advertisement

More from CCAFS | CGIAR Research Program on Climate Change, Agriculture and Food Security(20)

Recently uploaded(20)

Advertisement

See Potential for crowdsourcing and mobile phones Nov 10 2014

  1. Reducing the Costs of GHG Estimates in Agriculture to Inform Low Emissions Development,10-12 Nov 2014, Rome Italy The Potential for Crowdsourcing and Using Mobile Phone Technology Linda See Ecosystems Services and Management Program Geo-Wiki Team: Steffen Fritz, Ian McCallum, Christoph Perger, Martina Duerauer, Mathias Karner, Jon Nordling, Michael Obersteiner
  2. Overview • Terminology • Intro to Geo-Wiki – Campaigns – Hackathon – Gaming • Mobile devices and applications • Data collection for GHG accounting • Discussion
  3. Crowdsourcing • Outsourcing to the crowd (Howe, 2006) – E.g. Amazon’s Mechanical Turk • Using the crowd to collect data, solicit ideas, analyze data, do voluminous tasks that could otherwise not be done • Represents an untapped potential source of data for scientific research – Already being harnessed in ecology, conservation, species identification under the umbrella of citizen science
  4. Plethora of Terminology • Citizen science (+ extreme version) • PPSR • Volunteereed Geographic Information • GeoCollaboration / PPGIS • GeoWeb • Neogeography • Participatory sensing • Web mapping
  5. Context: Need for Improved Land Cover • Crucial baseline information for many applications/integrated assessment models • Overall and spatial disagreement when different products are compared • Need for more ground-based validation data • Confusing for users – Which one is correct? Which is the best product to use? • A number of studies have shown that the choice of land cover can have a significant affect on the final results
  6. Geo-Wiki: Visualization, Crowdsourcing and Validation Tool
  7. Large Disagreements in Cropland Can view these on: http://www.geo-wiki.org
  8. Showing Disagreement on Google Earth
  9. Example from a Competition (Humanimpact.geo-wiki.org)
  10. Data from Human Impact Competition
  11. Crowdsourcing Validation Data Number Competition Purpose of the Competition 1 Human Impact To validate a map of land availability for biofuel production 2 Hotspots of Map Disagreement To collect validation points in the areas were the GLC2000, MODIS and GlobCover disagree with one another 3 Wilderness To collect land cover and human impact in order to determine the amount of global wilderness. The locations used were the same as that of the Chinese 30 m land cover map 4 Global Validation Dataset To collect data at the same locations as the validation data assembled for the Chinese 30 m land cover map 5 & 6 Hackathon and IIASA Competition To collect data on the degree of cultivation and the degree of human settlement in Ethiopia in the context of land grabbing ~200,000 validation samples collected
  12. Outputs from Geo-Wiki: Cropland Map
  13. Outputs from Geo-Wiki: Map of Field Size
  14. Geo-Wiki Output: Global Map of Human Impact / Wilderness
  15. Outputs from Geo-Wiki: Downgrading of Land Availability for Biofuels Scenario Original figures (million ha) Adjusted for land cover (million ha) Adjusted for field size (millon ha) Adjusted for human impact (million ha) S1 320 98 42 34 S2 702 467 201 84 S3 1411 998 N/A 409 S4 1107 786 N/A 264 Fritz, S., See, L., van der Velde, M., Nalepa, R.A., Perger, C., Schill, C., McCallum, I., Schepaschenko, D., Kraxner, F., Cai, X., Zhang, X., Ortner, S., Hazarika, R., Cipriani, A., Di Bella, C., Rabia, A.H., Garcia, A., Vakolyuk, M., Singha, K., Beget, M.E., Erasmi, S., Albrecht, F., Shaw, B., Obersteiner, M. 2013. Downgrading recent estimates of land available for biofuel production. Environmental Science & Technology, 47(3), 1688-1694.
  16. Hackathon.geo-wiki.org • Organized by USAID • Challenge: – Collect information about cropland and settlement for Ethiopia – Overlay with location of land acquisitions – Look for evidence of effects on local populations • Extended to a competition for 3 weeks
  17. Land Grabbing Source: http://www.petergiovannini.com/Landgrabbing/
  18. More Outputs from a Hackathon
  19. Land Acquisition Area (from Land Matrix Database) + Clear Evidence of Settlements (from Geo-Wiki Hackathon) = Areas of Conflict
  20. Data Collected over Three Weeks
  21. Geo-Wiki Output: Interpolated Cropland Map
  22. Comparison through Differencing
  23. Accuracy Assessment Maps Accuracy measures (%) Overall accuracy User’s accuracy Producer’s accuracy All No Crop Crop No Crop Crop GLC-2000 77.3 90.5 48.1 79.5 69.6 MODIS 81.8 83.2 67.5 96.1 29.3 GlobCover 74.5 89.3 43.9 76.8 66.3 Crowdsourced cropland map 89.3 91.7 78.8 94.9 68.5 See, L., McCallum, I., Fritz, S., Perger, C., Kraxner, F., Obersteiner, M., Deka Baruah, U., Mili, N. and Ram Kalita, N. In press. Mapping Cropland in Ethiopia using Crowdsourcing. International Journal of Geosciences.
  24. Evolution of Crowdsourcing Cropland Capture collected validations African hybrid of cropland map number of Six campaigns Log Early competitions Launch Early games 2009 Time Today
  25. Multi-Platform Game
  26. Cropland Capture
  27. How Does it Work? • We started with a small pool of images already classified by experts • 90% of images the players get have already been classified • 10% of images not classified given to ‘good’ players è We assume the player to be correct on these images • The pool of classified images automatically increases
  28. Image 17365 Denmark Yes=69 No=1 Maybe=0
  29. Image 36318 Zimbabwe Yes=21 No=20 Maybe=2
  30. Incentives • Leaderboard • Weekly prizes: one random classification is picked; the person who did this classification receives a prize (last 5 weeks) • 3 final winners: at the end of the competition 3 winners were drawn to receiver bigger prizes, e.g. a tablet and smartphone
  31. ‘Softer’ Motivations • 66% of participants said they liked the idea of helping science • 24% were motivated by the prizes at the end • 24% like looking at the images/pictures • 19% were driven by the leaderboard • 29% said the game was fun • 25% said they stop playing in a given week when they realize they cannot get into the top 3 places
  32. Help from the Media
  33. Summary of the Competition • Ran for 25 weeks (15 Nov to 9 May 2014) • 3,014 players • 4,567,110 classifications • 187,673 unique images – 98,411 satellite images (250m to 1km.sq) – 89,232 photos
  34. Classifications by Device Device Number % iPhoneS 578,331 12.7 Other iPhones 616,537 13.5 iPads 698,762 15.3 Android Devices 1,636,627 35.8 Browsers 1,036,853 22.7 • Majority play on mobile devices • Apple products used more frequently than Android but not by that much
  35. Multiple Classifications
  36. Agreement of the Crowd
  37. Mobile Devices and Apps
  38. Geo-Wiki Pictures Mobile App
  39. Latest Version of the App
  40. FarmSupport Mobile App
  41. GEOSAF Early Warning App
  42. Grower’s Nation App http://www.growers-nation.org/
  43. ODK / GeoODK
  44. Sample Screens from GeoODK
  45. SATIDA GeoODK App
  46. SATIDA GeoODK App
  47. Household Survey App
  48. IFAD Household Survey App
  49. Household Survey App
  50. Ongoing Work • Cities Geo-Wiki – relevant to GHG emissions calculations • Further development of picture-based household survey app • New game building on Cropland Capture on forests and deforestation • Possible further development of FarmSupport app • Other related projects
  51. Data Collection for GHG Accounting • Ask farmers about management practices (fertilizer application, manure management) and cropping practices (crop types, crop residues, crop calendars) via mobile devices • Could be more icon-based than text-based • Could use voice recognition • Could use photos to provide evidence / QA • Could be SMS-based • Needs an incentive scheme (e.g. mobile credits, access to market and weather information) • Identified as bridging the gap (Paustian, 2013)
  52. Livestock Geo-Wiki
  53. Topics for Discussion • Who is the crowd? • Creating a community / motivation • Quality assurance • Protocols for data collection • Bias in the data • Barriers to uptake – literacy, technology
  54. Thanks for your Attention! Questions? Discussion
Advertisement