Successfully reported this slideshow.
We use your LinkedIn profile and activity data to personalize ads and to show you more relevant ads. You can change your ad preferences anytime.

Open Metrics for Open Repositories at OR2012


Published on

Slides for a paper on "Open Metrics for Open Repositories" based on the paper available from and presented by Nick Sheppard at the Seventh International Conference on Open Repositories (OR2012) held in Edinburgh from 9-13th July 2012.

Published in: Education, Technology
  • Be the first to comment

Open Metrics for Open Repositories at OR2012

  1. 1. Open Metrics for Open Repositories 20x20 Pecha Kucha delivered at OR2012 on Tuesday 10th July based on the unpublished paper available from Brian Kelly1, Nick Sheppard2, Jenny Delasalle3, Mark Dewey1, Owen Stephens4, Gareth J Johnson5 and Stephanie Taylor1 1 UKOLN, University of Bath, Bath, UK {B.Kelly, M.Dewey, S.Taylor} 2 UKCoRR/Leeds Metropolitan University, Leeds, UK {} 3 University of Warwick, Warwick, UK {} 4 Consultant, UK {} 5 UKCoRR/University of Leicester, Leicester, UK {}
  2. 2. OA still in its infancy?• Need for metrics• Understand how IRs are being used• Policy decisions• Technical infrastructure• Business data• Not just research• Open Data / OER• Open landscape
  3. 3. • New, alternative metrics• Use in evaluation and assessment• Exploit opportunities• Usage / bibliographic data• Make openly available• Adoption of altmetrics• The BOAI at 10 – Alma Swan (video)
  4. 4. The Finch report• Executive Summary• Gold vs Green• Estimated additional £50-60M• Green has failed!• What does failure look like?• What would success look like?• Evidence
  5. 5. Practice what you preach• Open it up• Article level (ideal)• Software / fragmentation• Wrangle your software• Need for accurate aggregation• Technical challenges Julian Kleyn (2008)
  6. 6. The Institutional Picture• Total number of records• Full-text / metadata only• Raw figure / Proportion of total• Metadata records accessed• Full-text downloaded• How records are accessed – Search/browse – Search engine referral – Referral from aggregation – Social media referral
  7. 7. EPrints
  8. 8. • Article level metrics• Open interface•• Use as advocacy tool
  9. 9. Dspace (GA plug-in) • More reliable than existing DS plugins • lost stats/inflated page views from robots • Stats available at three levels: – item, – collection – repository
  10. 10. The Big Picture• OPuS - Bath (EPrints)• Leeds Metropolitan• Proportion as measure?• RepUK / aggregations• Software issues – EPrints – DSpace – Also ran
  11. 11. Number of Full Text• Few IRs can easily provide data• RepUK (national aggregation service)• CORE• Discrepancies• Proportion as measure?• Full text only exemplars (19) (Should be HTML?)• Developing Research Management Infrastructure Brunel, City University London, Cranfield, Loughborough,• CRIS QMUL, RHUL, School of Advanced Study, Aberdeen, Birmingham, Cambridge, East London, Edinburgh, Exeter, Hull, St Andrews, Stirling, Surrey, Warwick, Nottingham
  12. 12. RepUK
  13. 13. DepositsOai_dc node occurences • Technical challengesOai_dc node occurences over time • OpenDoarOai_dc percentabe node occurences over timeMetadata quality – validity of XML and DRIVER • OAI-PMHcompliancedc:subject classification system occurrence • Data visualisationdc:language occurrence across UK repositoriesMatches against IANA Mime Types • Trends revealed
  14. 14. (Edinburgh Research Archive)154 IRs aggregated (28/06/12)• 397790 PDF• 20578 MS Word• 44018 jpeg• 122764 html
  15. 15. The Researchers’ Requirements• May have effect on deposit choice• PLoS/Arxiv display article-level metrics• Numbers aggregated (locations / versions)• PLoS show visitor numbers from PMC• May not deposit in IR to maximise numbers at preferred location• Few publishers display article-level metrics• Opportunity for IRs to engage with authors• Measuring impact (REF, RCUK)• altmetrics
  16. 16. Third Party Services• Mechanisms to harvest metadata/full-text• OAI-PMH• The search party didn’t turn up?• Slow growth• Google (Scholar)• CiteSeerX, CORE cache copies• Affects repository metrics• PIRUS2 / article level metrics
  17. 17. COnnecting REpositories (CORE)• Repository Analytics• Increase the visibility of content• Provide applications to aid content discovery• Enrich metadata
  18. 18. Conclusions • Yes it’s complicated! • Stats don’t tell the whole picture • Greatest value for operational / strategic purposes • Senior management • IR managers should be proactive • Culture of openness