1) The document discusses how baseball management's use of analytics and data-driven decision making can inform how research libraries manage collections in a digital environment.
2) It advocates shifting from traditional supply-side collection management to a demand-driven model based on analyzing actual user data and behaviors.
3) The document provides examples from North Carolina State University of how data analysis, statistical modeling, and evidence-based practices can help optimize collections expenditures and better meet user needs.
This document discusses how libraries can transition from traditional supply-driven collection models to more demand-driven, data-driven models that are more sustainable. It argues that usage data and analytics should be used more in collection decisions to lower costs and better meet user needs. Specific strategies mentioned include analyzing print and e-book usage patterns, using data to inform space planning, and collaborating through resource sharing networks. Challenges discussed include resistance to change and accounting for niche areas. The document advocates growing analytical skills, experimenting, and using data to make collections more vital to researchers.
Virginia tech collections_presentationGreg Raschke
This document discusses the need for libraries to shift from supply-driven to demand-driven collection strategies and practices due to unsustainable budget growth. It argues that libraries must lower costs by analyzing usage data, embracing patron-driven acquisition models, and investing in digital content delivery and curation to better serve user needs. The document also highlights examples from North Carolina State University of analyzing data to optimize collections spending and make evidence-based decisions.
The document discusses how library collections must change from a supply-driven to a demand-driven model due to unsustainable budget increases and the ability to deliver content digitally. It advocates lowering costs by analyzing usage data, embracing patron-driven acquisition, and investing in digital curation and technologies that provide content on demand. The author provides examples from North Carolina State University of analyzing print usage and expenditures to guide evidence-based collection decisions and ensure resources are used and vital.
The document discusses perspectives on consortial e-book acquisitions from librarians, publishers, and vendors. It summarizes findings from analyzing e-book purchasing data from three universities. Key points of agreement and disagreement between stakeholders are outlined regarding flexible acquisition options, usage data, and resource sharing models. Current realities of consortia patron-driven acquisition models are also presented.
Beyond Print Summit: TRLN History, Context, and MotivationsGreg Raschke
TRLN has a long history of collaboration dating back to the 1930s to share resources and collections across its member universities. As content has increasingly moved to digital formats, TRLN has worked to extend this collaboration to licensing e-journals, e-books, and databases for shared access. However, e-books present new challenges as they currently do not allow for the same sharing capabilities as print books due to copyright restrictions and digital rights management. TRLN aims to shape the future of e-books by spearheading new collaborative models that enable scholarship while leveraging the benefits of digital content and repositioning resource sharing in a hybrid print and electronic environment.
ALA Collections Review Presentation(070209)Greg Raschke
This document discusses challenges libraries face in collection development and management during difficult economic times. It outlines how collection reviews have evolved from 1995 to 2009 to increasingly rely on data-driven, evidence-based decision making including usage data, bibliometrics, and user feedback. The 2009 review emphasizes quantifying multiple data points to categorize content and identifying underperforming resources to reallocate funds. Ensuring the right content delivers high use and value relative to cost is key along with exploring new collaborative models.
Libraries Leading the Way on the 'Textbook Problem'Greg Raschke
Libraries are taking a leading role in addressing the high cost of textbooks by developing expertise in alternative textbook models. They are advocating for affordable options, educating faculty on alternatives, and directly supporting the creation and hosting of open educational resources. Some libraries purchase one copy of each required textbook and partner with bookstores. However, they are now putting more effort into transforming the textbook market by licensing and hosting open textbooks online, providing print-on-demand options, and working with companies developing new affordable textbook models.
ARL Collections Presentation: Moneyball, the Extra 2%, and What Baseball Mana...Greg Raschke
The document discusses how principles from baseball management and the book Moneyball can be applied to foster innovation in managing library collections. It suggests looking for inefficiencies in the market, questioning long-established assumptions, and applying significant innovations through statistical analysis and new approaches while emphasizing interpersonal skills. It recommends adapting tools like SAS software, experimenting, and getting faculty buy-in to combine analytical and people skills for developing collections.
This document discusses how libraries can transition from traditional supply-driven collection models to more demand-driven, data-driven models that are more sustainable. It argues that usage data and analytics should be used more in collection decisions to lower costs and better meet user needs. Specific strategies mentioned include analyzing print and e-book usage patterns, using data to inform space planning, and collaborating through resource sharing networks. Challenges discussed include resistance to change and accounting for niche areas. The document advocates growing analytical skills, experimenting, and using data to make collections more vital to researchers.
Virginia tech collections_presentationGreg Raschke
This document discusses the need for libraries to shift from supply-driven to demand-driven collection strategies and practices due to unsustainable budget growth. It argues that libraries must lower costs by analyzing usage data, embracing patron-driven acquisition models, and investing in digital content delivery and curation to better serve user needs. The document also highlights examples from North Carolina State University of analyzing data to optimize collections spending and make evidence-based decisions.
The document discusses how library collections must change from a supply-driven to a demand-driven model due to unsustainable budget increases and the ability to deliver content digitally. It advocates lowering costs by analyzing usage data, embracing patron-driven acquisition, and investing in digital curation and technologies that provide content on demand. The author provides examples from North Carolina State University of analyzing print usage and expenditures to guide evidence-based collection decisions and ensure resources are used and vital.
The document discusses perspectives on consortial e-book acquisitions from librarians, publishers, and vendors. It summarizes findings from analyzing e-book purchasing data from three universities. Key points of agreement and disagreement between stakeholders are outlined regarding flexible acquisition options, usage data, and resource sharing models. Current realities of consortia patron-driven acquisition models are also presented.
Beyond Print Summit: TRLN History, Context, and MotivationsGreg Raschke
TRLN has a long history of collaboration dating back to the 1930s to share resources and collections across its member universities. As content has increasingly moved to digital formats, TRLN has worked to extend this collaboration to licensing e-journals, e-books, and databases for shared access. However, e-books present new challenges as they currently do not allow for the same sharing capabilities as print books due to copyright restrictions and digital rights management. TRLN aims to shape the future of e-books by spearheading new collaborative models that enable scholarship while leveraging the benefits of digital content and repositioning resource sharing in a hybrid print and electronic environment.
ALA Collections Review Presentation(070209)Greg Raschke
This document discusses challenges libraries face in collection development and management during difficult economic times. It outlines how collection reviews have evolved from 1995 to 2009 to increasingly rely on data-driven, evidence-based decision making including usage data, bibliometrics, and user feedback. The 2009 review emphasizes quantifying multiple data points to categorize content and identifying underperforming resources to reallocate funds. Ensuring the right content delivers high use and value relative to cost is key along with exploring new collaborative models.
Libraries Leading the Way on the 'Textbook Problem'Greg Raschke
Libraries are taking a leading role in addressing the high cost of textbooks by developing expertise in alternative textbook models. They are advocating for affordable options, educating faculty on alternatives, and directly supporting the creation and hosting of open educational resources. Some libraries purchase one copy of each required textbook and partner with bookstores. However, they are now putting more effort into transforming the textbook market by licensing and hosting open textbooks online, providing print-on-demand options, and working with companies developing new affordable textbook models.
ARL Collections Presentation: Moneyball, the Extra 2%, and What Baseball Mana...Greg Raschke
The document discusses how principles from baseball management and the book Moneyball can be applied to foster innovation in managing library collections. It suggests looking for inefficiencies in the market, questioning long-established assumptions, and applying significant innovations through statistical analysis and new approaches while emphasizing interpersonal skills. It recommends adapting tools like SAS software, experimenting, and getting faculty buy-in to combine analytical and people skills for developing collections.
This document discusses the importance of libraries using data and metrics to inform decision making and communicate their value. It notes that while librarians are good at collecting statistics, they often lack the ability or willingness to analyze, interpret, and apply the data. Without using data to take action or inform strategies, it has little value beyond justifying budgets. The document provides examples of how libraries can select peer institutions for benchmarking, analyze trends over time, and use multiple data sets to understand user needs and behaviors. It emphasizes telling "stories" combined with data ("Stories + Stats") to communicate effectively with stakeholders.
Lecture presented by Vivian Praxedes D. Sy at PAARL's Summer Conference on the theme "Library Analytics: Data-driven Library Management", held at Pearl Hotel, Manila on 20-22 April 2016
This presentation was provided by Galadriel Chilton of The Ivy Plus Library Confederation, during the NISO training series "Assessment Practices and Metrics for the 21st Century (Session Three)," held on November 8, 2019.
This presentation was provided by Stuart Maxwell of Scholarly iQ, during the NFAIS Forethought event "Artificial Intelligence #2 – Processes for Media Analysis and Extraction" The webinar was held on May 20, 2020.
This presentation was provided by Elliot Felix of Brightspot, during the NISO training series "Assessment Practices and Metrics for the 21st Century (Session One)," held on October 25, 2019.
Getting the Most Out of Your E-Resources: Measuring Successkramsey
The document discusses measuring the usage and success of electronic resources. It provides an overview of NISO and standards they develop, including COUNTER and SUSHI. SUSHI allows for automated gathering of COUNTER usage reports to make collecting data easier for libraries. The document also discusses applying usage data, privacy concerns, and areas for future development.
This presentation was provided by Clara Chu and Merinda Kaye Hensley of The University of Illinois, during Session Eight of the NISO event "Assessment Practices and Metrics for the 21st Century," held on December 6, 2019.
This document summarizes the Library Impact Data Project, which aimed to show correlations between library usage data (books borrowed, e-resources accessed) and student attainment across multiple universities. Phase 1 found statistical significance between library usage and grades. Phase 2 added more student data points and found further correlations with demographics. The project aims to create a shared analytics service to allow libraries to analyze usage and benchmark against peers. Key areas for the next phase include developing an intuitive dashboard, addressing ethical issues around profiling individuals, and integrating additional data sources.
This presentation was provided by Joe Zucca of the University of Pennsylvania, during Session Five of the NISO event "Assessment Practices and Metrics for the 21st Century," held on November 22, 2019.
This presentation was provided by Heidi Nance of The Ivy Plus Libraries Confederation, during Session Six of the NISO event "Assessment Practices and Metrics for the 21st Century," held on December 6, 2019.
This presentation was provided by Steven Braun of Northeastern University Libraries during the NISO webinar, Using Analytics to Extract Value from the Library's Data, held on September 12, 2018.
The document discusses research information management systems (RIMs) and the role of libraries in supporting them. It describes RIMs as systems that collate fragmented institutional research data to reduce administrative burdens. Key functions of RIMs include automated data capture, integration with internal and external data sources, and providing analytics. The document argues that RIMs benefit institutions by centralizing research information for assessment, funding applications, and increasing visibility. Libraries are well-positioned to advise on RIMs and play a lead role in planning institutional research data collection and management.
This presentation was provided by Lisa Hinchliffe of The University of Illinois, during Session Seven of the NISO event "Assessment Practices and Metrics for the 21st Century," held on December 13, 2019.
This document discusses evaluating and selecting online resources. It begins by introducing the presenters and their interests in usage measurement and the concept of electronic resource usage in libraries. It then discusses how collection practices have shifted from a supply-side to a demand-driven model where use and demand analysis drive decisions. The document advocates for a data-driven approach using multiple variables like usage statistics, user feedback, and quality metrics to make large decisions. It outlines an agenda to discuss understanding e-resource use and value as well as negotiations. Finally, it questions how to best measure use and the value of understanding use as a complex process rather than just a number.
This document discusses methods for measuring the impact of citizen science projects online. It describes the development of a framework called MICS (Measuring Impact of Citizen Science) for assessing citizen science impact. MICS includes indicators for different domains like society, science, economy, environment and governance. The framework provides characteristics for each indicator such as its name, description, data type, and how data should be collected and analyzed. Case studies are being used to help implement and refine the MICS framework.
When Rubber Meets the Road: Rethinking Your Library Collections by Roger Scho...Charleston Conference
This document summarizes a presentation given by Roger C. Schonfeld at the Charleston Conference on rethinking library collections. Schonfeld discusses how user needs are shifting away from print journals towards electronic formats. While user needs are changing, print collections still need to be preserved. Library collaborations can help reduce costs by ensuring at least one print copy is preserved while giving libraries more flexibility in managing local print collections. Schonfeld proposes a risk-informed, research-based approach to categorize materials based on their preservation needs and determine optimal levels of print preservation.
Data-Informed Decision Making for Libraries - Athenaeum21Megan Hurst
Athenaeum21 presents three case studies of assessment and evaluation programs in libraries--one past, one current, and one future. The cases use three different modes of data gathering and analysis to show the power of understanding user needs and how well your organization is meeting them.
*Updated and reorganised following feedback in the breakouts*
While many librarians have developed mechanisms and
structures for managing local scholarship separate from
their standard resource management practices, the
intersection of the two content streams is occurring at
many institutions. During the past decade the presenters
have dedicated themselves to capturing best practices
of electronic resource management and mapping out
paths for creating open access workflows. Join them for a
lively discussion and interactive session where they outline
ways to bring these two initiatives together and identify the
teams needed.
Graham Stone
Jisc Collections
Peter McCracken
Cornell University
Jill Emery
Portland State University Library
Fostering Entrepreneurship Through Cooperative Data and Services
Panel Discussion
Moderator: Karen Coombs, Product Manager, OCLC Developer Network, OCLC; Cyril Oberlander, Interim Director, Milne Library, SUNY College at Geneseo; Annette Bailey, Digital Assets Librarian, Virginia Tech and LibX
How do you take something you don’t own and improve something you do? Find out from four library technologists. This panel discussion presented how they’ve drawn cooperative library data from the cloud to inexpensively innovate and better meet their users’ needs through OCLC Web Services and created LibX. OCLC Web services makes data on libraries and collections available for additional application development. It’s all made possible through library cooperation and the OCLC Developer Network community.
Clay Shirky, Fantasy Football, and Using Data to Glean the Future of Library ...Greg Raschke
This document discusses moving from traditional supply-driven library collections to demand-driven collections using data analysis. It notes that collection budgets are unsustainable under traditional models and that data can help lower costs by making collections more precise and responsive to user needs. The document advocates analyzing usage data to modify collecting practices and asserts that demand-driven, user-focused models will become a larger share of budgets. It provides examples of how North Carolina State University uses data like usage statistics, citations, and user feedback to evaluate resources and make evidence-based decisions about collections.
The document discusses using e-metrics to assess electronic collections. It defines e-metrics as standardized measurements that produce quantitative data extracted from using electronic resources. This can be used as a tool to assess effectiveness, efficiency, performance and quality of electronic resources. The document outlines various e-metrics analyses that can be done, including trend analysis, efficiency studies, and cost-benefit analysis using return on investment and cost per article reading. Examples of e-metrics reports are provided to illustrate utilization rate, satisfaction rate, efficiency rate, and usability rate analyses.
This document discusses the importance of libraries using data and metrics to inform decision making and communicate their value. It notes that while librarians are good at collecting statistics, they often lack the ability or willingness to analyze, interpret, and apply the data. Without using data to take action or inform strategies, it has little value beyond justifying budgets. The document provides examples of how libraries can select peer institutions for benchmarking, analyze trends over time, and use multiple data sets to understand user needs and behaviors. It emphasizes telling "stories" combined with data ("Stories + Stats") to communicate effectively with stakeholders.
Lecture presented by Vivian Praxedes D. Sy at PAARL's Summer Conference on the theme "Library Analytics: Data-driven Library Management", held at Pearl Hotel, Manila on 20-22 April 2016
This presentation was provided by Galadriel Chilton of The Ivy Plus Library Confederation, during the NISO training series "Assessment Practices and Metrics for the 21st Century (Session Three)," held on November 8, 2019.
This presentation was provided by Stuart Maxwell of Scholarly iQ, during the NFAIS Forethought event "Artificial Intelligence #2 – Processes for Media Analysis and Extraction" The webinar was held on May 20, 2020.
This presentation was provided by Elliot Felix of Brightspot, during the NISO training series "Assessment Practices and Metrics for the 21st Century (Session One)," held on October 25, 2019.
Getting the Most Out of Your E-Resources: Measuring Successkramsey
The document discusses measuring the usage and success of electronic resources. It provides an overview of NISO and standards they develop, including COUNTER and SUSHI. SUSHI allows for automated gathering of COUNTER usage reports to make collecting data easier for libraries. The document also discusses applying usage data, privacy concerns, and areas for future development.
This presentation was provided by Clara Chu and Merinda Kaye Hensley of The University of Illinois, during Session Eight of the NISO event "Assessment Practices and Metrics for the 21st Century," held on December 6, 2019.
This document summarizes the Library Impact Data Project, which aimed to show correlations between library usage data (books borrowed, e-resources accessed) and student attainment across multiple universities. Phase 1 found statistical significance between library usage and grades. Phase 2 added more student data points and found further correlations with demographics. The project aims to create a shared analytics service to allow libraries to analyze usage and benchmark against peers. Key areas for the next phase include developing an intuitive dashboard, addressing ethical issues around profiling individuals, and integrating additional data sources.
This presentation was provided by Joe Zucca of the University of Pennsylvania, during Session Five of the NISO event "Assessment Practices and Metrics for the 21st Century," held on November 22, 2019.
This presentation was provided by Heidi Nance of The Ivy Plus Libraries Confederation, during Session Six of the NISO event "Assessment Practices and Metrics for the 21st Century," held on December 6, 2019.
This presentation was provided by Steven Braun of Northeastern University Libraries during the NISO webinar, Using Analytics to Extract Value from the Library's Data, held on September 12, 2018.
The document discusses research information management systems (RIMs) and the role of libraries in supporting them. It describes RIMs as systems that collate fragmented institutional research data to reduce administrative burdens. Key functions of RIMs include automated data capture, integration with internal and external data sources, and providing analytics. The document argues that RIMs benefit institutions by centralizing research information for assessment, funding applications, and increasing visibility. Libraries are well-positioned to advise on RIMs and play a lead role in planning institutional research data collection and management.
This presentation was provided by Lisa Hinchliffe of The University of Illinois, during Session Seven of the NISO event "Assessment Practices and Metrics for the 21st Century," held on December 13, 2019.
This document discusses evaluating and selecting online resources. It begins by introducing the presenters and their interests in usage measurement and the concept of electronic resource usage in libraries. It then discusses how collection practices have shifted from a supply-side to a demand-driven model where use and demand analysis drive decisions. The document advocates for a data-driven approach using multiple variables like usage statistics, user feedback, and quality metrics to make large decisions. It outlines an agenda to discuss understanding e-resource use and value as well as negotiations. Finally, it questions how to best measure use and the value of understanding use as a complex process rather than just a number.
This document discusses methods for measuring the impact of citizen science projects online. It describes the development of a framework called MICS (Measuring Impact of Citizen Science) for assessing citizen science impact. MICS includes indicators for different domains like society, science, economy, environment and governance. The framework provides characteristics for each indicator such as its name, description, data type, and how data should be collected and analyzed. Case studies are being used to help implement and refine the MICS framework.
When Rubber Meets the Road: Rethinking Your Library Collections by Roger Scho...Charleston Conference
This document summarizes a presentation given by Roger C. Schonfeld at the Charleston Conference on rethinking library collections. Schonfeld discusses how user needs are shifting away from print journals towards electronic formats. While user needs are changing, print collections still need to be preserved. Library collaborations can help reduce costs by ensuring at least one print copy is preserved while giving libraries more flexibility in managing local print collections. Schonfeld proposes a risk-informed, research-based approach to categorize materials based on their preservation needs and determine optimal levels of print preservation.
Data-Informed Decision Making for Libraries - Athenaeum21Megan Hurst
Athenaeum21 presents three case studies of assessment and evaluation programs in libraries--one past, one current, and one future. The cases use three different modes of data gathering and analysis to show the power of understanding user needs and how well your organization is meeting them.
*Updated and reorganised following feedback in the breakouts*
While many librarians have developed mechanisms and
structures for managing local scholarship separate from
their standard resource management practices, the
intersection of the two content streams is occurring at
many institutions. During the past decade the presenters
have dedicated themselves to capturing best practices
of electronic resource management and mapping out
paths for creating open access workflows. Join them for a
lively discussion and interactive session where they outline
ways to bring these two initiatives together and identify the
teams needed.
Graham Stone
Jisc Collections
Peter McCracken
Cornell University
Jill Emery
Portland State University Library
Fostering Entrepreneurship Through Cooperative Data and Services
Panel Discussion
Moderator: Karen Coombs, Product Manager, OCLC Developer Network, OCLC; Cyril Oberlander, Interim Director, Milne Library, SUNY College at Geneseo; Annette Bailey, Digital Assets Librarian, Virginia Tech and LibX
How do you take something you don’t own and improve something you do? Find out from four library technologists. This panel discussion presented how they’ve drawn cooperative library data from the cloud to inexpensively innovate and better meet their users’ needs through OCLC Web Services and created LibX. OCLC Web services makes data on libraries and collections available for additional application development. It’s all made possible through library cooperation and the OCLC Developer Network community.
Clay Shirky, Fantasy Football, and Using Data to Glean the Future of Library ...Greg Raschke
This document discusses moving from traditional supply-driven library collections to demand-driven collections using data analysis. It notes that collection budgets are unsustainable under traditional models and that data can help lower costs by making collections more precise and responsive to user needs. The document advocates analyzing usage data to modify collecting practices and asserts that demand-driven, user-focused models will become a larger share of budgets. It provides examples of how North Carolina State University uses data like usage statistics, citations, and user feedback to evaluate resources and make evidence-based decisions about collections.
The document discusses using e-metrics to assess electronic collections. It defines e-metrics as standardized measurements that produce quantitative data extracted from using electronic resources. This can be used as a tool to assess effectiveness, efficiency, performance and quality of electronic resources. The document outlines various e-metrics analyses that can be done, including trend analysis, efficiency studies, and cost-benefit analysis using return on investment and cost per article reading. Examples of e-metrics reports are provided to illustrate utilization rate, satisfaction rate, efficiency rate, and usability rate analyses.
Brief presentation on data driven collection development or evidence based collection development. Generally, some of the things to watch out for and advice on how to view your data.
Effectively Applying Usage Statistics in E-Resource Collection Developmentralowe
This document discusses how libraries can effectively apply usage statistics in electronic resource collection development and decision making. It emphasizes two key elements: efficient data collection and clear analysis. Usage statistics should be used to track usage, expenditures, turnaways, and other metrics to assess the collection and inform decisions. Engaging librarians, faculty, and administrators in the process allows libraries to advocate for resources and demonstrate value. The document provides examples from Frostburg State University of surveys, underutilized journal reviews, and department collaborations that applied usage data to collection decisions.
My presentation given at the Association of Subscription Agents annual conference, Feb 2013.
It was titled Understanding how researchers and practitioners use STM information, but the specific theme was understanding how to design information products and services for researchs and practitioners against a background of information abundance (aka information overload).
A Comprehensive Review of Relevant Techniques used in Course Recommendation S...IRJET Journal
This document summarizes a research paper that reviewed techniques for course recommendation systems. It discussed four main recommendation approaches: content-based, collaborative filtering, knowledge-based, and hybrid systems. For each approach, it provided examples of previous research studies that utilized each approach. It also discussed challenges like cold starts, data sparsity, and privacy issues. Machine learning algorithms commonly used included clustering, classification, and association rule mining. The paper analyzed selected publications to evaluate different recommendation systems for online education. Overall, the document provided a comprehensive overview of course recommendation techniques and issues.
The document discusses using data and learning analytics to inform library investments and improve the student learning experience. It outlines an workshop on collecting and analyzing various data points like library usage, citations in reference managers, and quotes from readings. The goal is to move beyond typical metrics like satisfaction surveys and attendance to actual learning data that can better predict student outcomes and transform support. Privacy, transparency and monitoring versus engagement are important ethical considerations to discuss.
The document summarizes NIH's approach to data science and the ADDS mission. It discusses establishing a data ecosystem through community, policy, and infrastructure. The goals are to foster sustainability, efficiency, collaboration, reproducibility, and accessibility. NIH plans to seed the ecosystem through existing resources and funding. Example initiatives include establishing a data commons, standards, and training programs to develop a diverse data science workforce. The overall aim is to support a "digital enterprise" that enhances biomedical research and health outcomes.
This document provides an introduction to Snowball Metrics, which aims to establish standard metrics that research universities can use to understand their strengths and weaknesses. The Snowball Metrics approach is a bottom-up collaboration between research universities and Elsevier. It takes a pragmatic approach, agreeing on core metrics first and reusing existing standards where possible. The goal is for Snowball Metrics to become a globally recognized standard to inform evidence-based institutional decision-making, rather than ranking. Universities own the process of defining the metrics, while Elsevier provides project management support and technical expertise.
The document discusses gaining business intelligence from user activity data in libraries and higher education institutions. It outlines challenges in collecting and analyzing comprehensive user data from different systems. The Open University perspective is that most students do not visit physical libraries and sign up for individual courses rather than degrees. However, the university has significant online user traffic that could provide insights if integrated across various learning and library systems. Overcoming cultural, technical, and data challenges will be key to developing a comprehensive view of user activity data.
Data sets are often perceived as only being relevant for researchers.pdflakshmijewellery
Data sets are often perceived as only being relevant for researchers. As this course strives to help
students understand how data sets are used within business to make decisions, how do you
foresee the application of this course in your current line of work? Are there examples of times
in which knowing how to review and analyze data can be helpful in your current role?
Subject name Data Management, Analysis and Reporting
Solution
Data Management, Analysis and Reporting
Purpose of Data Management:
Appropriate information dealing with and administration is vital to the achievement and
reproducibility of astatistical investigation. Choice of the proper devices and proficient
utilization of these instruments can spare the scientist various hours, and permit different
specialists to use the results of their work. Also, as the span of databases in transportation keep
on develop, it is turning out to be progressively critical to put assets into the administration
ofthese information.
There are various auxiliary strides that should be performed both prior and then afterward
measurable examination of information. For instance, a database made out of various
information streams should be coordinated and incorporated into a solitary database for
investigation. Furthermore, now and again information must be changed into the favored
electronic configuration for an assortment of measurable bundles. At times, information acquired
from \"the field\" must be cleaned and repaired for information and estimation mistakes, and
reformatted. The accompanying areas talk about contemplations for building up a general
information collection,handling, and administration plan, and devices vital for fruitful usage of
that arrangement.The information gathering, taking care of, and administration arrange assumes
a critical part inside an examination extend. The arrangement gives a guide recording the stream
of information through the consecutive periods of accumulation, stockpiling, cleaning,
diminishment, investigation, lastly to documenting. Encourage, the administration arrange
archives the connections between the greater part of the product devices and projects important
to manage the information through this exploration life cycle.
The information taking care of and administration arrange should be produced before an
exploration extend starts. The arrangement, in any case, can develop as the analyst adapts more
about the information, and as new roads of information investigation are uncovered.
Data Input, Storage, Retrieval, and Preparation:
Are the information \"clean?\" The information input prepare generally presents grammatical
errors, miscodes, furthermore, mistakes into the information. (These blunders are unmistakably
unique in relation to arbitrary or estimation mistakes presented in the estimation procedure).
Distinctive capacity procedures bolster varying levels of information altering.
Are the information static, or will updates be accessible through the li.
The real challenge in the modern world is not producing information or storing information,
but apt and proper use of information by people. Since volume of information is growing in leaps
and bounds, the information needs of users are becoming more and more diverse and complex. In
this changing context information providers are facing a lot of challenges to capture, process, store
and disseminate the available information for actual users. The user studies provide a clear
understanding of the actual information needs of the user in order to readjust the existing
information systems or chose new ones. Various models of information needs and informationseeking behaviour have been discussed. Each modelrepresents a different but an overlapping or
similar approach to information seeking behavior of users. In order to satisfy the information need,the user actively undergoes the information seeking processes. Some factors like physiological,emotional, learning and demographic, etc. also deeply influence information seeking behaviour i.e.
some people have to face some obstacles. These barriers may be economic, social, environmental,
time related or geographical.Effectiveness of a professional depends upon dissemination and use of right information at
right time. Information and communication technologies have changed the way in which thelibraries provide their services. Users study provide deeperunderstanding of access to their
collections and services .The need and behavior of their users and satisfaction ratio of users are
new assessment techniques of libraries. Therefore an effort has been made to how determineinformation need and information seeking behavior of users.
Presenters: John Stephens, Lydia Hofstetter.
Presented at the Georgia Libraries Conference in Columbus, GA on 10/05/2017.
This session will give an overview of COUNTER statistics, including their development, use, and relation to other library metrics. The presentation will also discuss practical issues related to gathering, reporting, and common problems.
Patron-Driven Acquisition: Turning Theory into Practice (Part 2)ALATechSource
This document discusses patron-driven acquisition (PDA) in libraries, providing an overview of the topic and addressing related questions and considerations. It begins by defining PDA and explaining why libraries implement PDA programs. It then covers different PDA models for print and electronic resources, as well as budgeting, best practices for starting a PDA program, analyzing usage data, and controversies regarding PDA's effects on collections. The document aims to help libraries successfully plan and implement a PDA program while navigating related challenges.
The document discusses research information management systems (RIMs) and the role of libraries in supporting them. It describes RIMs as systems that collate fragmented institutional research data to reduce administrative burdens. Key functions of RIMs include automated data capture, integration with internal and external data sources, and providing analytics. The document argues that RIMs benefit institutions by centralizing research information for assessment, funding applications, and increasing visibility. Libraries are well-positioned to advise on RIMs and play a lead role in planning due to expertise in data management, bibliographic standards, and understanding researcher needs.
Big Data & DS Analytics for PAARL aims to help library participants relate Big Data and Data Science applications to library services. The speaker discusses Big Data concepts like the 3 V's of volume, velocity, and variety. Library data resources and analytics challenges are presented. Opportunities for libraries in Big Data include expertise in metadata, assessment, and collaboration. Building a Big Data culture requires openness, investment, training, and data sharing standards. Data governance differs from data management. Machine learning and social listening are explored as examples. Trends in data science domains and tools are shared.
Lecture presented by Fernan R. Dizon at PAARL's Summer Conference on the theme "Library Analytics: Data-driven Library Management", held at Pearl Hotel, Manila on 20-22 April 2016
Mike Mertens Directions for RDM day one summaryJisc
This document discusses directions for research data management in UK universities. It focuses on the business case and sustainability for implementing research data management plans and services. Key points include identifying the need, risks of not having plans, staffing and storage costs, advocacy efforts, and long-term preservation strategies. The document also discusses incentives for researchers to properly manage data, such as reward structures, compliance monitoring, opportunities for data publication and citation, integration of support systems, and aligning job descriptions with open data practices. Overall it provides guidance on justifying research data management programs through identifying institutional needs and risks, accounting for costs and scalability, and incentivizing researcher participation.
PAARL's 1st Marina G. Dayrit Lecture Series held at UP's Melchor Hall, 5F, Proctor & Gamble Audiovisual Hall, College of Engineering, on 3 March 2017, with Albert Anthony D. Gavino of Smart Communications Inc. as resource speaker on the topic "Using Big Data to Enhance Library Services"
Similar to Moneyball, Libraries, and more - Ithaka collections presentation (20)
Library Usage as a Map for Targeting OER AdvocacyGreg Raschke
This document summarizes a study conducted by Greg Raschke, Will Cross, Sydney Thompson, John Vickery, and Lillian Rigling on targeting open educational resource (OER) advocacy efforts based on library usage data at North Carolina State University. The study analyzed textbook price, enrollment, and library usage data from 2014-2016 to identify the top 25% most expensive and highly used textbooks. It identified around 100 textbooks per semester meeting these criteria across 49 departments. The study is being used to guide OER outreach and advocacy efforts like information sessions located near relevant departments. Next steps include further student outreach and exploring expanding the criteria to capture additional opportunities.
The Experiential Library: The Transformed Libraries Impact on Teaching and Le...Greg Raschke
This document discusses how transformed libraries can impact teaching and learning through experiential learning opportunities. It provides examples of how NC State Libraries has created immersive, hands-on, and interactive learning experiences for students through spaces and programs focused on visualization, making, code and art, workshops, and visiting scholars. It acknowledges initial challenges but emphasizes leveraging opportunities through expertise, programming, and constant integration with teaching to create experiential libraries that can distinguish universities and support deeper learning.
Preparing for New Roles and Transformed Libraries: Models and ImplementationGreg Raschke
Libraries are engaging the research and teaching enterprises of their parent organizations in emerging areas. At the same time, library spaces and technologies are being adapted to support interdisciplinary collaboration, immersive learning environments, life-cycle support for research, and digitally centered scholarship. A key challenge and opportunity in realizing and fully leveraging this strategy is the transformation of the roles of subject specialists and academic technology librarians to support deeper collaboration around these emerging services and add value across the research and teaching life cycles. Transforming the skills, attitudes, goals, and priorities of practicing research librarians, while also better matching the pace of change in library program curricula to these enormous changes, is fundamental to achieving integration and value-addition at scale. This session focuses on both the imperative to increase the pace and structure of change in library curricula and for libraries to provide effective transformational development opportunities for librarians.
Digital Scholarship Spaces: Building Communities & Enabling CollaborationGreg Raschke
The document discusses new models of content creation and scholarship that are taking place in library spaces. It describes how the NC State University Libraries worked to incubate projects that brought faculty and students together to work on interdisciplinary projects. Examples included a program called Visiting Scholar that hosted lectures and workshops, and a Code+Art project that combined computer science and digital art. The document also notes some challenges in strategically managing staff resources for these new types of projects and determining how to disseminate and reproduce the works created through the programs.
Preparing Research Librarians for Transformed Libraries: Creating a Community...Greg Raschke
Preparing research and subject specialist librarians for transformed libraries through creating a community of practice. Explores informal, organic and formal, credentialed development opportunities as part of a community of practice for testing, developing, and extending research life-cycle services.
The document discusses perspectives on consortial e-book acquisitions from librarians, publishers, and vendors. It summarizes findings from analyzing e-book purchasing data from three universities. Key points of agreement and disagreement between stakeholders are outlined regarding flexible acquisition options, usage data, and resource sharing models. Current realities of consortia patron-driven acquisition models are also presented.
UNC Collaborative Book Purchasing PlanGreg Raschke
The document discusses the UNC System Libraries' collaborative book purchasing plan which aims to minimize duplication across the 16 institutions while maintaining local autonomy and shared discounts. It analyzed 11.1 million book copies held, finding 58% held by 3 or less libraries. The plan follows a buyers club model and brings institutions onto unified discount rates with YBP ranging from 18-19% for cloth firm/approval orders to 9% for continuations. estimated annual savings of $275,000+ for the system through cooperative purchasing.
Open Textbook Models: View from the LibraryGreg Raschke
The document summarizes efforts at the University of Massachusetts Amherst to promote open educational resources (OER). It discusses initial events during Open Access Weeks in 2009 and 2010 to raise awareness. A larger university event on September 21, 2010 provided education on OER. Publicity efforts gained support from campus groups. The libraries' Scholarly Communication Office and OER LibGuide provide resources to support OER adoption.
Removing Uninteresting Bytes in Software FuzzingAftab Hussain
Imagine a world where software fuzzing, the process of mutating bytes in test seeds to uncover hidden and erroneous program behaviors, becomes faster and more effective. A lot depends on the initial seeds, which can significantly dictate the trajectory of a fuzzing campaign, particularly in terms of how long it takes to uncover interesting behaviour in your code. We introduce DIAR, a technique designed to speedup fuzzing campaigns by pinpointing and eliminating those uninteresting bytes in the seeds. Picture this: instead of wasting valuable resources on meaningless mutations in large, bloated seeds, DIAR removes the unnecessary bytes, streamlining the entire process.
In this work, we equipped AFL, a popular fuzzer, with DIAR and examined two critical Linux libraries -- Libxml's xmllint, a tool for parsing xml documents, and Binutil's readelf, an essential debugging and security analysis command-line tool used to display detailed information about ELF (Executable and Linkable Format). Our preliminary results show that AFL+DIAR does not only discover new paths more quickly but also achieves higher coverage overall. This work thus showcases how starting with lean and optimized seeds can lead to faster, more comprehensive fuzzing campaigns -- and DIAR helps you find such seeds.
- These are slides of the talk given at IEEE International Conference on Software Testing Verification and Validation Workshop, ICSTW 2022.
Threats to mobile devices are more prevalent and increasing in scope and complexity. Users of mobile devices desire to take full advantage of the features
available on those devices, but many of the features provide convenience and capability but sacrifice security. This best practices guide outlines steps the users can take to better protect personal devices and information.
Alt. GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using ...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...SOFTTECHHUB
The choice of an operating system plays a pivotal role in shaping our computing experience. For decades, Microsoft's Windows has dominated the market, offering a familiar and widely adopted platform for personal and professional use. However, as technological advancements continue to push the boundaries of innovation, alternative operating systems have emerged, challenging the status quo and offering users a fresh perspective on computing.
One such alternative that has garnered significant attention and acclaim is Nitrux Linux 3.5.0, a sleek, powerful, and user-friendly Linux distribution that promises to redefine the way we interact with our devices. With its focus on performance, security, and customization, Nitrux Linux presents a compelling case for those seeking to break free from the constraints of proprietary software and embrace the freedom and flexibility of open-source computing.
Full-RAG: A modern architecture for hyper-personalizationZilliz
Mike Del Balso, CEO & Co-Founder at Tecton, presents "Full RAG," a novel approach to AI recommendation systems, aiming to push beyond the limitations of traditional models through a deep integration of contextual insights and real-time data, leveraging the Retrieval-Augmented Generation architecture. This talk will outline Full RAG's potential to significantly enhance personalization, address engineering challenges such as data management and model training, and introduce data enrichment with reranking as a key solution. Attendees will gain crucial insights into the importance of hyperpersonalization in AI, the capabilities of Full RAG for advanced personalization, and strategies for managing complex data integrations for deploying cutting-edge AI solutions.
Essentials of Automations: The Art of Triggers and Actions in FMESafe Software
In this second installment of our Essentials of Automations webinar series, we’ll explore the landscape of triggers and actions, guiding you through the nuances of authoring and adapting workspaces for seamless automations. Gain an understanding of the full spectrum of triggers and actions available in FME, empowering you to enhance your workspaces for efficient automation.
We’ll kick things off by showcasing the most commonly used event-based triggers, introducing you to various automation workflows like manual triggers, schedules, directory watchers, and more. Plus, see how these elements play out in real scenarios.
Whether you’re tweaking your current setup or building from the ground up, this session will arm you with the tools and insights needed to transform your FME usage into a powerhouse of productivity. Join us to discover effective strategies that simplify complex processes, enhancing your productivity and transforming your data management practices with FME. Let’s turn complexity into clarity and make your workspaces work wonders!
Climate Impact of Software Testing at Nordic Testing DaysKari Kakkonen
My slides at Nordic Testing Days 6.6.2024
Climate impact / sustainability of software testing discussed on the talk. ICT and testing must carry their part of global responsibility to help with the climat warming. We can minimize the carbon footprint but we can also have a carbon handprint, a positive impact on the climate. Quality characteristics can be added with sustainability, and then measured continuously. Test environments can be used less, and in smaller scale and on demand. Test techniques can be used in optimizing or minimizing number of tests. Test automation can be used to speed up testing.
Unlock the Future of Search with MongoDB Atlas_ Vector Search Unleashed.pdfMalak Abu Hammad
Discover how MongoDB Atlas and vector search technology can revolutionize your application's search capabilities. This comprehensive presentation covers:
* What is Vector Search?
* Importance and benefits of vector search
* Practical use cases across various industries
* Step-by-step implementation guide
* Live demos with code snippets
* Enhancing LLM capabilities with vector search
* Best practices and optimization strategies
Perfect for developers, AI enthusiasts, and tech leaders. Learn how to leverage MongoDB Atlas to deliver highly relevant, context-aware search results, transforming your data retrieval process. Stay ahead in tech innovation and maximize the potential of your applications.
#MongoDB #VectorSearch #AI #SemanticSearch #TechInnovation #DataScience #LLM #MachineLearning #SearchTechnology
TrustArc Webinar - 2024 Global Privacy SurveyTrustArc
How does your privacy program stack up against your peers? What challenges are privacy teams tackling and prioritizing in 2024?
In the fifth annual Global Privacy Benchmarks Survey, we asked over 1,800 global privacy professionals and business executives to share their perspectives on the current state of privacy inside and outside of their organizations. This year’s report focused on emerging areas of importance for privacy and compliance professionals, including considerations and implications of Artificial Intelligence (AI) technologies, building brand trust, and different approaches for achieving higher privacy competence scores.
See how organizational priorities and strategic approaches to data security and privacy are evolving around the globe.
This webinar will review:
- The top 10 privacy insights from the fifth annual Global Privacy Benchmarks Survey
- The top challenges for privacy leaders, practitioners, and organizations in 2024
- Key themes to consider in developing and maintaining your privacy program
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/06/building-and-scaling-ai-applications-with-the-nx-ai-manager-a-presentation-from-network-optix/
Robin van Emden, Senior Director of Data Science at Network Optix, presents the “Building and Scaling AI Applications with the Nx AI Manager,” tutorial at the May 2024 Embedded Vision Summit.
In this presentation, van Emden covers the basics of scaling edge AI solutions using the Nx tool kit. He emphasizes the process of developing AI models and deploying them globally. He also showcases the conversion of AI models and the creation of effective edge AI pipelines, with a focus on pre-processing, model conversion, selecting the appropriate inference engine for the target hardware and post-processing.
van Emden shows how Nx can simplify the developer’s life and facilitate a rapid transition from concept to production-ready applications.He provides valuable insights into developing scalable and efficient edge AI solutions, with a strong focus on practical implementation.
GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...Neo4j
Leonard Jayamohan, Partner & Generative AI Lead, Deloitte
This keynote will reveal how Deloitte leverages Neo4j’s graph power for groundbreaking digital twin solutions, achieving a staggering 100x performance boost. Discover the essential role knowledge graphs play in successful generative AI implementations. Plus, get an exclusive look at an innovative Neo4j + Generative AI solution Deloitte is developing in-house.
Enchancing adoption of Open Source Libraries. A case study on Albumentations.AIVladimir Iglovikov, Ph.D.
Presented by Vladimir Iglovikov:
- https://www.linkedin.com/in/iglovikov/
- https://x.com/viglovikov
- https://www.instagram.com/ternaus/
This presentation delves into the journey of Albumentations.ai, a highly successful open-source library for data augmentation.
Created out of a necessity for superior performance in Kaggle competitions, Albumentations has grown to become a widely used tool among data scientists and machine learning practitioners.
This case study covers various aspects, including:
People: The contributors and community that have supported Albumentations.
Metrics: The success indicators such as downloads, daily active users, GitHub stars, and financial contributions.
Challenges: The hurdles in monetizing open-source projects and measuring user engagement.
Development Practices: Best practices for creating, maintaining, and scaling open-source libraries, including code hygiene, CI/CD, and fast iteration.
Community Building: Strategies for making adoption easy, iterating quickly, and fostering a vibrant, engaged community.
Marketing: Both online and offline marketing tactics, focusing on real, impactful interactions and collaborations.
Mental Health: Maintaining balance and not feeling pressured by user demands.
Key insights include the importance of automation, making the adoption process seamless, and leveraging offline interactions for marketing. The presentation also emphasizes the need for continuous small improvements and building a friendly, inclusive community that contributes to the project's growth.
Vladimir Iglovikov brings his extensive experience as a Kaggle Grandmaster, ex-Staff ML Engineer at Lyft, sharing valuable lessons and practical advice for anyone looking to enhance the adoption of their open-source projects.
Explore more about Albumentations and join the community at:
GitHub: https://github.com/albumentations-team/albumentations
Website: https://albumentations.ai/
LinkedIn: https://www.linkedin.com/company/100504475
Twitter: https://x.com/albumentations
Large Language Model (LLM) and it’s Geospatial Applications
Moneyball, Libraries, and more - Ithaka collections presentation
1. Moneyball, the Extra 2%, and What Baseball
Management Can Teach Us About Fostering
Innovation in Managing Collections
Greg Raschke
North Carolina State University
Ithaka
March 8, 2012
4. Looking Deeper and Questioning
Assumptions
Identifying market inefficiencies.
Apply and acculturate significant
innovation.
Question long-established wisdom.
Test what is “known” with in-depth
analysis, statistical modeling, and
new approaches.
Emphasize interpersonal skills in
leveraging new knowledge and
approaches.
5. Supply-Side Collections
Print-based, unpredictable demand,
and legitimate need for just in case
collections
Lead to judging quality by size (as in
the ARL rankings) and libraries were
then held captive to this standard
Contributed to inelastic demand for
journals and a combination of
speculative and package buying
Use is secondary to size, dollars
expended, and other input measures
Credit to David Lewis
(http://ulib.iupui.edu/users/dlewis)
7. Demand-Driven Collections
Make information easily, widely,
and cheaply available
Collections as drivers of
research, teaching, and learning
To make special or unique
collections held/managed by the
library available to the user
community and the world
8. Demand-Driven – Changing Practice
Tension between time-honored role as custodians of scholarship
versus enabling digital environment for scholars
Not just PDA – portfolio of approaches, but certainly more responsive
Utilize new tools and techniques to become advanced analysts
Truly embrace evidence based decision making
Look at how collections are actually used, not at expressed need
9. Demand-Driven – More Assumptions
Less tolerance for and less
investment in lower use general
collections
Resource management based
increasingly on use
Modify collecting based on
changes in the actual use
Risks of doing nothing –
newspapers
10. Demand-Driven – Assertions
Rewards of adapting – more
used and vital than ever
Use based and user driven
collecting models will take
growing share of budget
Bet on numbers
Bet on good and quick
Put resources into enabling
digital environment for scholars
and custodian role will come out
of that strategy
11. Why So Much Data?
Data analysis is a key component in solving/managing:
Increasing pressure for accountability
Increasing capability to gather and analyze data
Increasing precision in the way we build collections and expend
resources
Advocacy
Changing practice and data analysis at NCSU
12.
13. Serials Review 2009 – Open, Data-
Driven, and Real-Time Analysis
Standardized usage data Usage ((07 usage+08 usage/2)+
(where available) (publications*10)+ (citations*5)+
Bibliometrics - publication data (Impact Factor)
and citation patterns (e.g Community Feedback ((Weighted
LJUR) Ranking x % Match) x Total #
Impact factor and eigenfactor Rankings) + 0.1 x # of "1s“
User community feedback via Price/feedback value
interactive, database-driven Price/use
applications
Merge results to filter out top 20%
Weigh/calculate/quantify user
feedback and bottom 20%
Weigh price against multiple
data points
14. Looking closer – Finding balance
An example - a closer look at print item usage
Traditional ILS reporting tools can make this difficult
Advanced analytical tools can help
What types of questions can we ask?
Should Patron-Driven records not purchased be purged after 2 years?
How does print item usage break down?
Do print items even get used?
15. If it’s not used after 2 years…
Should PDA records
be purged?
Maybe…
We haven’t even hit
50% usage
But what if we take a
longer view…
16. If it’s not used after 2 years…
Things begin to
look different
17. Looking even closer…
How does print
item use break
down?
Single circ
usage is
consistently
~14%
Would this
change in a
PDA only
world?
22. Measurable Uses of the Collection 2009/2010
Measurable Uses of the Collection 2009/2010
Full-text journal downloads* 3,672,600
Database use 1,989,972
Print book circulations/renewals 525,430
Digital collections requests 471,403
E-books 149,815
Reserves** 327,267
Total Uses 7,136,487
* Includes use of NC LIVE full-text content
** Includes textbook, print, and e-reserves usage
23. From Assumptions to Assertions to
Practice
Grow/develop/hire analysts.
Adapt statistical tools such as SAS software.
Partner with digital library/technologists.
Develop positive arbitrage.
Put resources into enabling digital environment for scholars.
Experiment – budget for it, reward it.
Work hard to get the faculty to buy into new approaches.
Combine analytical approaches with the people skills .
“…there was a bias toward what people saw with their own eyes,
or thought they had seen. The human mind played tricks on itself
when it relied exclusively on what it saw, and every trick it played
was a financial opportunity for someone who saw through the
illusion to the reality”.
Editor's Notes
Moneyball, the Extra 2%, and What Baseball Management Can Teach Us About Fostering Innovation in Managing Collections I have been thinking about the intersection of modern baseball management with innovation in managing collections quite a bit and have been trying to have a bit of fun with it. I was not sure the Moneyball analogy completely worked, but then Anne Kenney brought it up independently in a conversation about libraries and Brad Pitt made it into a movie – so I pressed-on assuming if it is good enough for Anne and Brad - then it is a good enough comparison for me. You might conclude that this is a forced analogy that means I spend too much time thinking about baseball and my wife would whole-heartedly agree, but I do think the comparisons about innovation developing and managing baseball teams and research library collections work well. There is much we can learn. Advance slide.
This is the title my boss recommended. Advance slide.
Briefly - Moneyball is the story of the success of the Oakland A’s in the late 1990’s and early 2000’s and The Extra 2% is the less popular, but just as important, story of the success of the Tampa Bay Rays from 2007 to today. Both teams had to struggle against inferior revenue streams, dysfunctional markets, and competition with significantly better financing and resource basis – so the New York Yankees and Boston Red Sox for example. Both had to find new approaches to building and running a successful baseball team in order to compete, maintain vitality, and be successful in serving their primary users – fans. I would argue that today many libraries find themselves in a very similar position.
My wife asked me if this guy was supposed to be a metaphor for looking deeper and analytical thinking and I can tell you no. In baseball looking deeper and questioning assumptions means questioning reliance on old fashioned numbers and axioms that do not tell you much about past, current, or likely future success - using statistical analysis to identify market inefficiencies. It also means not relying on grizzled hardened scouts – that is what this guy is – who think can predict performance based on what they see and rely on paradigms that dominated the game until the late 1990's rather than what they can test, question, analyze, and project based on good information - and then breeding that into every element of the organizational culture. I suppose this guy could easily be a grizzled hardened bibliographer as well – maybe he works in one of your libraries, but I doubt it – most of us are well past the grizzled back-room selector models, but we still hold onto many of those old practices as we turn our big ships to fully embrace the digital environment. Many of us are changing practice and working to fully embrace digital content as the primary means for discovering and disseminating information, but we are not doing it fast enough and would benefit from looking outside of libraries for ideas to expedite the pace of innovation in collections. Here are some specific ideas I think research libraries can increasingly employ.
Peak of supply-side collections occurred in 1990's. Print-based, unpredictable demand, and legitimate need for just in case collections. Lead to judging quality by size (as in the ARL rankings) and libraries were then held captive to this standard. Contributed to inelastic demand for journals and a combination of speculative and package buying. Use is secondary to size, dollars expended, and other input measures. Credit to David Lewis (http://ulib.iupui.edu/users/dlewis).
The shift from supply-side development of collections (print-based, large inventories, unpredictable supply and demand periods, and judging collections largely on size) to more directly demand-driven approach has been underway for quite a while, but the pace is definitely accelerating for many reasons, including: Technology and increasing amount of content on open networks Changes in publishing Supply-chain capabilities and print-on-demand Increased accountability for libraries and higher education, but One reason trumps all others – economics The combination of dysfunctional markets for scholarly information and significant economic downturn have brought many of us to the level of the Oakland A’s and Tampa Bay Rays – inexorable economic factors driving us to increasingly innovate in our management of collections, question established norms, and develop new models for managing collections.
Collections for the sake of collections is dead, but much of the mission for academic and research library remains... Make information easily, widely, and cheaply available Collections as drivers of research, teaching, and learning To make special or unique collections held/managed by the library available to the user community and the world
Tension between time-honored role as custodians of scholarship versus enabling digital environment for scholars. Not just PDA – portfolio of approaches, but certainly more responsive and more analytical. Utilize new tools and techniques to become advanced analysts. Truly embrace evidence based decision making. Look at how collections are actually used, not at expressed need.
More assumptions Less tolerance for and less investment in lower use general collections Resource management based increasingly on use Modify collecting based on changes in the actual use Risks of doing nothing – newspapers “ I don’t know. Nobody knows. We’re collectively living through 1500, when it’s easier to see what’s broken than what will replace it. The internet turns 40 this fall. Access by the general public is less than half that age. Web use, as a normal part of life for a majority of the developed world, is less than half that age. We just got here. Even the revolutionaries can’t predict what will happen.” – Clay Shirky If we can not predict - then how can we at least better understand what is going on with the changing nature of scholarship.
Anybody know who these guys are - good - random fantasy football draft picture I pulled off of Google. Tell fantasy football story. Use based and user driven collecting models will take growing share of budget Bet on numbers Bet on good and quick Put resources into enabling digital environment for scholars and custodian role will come out of that strategy Rewards of adapting – more used and vital than ever - Ithaka studies are interesting - libraries need healthy fear they provide - but best do not need to worry - gateway role can decline if demonstrate use
Data analysis is a key component in solving/managing: Increasing pressure for accountability Increasing capability to gather and analyze data Increasing precision in the way we build collections and expend resources Advocacy Changing practice and data analysis at NCSU It is also something that not everyone was doing - so you can reap some competitive advantage.
We developed a database to solicit community input on journal titles, weight community input, and pair it with eigenfactor, snip, usage data, publication rates, and citation rates among our community to develop a weighted valuation for each journal title in our collection. Ivy and colleagues at CDL have done much of this work as well. Piggybacked on work at University of Washington and CDL to roll the data up to analyze, manage, and negotiate packages with publishers
13,000+ points of data from 700 users – how do you at least run an initial filter through that data? Relationships between usage data and community feedback data. Way more open and data-driven process than ever before where capturing data and feedback and analyzing it in real-time. We developed our ERM system to ingest statistics related to every continuing resource we own and enables a myriad of reports based on usage rates and expenditures - and here is the key - we use it consistently, not just when we have cuts or journal reviews and annually review the bottom 20% of our continuing resources based on overall performance and make decisions accordingly to enable investment in emerging areas.
We put our entire book collection and circulation activity into SAS. We ran a myriad of statistical analyses that have informed everything from cutting approval plan expenditures down by 45%, scoping our demand-driven e-book collections program, and heavily influencing and reducing our item-by-item book selection. Saved hundreds of thousands annually on our various book collecting efforts, improved our usage rates, broadened our scope in targeted subjects, and reinvested in new areas.
We developed a database where we actively map collections expenditures by fund/subject to university data points such as number of phd’s, faculty members, enrollment, grant dollars, and can manipulate those at the research center, departmental, and college level. We can use that real-time information to mine for overvalued areas, under-invested subjects, test ourselves against university expenditures and trends, and advocate for the preservation and addition of library resources. It has been very effective in balancing our collections portfolio and helping to advocate for more funds.
Moving quickly from changing assumptions to assertions and changing practice. Do whatever you have to do to grow, develop, or hire statistical analysts to be part of your collections team. Change your staff and your staffing models with at least one new hire. People are the most important tool for innovation - that is nothing new - but it requires real change with staff to truly change culture and breed innovation. Statistical tools - build, adopt, adapt, and start using them broadly across the collections program. Some brief examples. Partnering with digital library colleagues. Digital library innovation has not been applied enough to the business of building and managing collections. We need to do more to foster collaboration among collections and digital library colleagues. A couple of quick examples. These efforts have been imperfect and we can get better - but we are making progress and taking Steve Jobs advice and stealing ideas from other libraries whenever we can. Develop positive arbitrage Positive arbitrage is a concept at the heart of The Extra 2%. the practice of taking advantage of a price difference between two or more markets or a transaction that involves no negative cash flow at any probabilistic or temporal state and a positive cash flow in at least one state. It's use in the book and my use here is kind of a bastardization of the term, but the key concept is finding and generating value in everything you do with collections. Library collections work has excelled at creating negative arbitrage and value – we exchanged dollars for books that did not circulate, purchased journals we did not want, and often signed contracts that favored the publishers and vendors we work with. Sometimes there are good reasons for doing this, but often it was due to basing decisions on past practice and not having the type of data and analysis that would help us find the best value in expending collections resources – help us find the extra 2% or 5% or 10% to maximize the impact of collections. Free collections funds to invest in new and emerging areas. Free funds to invest in digital media, collaborative efforts, and the kinds of great digital library projects we have heard about here that are non-traditional, but critical areas of value for scholars and libraries. If you budget for experimentation and reward innovation (even innovative failures) - then you can reinforce your efforts to change the staff, change the culture to consistently question assumptions, experiment with new approaches, generate innovation, and bring more content and value to users. Both of the teams referenced today had people with dynamic interpersonal skills to carry forward the combination of analytical people and analytical tools to drive collections work. They met resistance as some constituencies dug in, but persevered by demonstrating impact and working the necessary political channels to create buy-in. Libraries tend to be very good at generating positive political arbitrage with faculty and other users. That needs to be maintained as we work on new models. For analytical approaches to be successfully applied in local environments - the user community and especially faculty - have to accept them in time. Analysis does not obviate the political and interpersonal work of managing collections, but I am certain that if we use analytical tools to increase access to the collections users want in the format in which they want to consume them, then just like winning in baseball, your fan-base will buy-in.