Brown, Christopher C. “Harvesting HathiTrust Documents: A New Model for Online Access.” Presentation given at the 2011 Missouri Government Documents Conference, 7 June 2011, Columbia, MO.
An evaluation of taxonomic name finding & next steps in Biodiversity Heritage...Chris Freeland
This document discusses an evaluation of taxonomic name finding in the Biodiversity Heritage Library (BHL) and next steps. The evaluation found that the TaxonFinder algorithm identified names with 28.2% precision and 23.3% recall, while accounting for optical character recognition errors. To improve name finding, the document recommends enhancing fuzzy retrieval algorithms, adding exception rules to overcome errors, and conducting more evaluations. It also outlines plans to create an article repository and additional services to further develop the BHL.
Biodiversity informatics: digitising the living worldRoderic Page
The document discusses biodiversity informatics and efforts to digitize information about the living world. It notes that GBIF has over 500 million specimen records but that none of the Kenya data actually comes from Kenya. It also discusses challenges with the Biodiversity Heritage Library in emphasizing older publications over recent scientific articles and representing information by scanned book pages rather than articles. The document raises issues about ensuring all scientific names are published and unique, and about making publications containing names accessible online.
CATA is exploiting decreasing information costs and the evolution of social networks to transition its business model. It is adding a fee-for-knowledge model to its existing fee-for-membership model in order to move customers away from passively belonging to CATA and toward gaining value through transactional knowledge sharing on CATA's social network. This new approach is paying off as major companies like BMW and governments are utilizing CATA's social network and knowledge-sharing services.
Nirvana fue una influyente banda de rock estadounidense formada en los años 80 por Kurt Cobain y Krist Novoselic. Sus álbumes Bleach, Nevermind y In Utero cosecharon gran éxito comercial y los convirtió en uno de los grupos de rock más importantes de los 90. La banda sufrió varios cambios de bateristas antes de consolidar su formación clásica con Dave Grohl.
In this presentation the following topics are covered: How to use different formats in WordPress, Configuring WordPress, Selecting a template, Creating privacy, discussion and permalink settings and also deciding on a content strategy along with how to choose concise keywords for a blog.
An evaluation of taxonomic name finding & next steps in Biodiversity Heritage...Chris Freeland
This document discusses an evaluation of taxonomic name finding in the Biodiversity Heritage Library (BHL) and next steps. The evaluation found that the TaxonFinder algorithm identified names with 28.2% precision and 23.3% recall, while accounting for optical character recognition errors. To improve name finding, the document recommends enhancing fuzzy retrieval algorithms, adding exception rules to overcome errors, and conducting more evaluations. It also outlines plans to create an article repository and additional services to further develop the BHL.
Biodiversity informatics: digitising the living worldRoderic Page
The document discusses biodiversity informatics and efforts to digitize information about the living world. It notes that GBIF has over 500 million specimen records but that none of the Kenya data actually comes from Kenya. It also discusses challenges with the Biodiversity Heritage Library in emphasizing older publications over recent scientific articles and representing information by scanned book pages rather than articles. The document raises issues about ensuring all scientific names are published and unique, and about making publications containing names accessible online.
CATA is exploiting decreasing information costs and the evolution of social networks to transition its business model. It is adding a fee-for-knowledge model to its existing fee-for-membership model in order to move customers away from passively belonging to CATA and toward gaining value through transactional knowledge sharing on CATA's social network. This new approach is paying off as major companies like BMW and governments are utilizing CATA's social network and knowledge-sharing services.
Nirvana fue una influyente banda de rock estadounidense formada en los años 80 por Kurt Cobain y Krist Novoselic. Sus álbumes Bleach, Nevermind y In Utero cosecharon gran éxito comercial y los convirtió en uno de los grupos de rock más importantes de los 90. La banda sufrió varios cambios de bateristas antes de consolidar su formación clásica con Dave Grohl.
In this presentation the following topics are covered: How to use different formats in WordPress, Configuring WordPress, Selecting a template, Creating privacy, discussion and permalink settings and also deciding on a content strategy along with how to choose concise keywords for a blog.
Presentation given at the DASISH Workshop on PID services, December 8-9., 2014, GESIS Leibniz Institute of Social Sciences, Cologne.
Presentation of DataCite, its history, structure and services followed by a short introduction of the DOI registration service for social and economic data (da|ra).
The presentation explores the trend towards a scholarly communication system that is friendly to machines. It presents 3 exhibits illustrating the trend and 1 exhibit illustrating inertia in the system. It makes the point that machine-actionability can be much easier achieved if content and metadata are available in Open Access and under a permissive Creative Commons license. It also observes that even with content and metadata openly available, new costs related to advanced tools to explore the scholarly record will emerge. Finally, it points at significant challenges regarding the persistence of the scholarly record in light of increasingly interconnected and actionable content and advanced tools to interact with it.
The slides were used for a plenary presentation at the LIBER 2011 Conference in Barcelona, Spain, on June 30 2011.
BHL Technologies: Review for BHL-AustraliaChris Freeland
This document summarizes the history, technologies, and services of the Biodiversity Heritage Library (BHL). It discusses how BHL was developed from earlier projects, its usage statistics, scanning and content acquisition methods, data mining capabilities, and APIs. It provides details on name finding, CiteBank integration, and opportunities to expand BHL into a global resource through new technologies and partnerships.
Bringing Digital Humanities to the wider public: libraries as incubator for D...Martijn Kleppe
Keynote presented at Language Technologies & Digital Humanities Conferences 2018, Ljubljana, Slovenia.
Abstract:
Digital Humanities researchers rely on large digital datasets. Since the National Library of the Netherlands (KB) has been digitizing its collection for about ten years, their datasets are popular amongst DH scholars that focus on historical newspapers, periodicals and books. By not only supporting researchers by giving them access to datasets, but also by collaborating with them, the KB aims to incorporate DH research results in their services and products. We do this by sharing our prototype tools and code on our online Lab , invite academics to come and work as researcher-in-residence and are full partner in research projects. In this talk I will describe the challenges and opportunities for libraries and academics when they collaborate. What can researchers gain from collaborating with libraries? And how can libraries bring the affordances of DH research to the wider public?
This document discusses the transition of libraries and their resources to cloud-based systems. It notes that while physical library usage like book circulation and reference visits have declined, the number of visitors to libraries has increased. This suggests libraries are becoming more of a community gathering place. The document outlines some cloud-based tools that can be used for collaboration and storage. It also discusses how cloud computing will impact academia through more flexible access to computing resources and the potential for institutions to share costs and talents. However, it notes there are also security, privacy and control challenges to consider with cloud-based systems.
This document discusses building an online profile as a scientist in the era of big data and open science. It begins with an overview of the speaker's background working in academia, industry, and as an entrepreneur. The speaker then discusses various online tools and platforms that scientists can use to share their work and expertise, such as ORCID, LinkedIn, Google Scholar, SlideShare, and ResearchGate. He emphasizes the importance of making contributions openly available online in order to increase visibility and measure impact through alternative metrics. The speaker also provides examples of using these tools to showcase his own career and publications.
Workshop 5: Uptake of, and concepts in text and data miningRoss Mounce
Content mining involves large-scale computer-aided information extraction from various types of digital content such as text, images, videos, and metadata. It can be used to extract useful information from the vast amounts of scholarly literature available online. Some examples of content mining include recomputing statistical tests reported in papers, finding recent publications using specimens from museums, and identifying associations between weevils and their host plants mentioned together in papers. However, much of the potential of content mining is not realized due to challenges such as fragmented publication of literature across many platforms, lack of standardized formats like XML that enable sophisticated searches, and publishers not making full text and metadata openly available.
The Rolex Learning Center at EPFL: a new building for a new vision in collect...Thomas Guignard
The new Rolex Learning Center at EPFL provides a new vision for collection development and library services. It includes 15,000 square meters of work spaces, group study rooms, cafes, and other amenities to serve as a hub for learning, research, and collaboration. The building reflects evolving teaching models, libraries, and acquisition models toward more digital resources and on-demand access. It aims to provide a seamless experience for patrons across physical and digital collections.
This document summarizes a NISO webinar on guidelines and resources for developing data access plans in response to the Office of Science and Technology Policy's (OSTP) 2013 memo. The memo directs large federal funding agencies to develop public access plans for research results. The webinar outlines the required elements of these plans and provides existing guidelines and resources that can help agencies meet digital data requirements, such as standards for data dissemination, description, and long-term preservation. Speakers from the Interuniversity Consortium for Political and Social Research discuss how agencies can leverage existing infrastructure and best practices to develop plans that maximize access to and reuse of federal research data.
This document summarizes the costs and benefits of preserving digital research data. It discusses how most data preservation activities are publicly funded and how costs include personnel, hardware, and software costs. While costs can be difficult to quantify, benefits include scientific progress through data reuse and valorization. The value of data increases over time through reuse, but preserving data leads to increasing storage and access costs as data volumes grow. Data archiving services may charge for data reuse or deposit. The document also introduces the Data Archiving and Networked Services (DANS) organization, its activities in archiving datasets, and policies around earning money from archived data.
DataCite is a global consortium that provides standards and services to make research data findable, accessible, and citable. It assigns digital object identifiers (DOIs) to datasets and other research works to improve how they are located and referenced. DataCite was founded in 2009 and is supported by many research institutions and libraries around the world. It develops best practices for data publishing and sharing and operates metadata and search services to help researchers discover and access data.
DataCite - services and support for opening up research dataHerbert Gruttemeier
This document provides an overview of DataCite, a global consortium that provides services and support for opening up research data. DataCite maintains the DataCite Metadata Store (MDS) where members can mint Digital Object Identifiers (DOIs) and register metadata for datasets and other research objects. It also operates services like the DataCite Metadata Search, an OAI provider for harvesting metadata, and statistics on DOI registration and resolution. DataCite aims to improve the scholarly infrastructure around research data by establishing standards, best practices, and working with data centers and repositories.
This document discusses the Biodiversity Heritage Library (BHL) project. It provides statistics on the size and usage of the BHL digital library, describes the workflow and technologies used to digitize materials. It also summarizes services provided by BHL, including name finding, an application programming interface (API), and links to citation databases. Plans are outlined to improve display of articles and develop new tools for working with bibliographies and citations.
Interlinking Data and Knowledge in Enterprises, Research and Society with Lin...Christoph Lange
The Linked Data paradigm has emerged as a powerful enabler for data and knowledge interlinking and exchange using standardised Web technologies.
In this article, we discuss our vision how the Linked Data paradigm can be employed to evolve the intranets of large organisations -- be it enterprises, research organisations or governmental and public administrations -- into networks of internal data and knowledge.
In particular for large enterprises data integration is still a key challenge. The Linked Data paradigm seems a promising approach for integrating enterprise data. Like the Web of Data, which now complements the original document-centred Web, data intranets may help to enhance and flexibilise the intranets and service-oriented architectures that exist in large organisations. Furthermore, using Linked Data gives enterprises access to 50+ billion facts from the growing Linked Open Data (LOD) cloud. As a result, a data intranet can help to bridge the gap between structured data management (in ERP, CRM or SCM systems) and semi-structured or unstructured information in documents, wikis or web portals, and make all of these sources searchable in a coherent way.
Keynote at Baltic DB&IS 2014, 9 June 2014, Tallinn, Estonia
The document summarizes a talk given by Dr. Johannes Keizer on the CIARD (Coherence in Information for Agricultural Research for development) initiative and a global infrastructure for linked open data (LOD). The CIARD initiative aims to provide open access to agricultural research by promoting standards and sharing information. It involves institutions contributing their research outputs through the CIARD RING and adopting standards. The infrastructure proposed includes distributed repositories linked through vocabularies and LOD. Tools are being developed to generate LOD and link datasets through shared concepts.
This is one out of a series of presentations which I have given during a recent trip to the United States. I will make them all public, but content does not vary a lot between some of them
Building a Vast Library of Life: The Biodiversity Heritage Library Looks to t...Martin Kalfatovic
The Biodiversity Heritage Library (BHL) is a global digital library containing over 159,000 items and 45 million pages of biodiversity literature. Started in 2006 with 10 institutions, BHL has grown to include 23 member institutions and affiliates worldwide. BHL aims to make biodiversity literature openly accessible online to support research and address the "taxonomic impediment". It does this through mass digitization of materials and partnerships to digitize copyrighted works. BHL has seen large growth in both its digital collections and global users, while continuing to add new content types and features to better serve the biodiversity community.
contentmine.org (funded by Shuttleworth Foundation) has developed tools and workshops to allow anyone to mine scientific content. This 10-minute presentation at Wellcome Trust encourages you to become involved - no previous knowledge required.
Migrating Government Publications without Going South: Our Alma/Primo ExperienceChristopher Brown
This document discusses the University of Denver Main Library's migration from their previous integrated library system to Alma/Primo and the challenges of providing access to their government documents collection. Key points:
1) All of the library's government documents (~850,000 items) were moved off-site, making browsing the physical collection impossible.
2) Primo's virtual browse feature for government documents classified using the SuDocs system was inadequate.
3) The librarian created a custom "Browse the Documents Stacks" feature using a Microsoft Access database and Springshare's Libguides to virtually replicate browsing the collection in SuDocs order.
4) This solution helped users find documents, but the librarian
Downsizing Your Depository: Dealing with Mandates from Your AdministrationChristopher Brown
This document discusses strategies for downsizing a federal depository collection in response to an administration mandate. It recommends first amending item selections to receive only online materials. It then discusses analyzing current holdings to identify materials for removal, notifying regional libraries, and using various tools to generate Needs and Offers lists for legacy print materials. Specific series like Congressional Hearings that present challenges are also addressed. Working closely with the regional library is emphasized throughout the downsizing process.
More Related Content
Similar to Harvesting HathiTrust Documents: A New Model for Online Access
Presentation given at the DASISH Workshop on PID services, December 8-9., 2014, GESIS Leibniz Institute of Social Sciences, Cologne.
Presentation of DataCite, its history, structure and services followed by a short introduction of the DOI registration service for social and economic data (da|ra).
The presentation explores the trend towards a scholarly communication system that is friendly to machines. It presents 3 exhibits illustrating the trend and 1 exhibit illustrating inertia in the system. It makes the point that machine-actionability can be much easier achieved if content and metadata are available in Open Access and under a permissive Creative Commons license. It also observes that even with content and metadata openly available, new costs related to advanced tools to explore the scholarly record will emerge. Finally, it points at significant challenges regarding the persistence of the scholarly record in light of increasingly interconnected and actionable content and advanced tools to interact with it.
The slides were used for a plenary presentation at the LIBER 2011 Conference in Barcelona, Spain, on June 30 2011.
BHL Technologies: Review for BHL-AustraliaChris Freeland
This document summarizes the history, technologies, and services of the Biodiversity Heritage Library (BHL). It discusses how BHL was developed from earlier projects, its usage statistics, scanning and content acquisition methods, data mining capabilities, and APIs. It provides details on name finding, CiteBank integration, and opportunities to expand BHL into a global resource through new technologies and partnerships.
Bringing Digital Humanities to the wider public: libraries as incubator for D...Martijn Kleppe
Keynote presented at Language Technologies & Digital Humanities Conferences 2018, Ljubljana, Slovenia.
Abstract:
Digital Humanities researchers rely on large digital datasets. Since the National Library of the Netherlands (KB) has been digitizing its collection for about ten years, their datasets are popular amongst DH scholars that focus on historical newspapers, periodicals and books. By not only supporting researchers by giving them access to datasets, but also by collaborating with them, the KB aims to incorporate DH research results in their services and products. We do this by sharing our prototype tools and code on our online Lab , invite academics to come and work as researcher-in-residence and are full partner in research projects. In this talk I will describe the challenges and opportunities for libraries and academics when they collaborate. What can researchers gain from collaborating with libraries? And how can libraries bring the affordances of DH research to the wider public?
This document discusses the transition of libraries and their resources to cloud-based systems. It notes that while physical library usage like book circulation and reference visits have declined, the number of visitors to libraries has increased. This suggests libraries are becoming more of a community gathering place. The document outlines some cloud-based tools that can be used for collaboration and storage. It also discusses how cloud computing will impact academia through more flexible access to computing resources and the potential for institutions to share costs and talents. However, it notes there are also security, privacy and control challenges to consider with cloud-based systems.
This document discusses building an online profile as a scientist in the era of big data and open science. It begins with an overview of the speaker's background working in academia, industry, and as an entrepreneur. The speaker then discusses various online tools and platforms that scientists can use to share their work and expertise, such as ORCID, LinkedIn, Google Scholar, SlideShare, and ResearchGate. He emphasizes the importance of making contributions openly available online in order to increase visibility and measure impact through alternative metrics. The speaker also provides examples of using these tools to showcase his own career and publications.
Workshop 5: Uptake of, and concepts in text and data miningRoss Mounce
Content mining involves large-scale computer-aided information extraction from various types of digital content such as text, images, videos, and metadata. It can be used to extract useful information from the vast amounts of scholarly literature available online. Some examples of content mining include recomputing statistical tests reported in papers, finding recent publications using specimens from museums, and identifying associations between weevils and their host plants mentioned together in papers. However, much of the potential of content mining is not realized due to challenges such as fragmented publication of literature across many platforms, lack of standardized formats like XML that enable sophisticated searches, and publishers not making full text and metadata openly available.
The Rolex Learning Center at EPFL: a new building for a new vision in collect...Thomas Guignard
The new Rolex Learning Center at EPFL provides a new vision for collection development and library services. It includes 15,000 square meters of work spaces, group study rooms, cafes, and other amenities to serve as a hub for learning, research, and collaboration. The building reflects evolving teaching models, libraries, and acquisition models toward more digital resources and on-demand access. It aims to provide a seamless experience for patrons across physical and digital collections.
This document summarizes a NISO webinar on guidelines and resources for developing data access plans in response to the Office of Science and Technology Policy's (OSTP) 2013 memo. The memo directs large federal funding agencies to develop public access plans for research results. The webinar outlines the required elements of these plans and provides existing guidelines and resources that can help agencies meet digital data requirements, such as standards for data dissemination, description, and long-term preservation. Speakers from the Interuniversity Consortium for Political and Social Research discuss how agencies can leverage existing infrastructure and best practices to develop plans that maximize access to and reuse of federal research data.
This document summarizes the costs and benefits of preserving digital research data. It discusses how most data preservation activities are publicly funded and how costs include personnel, hardware, and software costs. While costs can be difficult to quantify, benefits include scientific progress through data reuse and valorization. The value of data increases over time through reuse, but preserving data leads to increasing storage and access costs as data volumes grow. Data archiving services may charge for data reuse or deposit. The document also introduces the Data Archiving and Networked Services (DANS) organization, its activities in archiving datasets, and policies around earning money from archived data.
DataCite is a global consortium that provides standards and services to make research data findable, accessible, and citable. It assigns digital object identifiers (DOIs) to datasets and other research works to improve how they are located and referenced. DataCite was founded in 2009 and is supported by many research institutions and libraries around the world. It develops best practices for data publishing and sharing and operates metadata and search services to help researchers discover and access data.
DataCite - services and support for opening up research dataHerbert Gruttemeier
This document provides an overview of DataCite, a global consortium that provides services and support for opening up research data. DataCite maintains the DataCite Metadata Store (MDS) where members can mint Digital Object Identifiers (DOIs) and register metadata for datasets and other research objects. It also operates services like the DataCite Metadata Search, an OAI provider for harvesting metadata, and statistics on DOI registration and resolution. DataCite aims to improve the scholarly infrastructure around research data by establishing standards, best practices, and working with data centers and repositories.
This document discusses the Biodiversity Heritage Library (BHL) project. It provides statistics on the size and usage of the BHL digital library, describes the workflow and technologies used to digitize materials. It also summarizes services provided by BHL, including name finding, an application programming interface (API), and links to citation databases. Plans are outlined to improve display of articles and develop new tools for working with bibliographies and citations.
Interlinking Data and Knowledge in Enterprises, Research and Society with Lin...Christoph Lange
The Linked Data paradigm has emerged as a powerful enabler for data and knowledge interlinking and exchange using standardised Web technologies.
In this article, we discuss our vision how the Linked Data paradigm can be employed to evolve the intranets of large organisations -- be it enterprises, research organisations or governmental and public administrations -- into networks of internal data and knowledge.
In particular for large enterprises data integration is still a key challenge. The Linked Data paradigm seems a promising approach for integrating enterprise data. Like the Web of Data, which now complements the original document-centred Web, data intranets may help to enhance and flexibilise the intranets and service-oriented architectures that exist in large organisations. Furthermore, using Linked Data gives enterprises access to 50+ billion facts from the growing Linked Open Data (LOD) cloud. As a result, a data intranet can help to bridge the gap between structured data management (in ERP, CRM or SCM systems) and semi-structured or unstructured information in documents, wikis or web portals, and make all of these sources searchable in a coherent way.
Keynote at Baltic DB&IS 2014, 9 June 2014, Tallinn, Estonia
The document summarizes a talk given by Dr. Johannes Keizer on the CIARD (Coherence in Information for Agricultural Research for development) initiative and a global infrastructure for linked open data (LOD). The CIARD initiative aims to provide open access to agricultural research by promoting standards and sharing information. It involves institutions contributing their research outputs through the CIARD RING and adopting standards. The infrastructure proposed includes distributed repositories linked through vocabularies and LOD. Tools are being developed to generate LOD and link datasets through shared concepts.
This is one out of a series of presentations which I have given during a recent trip to the United States. I will make them all public, but content does not vary a lot between some of them
Building a Vast Library of Life: The Biodiversity Heritage Library Looks to t...Martin Kalfatovic
The Biodiversity Heritage Library (BHL) is a global digital library containing over 159,000 items and 45 million pages of biodiversity literature. Started in 2006 with 10 institutions, BHL has grown to include 23 member institutions and affiliates worldwide. BHL aims to make biodiversity literature openly accessible online to support research and address the "taxonomic impediment". It does this through mass digitization of materials and partnerships to digitize copyrighted works. BHL has seen large growth in both its digital collections and global users, while continuing to add new content types and features to better serve the biodiversity community.
contentmine.org (funded by Shuttleworth Foundation) has developed tools and workshops to allow anyone to mine scientific content. This 10-minute presentation at Wellcome Trust encourages you to become involved - no previous knowledge required.
Similar to Harvesting HathiTrust Documents: A New Model for Online Access (20)
Migrating Government Publications without Going South: Our Alma/Primo ExperienceChristopher Brown
This document discusses the University of Denver Main Library's migration from their previous integrated library system to Alma/Primo and the challenges of providing access to their government documents collection. Key points:
1) All of the library's government documents (~850,000 items) were moved off-site, making browsing the physical collection impossible.
2) Primo's virtual browse feature for government documents classified using the SuDocs system was inadequate.
3) The librarian created a custom "Browse the Documents Stacks" feature using a Microsoft Access database and Springshare's Libguides to virtually replicate browsing the collection in SuDocs order.
4) This solution helped users find documents, but the librarian
Downsizing Your Depository: Dealing with Mandates from Your AdministrationChristopher Brown
This document discusses strategies for downsizing a federal depository collection in response to an administration mandate. It recommends first amending item selections to receive only online materials. It then discusses analyzing current holdings to identify materials for removal, notifying regional libraries, and using various tools to generate Needs and Offers lists for legacy print materials. Specific series like Congressional Hearings that present challenges are also addressed. Working closely with the regional library is emphasized throughout the downsizing process.
This talk from Chris Brown will focus on tools, techniques, and ideas for downsizing your documents collection. Various themes will be addressed, including: item deselection, weeding tools, offsite storage, DDM2, and getting rid of “tonnage”. Chris will also introduce you to his own item selection/deselection tool.
Web-scale Discovery Tools and the Backgrounding of Government InformationChristopher Brown
Brown, Christopher. C. “Web-scale Discovery Tools and the Backgrounding of Government Information.” Presentation given at the Electronic Resources & Libraries Conference, 23 February 2015, Austin, TX.
Web-scale discovery tools have opened up new realms of discovery for libraries and their electronic resources. However, government information is being short-changed and backgrounded in these tools. Find out why this is happening, the diagnostic tests to show that this is so, and what we can do about it.
Brown, Christopher C. “The Darkening of Government Information.” Presentation given at the Western States Government Information Virtual Conference, 7 August 2014, online.
Collecting Usage Statistics for E-Government ResourcesChristopher Brown
Brown, Christopher C. “Collecting Usage Statistics for E-Government Resources.” Online Webinar, Federal Depository Library Program. Presentation through the U.S. Government Printing Office iCohere platform, 20 May 2014.
Outbound Harvesting with Encore as a Library Space-Saving Strategy : The Cas...Christopher Brown
Brown, Christopher C. “Outbound Harvesting with Encore as a Library Space-Saving Strategy : The Case of HathiTrust Docs.” Presentation given at the Innovative Users Group at ALA Midwinter, 7 January 2011, San Diego, CA.
Brown, Christopher C. “Item Deselection on the Fast Track.” Presentation given at the 2011 Missouri Government Documents Conference, 6 June 2011, Columbia, MO.
Going All-Electronic and Keeping Track of It: Clickthrough Statistics for On...Christopher Brown
Brown, Christopher C. “Going All-Electronic and Keeping Track of It: Clickthrough Statistics for Online Document Usage.” Presentation given at the 2011 Missouri Government Documents Conference, 7 June 2011, Columbia, MO.
The Three Googles: How I Teach Google in an Academic SettingChristopher Brown
Brown, Christopher C. “The Three Googles: How I Teach Google in an Academic Setting.” Presentation given at the CoALA Spring Workshop, 10 May 2013, online.
Brown, Christopher C. “The Front Face of the ERM: How we Left Our Home-Grown
Database Management System and Enbraced a More Innovative One.” Presentation
given at the Innovative Users Group 2013, 25 April 2013, San Francisco, CA.
Planning the Six-State Virtual Government Information ConferenceChristopher Brown
Brown, Christopher C., Janet Fisher, Peggy Jobe and Jennie Gerke. Planning the Six-State Virtual Government Information Conference. Presentation given at the Fall 2012 Depository Library Conference, 16 October 2012, Arlington, VA.
Fiche Online: A Vision for Digitizing All Documents FicheChristopher Brown
Brown, Christopher C. Fiche Online: A Vision for Digitizing All Documents Fiche. Presentation given at the Fall 2012 Depository Library Conference, 15 October 2012, Arlington, VA.
Brown, Christopher C. “Summon and the Art of Discovery.” Presentation given at the
Serials Solutions Western Regional Meeting, 14 September 2012, Aurora, CO.
When there is no Vendor: Statistics for Free Clickthroughs via the Online Cat...Christopher Brown
The document describes a system implemented at the University of Denver to track clickthrough statistics for online resources accessible through the library catalog. For over 8 years, DU has tracked clicks on URLs in catalog records to measure usage of free resources like government documents and publications from other organizations. This provides statistics not available from vendors to help with collection decisions. The system prepends URLs with a code that redirects users through a server collecting data before sending them to the final site. This allows analysis of what resources are most popular and identification of broken links to improve access.
ISO/IEC 27001, ISO/IEC 42001, and GDPR: Best Practices for Implementation and...PECB
Denis is a dynamic and results-driven Chief Information Officer (CIO) with a distinguished career spanning information systems analysis and technical project management. With a proven track record of spearheading the design and delivery of cutting-edge Information Management solutions, he has consistently elevated business operations, streamlined reporting functions, and maximized process efficiency.
Certified as an ISO/IEC 27001: Information Security Management Systems (ISMS) Lead Implementer, Data Protection Officer, and Cyber Risks Analyst, Denis brings a heightened focus on data security, privacy, and cyber resilience to every endeavor.
His expertise extends across a diverse spectrum of reporting, database, and web development applications, underpinned by an exceptional grasp of data storage and virtualization technologies. His proficiency in application testing, database administration, and data cleansing ensures seamless execution of complex projects.
What sets Denis apart is his comprehensive understanding of Business and Systems Analysis technologies, honed through involvement in all phases of the Software Development Lifecycle (SDLC). From meticulous requirements gathering to precise analysis, innovative design, rigorous development, thorough testing, and successful implementation, he has consistently delivered exceptional results.
Throughout his career, he has taken on multifaceted roles, from leading technical project management teams to owning solutions that drive operational excellence. His conscientious and proactive approach is unwavering, whether he is working independently or collaboratively within a team. His ability to connect with colleagues on a personal level underscores his commitment to fostering a harmonious and productive workplace environment.
Date: May 29, 2024
Tags: Information Security, ISO/IEC 27001, ISO/IEC 42001, Artificial Intelligence, GDPR
-------------------------------------------------------------------------------
Find out more about ISO training and certification services
Training: ISO/IEC 27001 Information Security Management System - EN | PECB
ISO/IEC 42001 Artificial Intelligence Management System - EN | PECB
General Data Protection Regulation (GDPR) - Training Courses - EN | PECB
Webinars: https://pecb.com/webinars
Article: https://pecb.com/article
-------------------------------------------------------------------------------
For more information about PECB:
Website: https://pecb.com/
LinkedIn: https://www.linkedin.com/company/pecb/
Facebook: https://www.facebook.com/PECBInternational/
Slideshare: http://www.slideshare.net/PECBCERTIFICATION
This document provides an overview of wound healing, its functions, stages, mechanisms, factors affecting it, and complications.
A wound is a break in the integrity of the skin or tissues, which may be associated with disruption of the structure and function.
Healing is the body’s response to injury in an attempt to restore normal structure and functions.
Healing can occur in two ways: Regeneration and Repair
There are 4 phases of wound healing: hemostasis, inflammation, proliferation, and remodeling. This document also describes the mechanism of wound healing. Factors that affect healing include infection, uncontrolled diabetes, poor nutrition, age, anemia, the presence of foreign bodies, etc.
Complications of wound healing like infection, hyperpigmentation of scar, contractures, and keloid formation.
Temple of Asclepius in Thrace. Excavation resultsKrassimira Luka
The temple and the sanctuary around were dedicated to Asklepios Zmidrenus. This name has been known since 1875 when an inscription dedicated to him was discovered in Rome. The inscription is dated in 227 AD and was left by soldiers originating from the city of Philippopolis (modern Plovdiv).
Gender and Mental Health - Counselling and Family Therapy Applications and In...PsychoTech Services
A proprietary approach developed by bringing together the best of learning theories from Psychology, design principles from the world of visualization, and pedagogical methods from over a decade of training experience, that enables you to: Learn better, faster!
Beyond Degrees - Empowering the Workforce in the Context of Skills-First.pptxEduSkills OECD
Iván Bornacelly, Policy Analyst at the OECD Centre for Skills, OECD, presents at the webinar 'Tackling job market gaps with a skills-first approach' on 12 June 2024
Philippine Edukasyong Pantahanan at Pangkabuhayan (EPP) CurriculumMJDuyan
(𝐓𝐋𝐄 𝟏𝟎𝟎) (𝐋𝐞𝐬𝐬𝐨𝐧 𝟏)-𝐏𝐫𝐞𝐥𝐢𝐦𝐬
𝐃𝐢𝐬𝐜𝐮𝐬𝐬 𝐭𝐡𝐞 𝐄𝐏𝐏 𝐂𝐮𝐫𝐫𝐢𝐜𝐮𝐥𝐮𝐦 𝐢𝐧 𝐭𝐡𝐞 𝐏𝐡𝐢𝐥𝐢𝐩𝐩𝐢𝐧𝐞𝐬:
- Understand the goals and objectives of the Edukasyong Pantahanan at Pangkabuhayan (EPP) curriculum, recognizing its importance in fostering practical life skills and values among students. Students will also be able to identify the key components and subjects covered, such as agriculture, home economics, industrial arts, and information and communication technology.
𝐄𝐱𝐩𝐥𝐚𝐢𝐧 𝐭𝐡𝐞 𝐍𝐚𝐭𝐮𝐫𝐞 𝐚𝐧𝐝 𝐒𝐜𝐨𝐩𝐞 𝐨𝐟 𝐚𝐧 𝐄𝐧𝐭𝐫𝐞𝐩𝐫𝐞𝐧𝐞𝐮𝐫:
-Define entrepreneurship, distinguishing it from general business activities by emphasizing its focus on innovation, risk-taking, and value creation. Students will describe the characteristics and traits of successful entrepreneurs, including their roles and responsibilities, and discuss the broader economic and social impacts of entrepreneurial activities on both local and global scales.
Harvesting HathiTrust Documents: A New Model for Online Access
1. Harvesting HathiTrust Documents: A New
Model for Online Access
Christopher C. Brown
University of Denver, Penrose Library
(303) 871-3404
cbrown@du.edu
2011 Missouri Government Documents Conference
2. DR, IR,
Digital Texts
Inbound Harvesting
Outbound Harvesting
This presentation will show how Encore harvesting can be
used to mitigate a space problem in a library, substituting
online access for the need for physical access to the
collection. The government documents collection will be the
primary focus.
Note: Encore is the next-generation catalog interface produced by Innovative Interfaces, Inc.
3. Collection Downsizing?
Malpas, Constance. 2011. Cloud-sourcing Research Collections: Managing Print in the Mass-digitized
Library Environment. Dublin, Ohio: OCLC Research.
http://www.oclc.org/research/publications/library/2011/2011-01.pdf.
4. About University of Denver
Depository since 1909
Historically a 70-75% selective
Now a 4.8% selective, but receive 100%
of online cataloging
Adding URLs to historic documents
Currently 100% of our paper documents
are in storage
We are remodeling our library. Under the
remodeling plan, all docs will remain in
remote storage.
5. Partial Solution: Using Encore for
Outbound Harvesting
All documents off-site
Our users are accustomed to using electronic documents
Need to divert attention away from physical collection
holdings
Encore harvesting of Hathi Trust can do this
6. PD = where docs generally live
Hathi Trust Attributes
From: http://www.hathitrust.org/rights_database
7. Sampling Method
I wanted to see how many government documents were in
our Hathi Trust harvest
Limit to Hathi Trust for a given year
Examine first result on each page of 25 results (4% of
results) [limitation: Encore only displays first 1,000 results]
8. Harvesting Hathi Docs: The Stats
Date Range
2000-2009
1990-1999
1980-1989
1970-1979
1960-1969
1950-1959
1940-1949
1930-1939
1920-1929
1910-1919
1900-1909
1890-1899
1880-1889
1870-1879
1860-1869
Hathi Totals
505,682
709,214
723,657
631,110
546,914
281,615
184,755
175,103
175,226
175,148
179,018
112,295
83,950
58,624
50,907
4,593,218
Hathi All Pub
Domain
pdus + pd
14,140
29,163
33,753
28,633
21,244
20,861
17,096
16,237
66,563
169,923
153,284
110,605
82,809
57,826
50,337
872,474
Hathi pdus DU pd Harvest
726
13,369
880
28,164
1,204
32,321
2,046
26,189
1,987
18,991
863
19,893
600
16,253
654
15,317
27,108
28,854
75,955
61,230
70,900
47,999
50,502
34,742
38,928
23,855
27,202
17,751
2,273
45,790
301,828
430,718
Docs Sampling
13,340
99.78%
26,662
94.67%
31,370
97.06%
25,607
97.78%
7,668
40.38%
3,888
19.54%
3,771
23.21%
2,600
16.97%
1,529
5.30%
4,124
6.73%
2,265
4.72%
596
1.72%
699
2.93%
319
1.80%
248
0.54%
124,686
28.95%
Statistics as of mid-March, 2011
The Docs Sampling columns show the estimated numbers of docs per
year and the estimated percentage of docs per year from the Harvest
9. Malpas: Docs about 3% of Hathi Total
and 15% of Public Domain
GovDocs: 3% overall
GovDocs: 15% of Public Domain
10. Hathi Docs Usage in Proportion to Docs
Distribution
Sources: 1895-1976 data: Monthly Catalog, 1895-1976 (ProQuest);1976 onward data: CGP
13. Stripped-Out Fields
008 fixed field data
650 subfields other than “a”
500 notes
5xx shipping list info
300 subfields after “a”
086 SuDocs number
14. Use Stats for Hathi Trust?
Statistics from Google Analytics
•Statistics for all Hathi Trust records accessed, not just documents
•Spikes in usage are docs librarian (my) testing, not real users
17. Conclusions
Documents content in HathiTrust can provide a
suitable surrogate for a limited subset of documents,
but not a wholesale replacement.
HathiTrust documents can be used as surrogates for
selected titles, especially larger serial runs. But it is
difficult at this time to isolate those titles.
HathiTrust is definitely worth harvesting into local
catalogs or other digital repositories.