Metadata plays a key role in describing and enhancing archival collections. This presentation covers the standards, metadata elements and tools chosen to describe two unique archival collections at Solent University Library. Using the International Standard Archival Description (General) (ISAD (G)), as a framework for the creation and management of archival descriptions, EAD as a serialisation format and the Jisc’s Archives Hub Editor as an application, the archival works of Philip Mackie and Ken Russell were described. A decision was made to comply with ISAD(G)’s mandatory elements such as Reference code, Title, Name of Creator, Dates of Creation, Extent of the Unit of Description and Level of description. Other metadata elements were also identified to record context, provenance, and access points. Because of this work, the library saw noticeable increase in the interest of its archival collections both within and outside the university. This presentation includes a step-by-step explanation of how the Archives Hub Editor was used to describe the works of Philip Mackie and the works of Ken Russell which includes scripts, photographs, scripts, correspondence, and other documents.
Paper presented at the CILIP Metadata and Discovery Group (MDG) Conference & UKCoR RDA Day (6th - 8th Sept 2023 at IET Austin Court, Birmingham).
The NBK and the UK Distributed Print Book Collection / Rozz Evans (University...CILIP MDG
The NBK and the UK Distributed Print Book Collection / Rozz Evans (Head of Collection Strategy, University College London) and Bethan Ruddock (Senior Product Owner, Digital Resources, Jisc).
In May 2022 the Research Libraries UK (RLUK) Board of Directors endorsed a vision for a UK Distributed Print Book Collection (UK DPBC) as part of the RLUK Library Transforming Strategy of which ‘Collective Collections’ is one of the five key strands. It is envisaged that this shared print collection will extend beyond RLUK libraries, with SCONUL, national libraries, special libraries and Jisc all playing a role. This model will help to ensure preservation of and access to shared print holdings in the UK for current and future users.
A key part of the infrastructure for this work is the National Bibliographic Knowledgebase (NBK). Initial scoping work has been done using the holdings information on the NBK to provide an overview of how many print monograph titles in total are held in NBK libraries and numbers of holding libraries (where there are 10 or fewer records in the NBK). The RLUK Collections Strategy Network (CSN) is developing guidelines for libraries who wish to participate in the UK Distributed Print Book Collection (UK DPBC) and is using the NBK data analysis to help identify the minimum number of copies which should be held nationally. The intention is that participating libraries will add retention statements to the catalogue records of any books they can commit to keep in their collections, and that these retention commitments will be surfaced in the NBK.
This presentation will cover the background and aims of the UK Distributed Print Book Collection strategy; how the NBK data has been analysed so far; the role of high-quality metadata in enabling this analysis; and the future stages of the project.
Paper presented at the CILIP Metadata and Discovery Group (MDG) Conference & UKCoR RDA Day (6th - 8th Sept 2023 at IET Austin Court, Birmingham).
Yes we can! Implementing digitization requests in AlmaFrançois Renaville
Enabling patron digitization requests is one of those (new) exciting services that end users appreciate! Alma allows patrons and staff members to request the creation of a digital file for any physical or electronic item. Alma digitization request workflow is configurable according to the library’s requirements, created via Alma or Primo, partial or full, limited to some user groups, limited in numbers, with or without approval process, digitized materials added to collections and made accessible to registered and/or guest users via Primo, etc. So Alma libraries can deal with a lot of different options to satisfy their patrons' needs.
This session will present the development and launch of digital fulfillment from the perspective of 3 institutions (the University of Otago Library, the University of Liege Library, and the University of Sheffield Library) focusing on the strategic choices that were made, the implementation, and the final delivery of the services to their communities.
This document introduces linked data and discusses how publishing data as linked RDF triples on the web allows for a global linked database. It explains that linked data uses HTTP URIs to identify things and links data from different sources to be queried using SPARQL. Publishing linked data provides benefits like being able to integrate and discover related data on the web. Tools are available to convert existing data or publish new data as linked open data.
How to use the Cataloguing Code Ethics at your day job : a hands-on workshop ...CILIP MDG
In 2021, the joint USA, Canada and UK Cataloguing Ethics Steering Committee published the Cataloguing Code of Ethics. The document is a community response to clearly articulated needs for a code of ethics addressing the specific responsibilities of cataloguers and metadata managers engaged in creating, sharing, enriching, and preserving metadata in an increasingly global and technology dominated cataloguing ecosystem.
The Code was created using the work of volunteers drawn from the cataloguing community of practice in the USA, Canada, and the UK but also from Australia, Mexico, and Israel. It is a concise practical tool for reviewing cataloguing policy and practice through an ethical lens and is intended for use by practitioners (regardless of where they are on their career journey and whether they work for the cultural and heritage sector or for the companies that sell systems, content, or metadata); standards developers; students; and educators.
The ten statements of ethical principles embody a commitment to: understanding and meeting user needs; ensuring access to resources and metadata; promoting collaborative partnerships; critically applying standards; acknowledging bias; respecting the privacy and preferences of those who create and publish information resources; advocating for the value of cataloguing work and access to appropriate education and training to build a robust and sustainable workforce; promoting diversity, equity and inclusion; taking responsibility for decision-making and ensuring transparency in working practices. It is intended that the Code will provide agency and empowerment for cataloguers and metadata managers, both now and in the future, as they seek to deliver solutions to ethical dilemmas that are part and parcel of day-to-day cataloguing work.
The session will include a short introduction to the Code followed by group work that will allow attendees to learn how to use the Code to devise practical solutions. Each of the ten statements will be examined under the broader topic headings of; audiences, tasks, and hurdles; collaboration; and education, jobs and advocacy. The presenters will provide practical examples of ethical issues applying to each principle and invite attendees to contribute their own experiences and questions. Feedback will be gathered and shared for mutual benefit.
Paper presented at the CILIP Metadata and Discovery Group (MDG) Conference & UKCoR RDA Day (6th - 8th Sept 2023 at IET Austin Court, Birmingham).
Dal riconoscimento facciale al riconoscimento di frodi o difetti di fabbricazione, l'analisi di immagini e video che sfruttano tecniche di intelligenza artificiale, si stanno evolvendo e raffinando a ritmi elevati. In questo webinar esploreremo le possibilità messe a disposizione dai servizi AWS per applicare lo stato dell'arte delle tecniche di computer vision a scenari reali.
Building, Evaluating, and Optimizing your RAG App for ProductionSri Ambati
The document discusses optimizing question answering systems called RAG (Retrieve-and-Generate) stacks. It outlines challenges with naive RAG approaches and proposes solutions like improved data representations, advanced retrieval techniques, and fine-tuning large language models. Table stakes optimizations include tuning chunk sizes, prompt engineering, and customizing LLMs. More advanced techniques involve small-to-big retrieval, multi-document agents, embedding fine-tuning, and LLM fine-tuning.
BrightonSEO October 2022 - Martijn Scheybeler - SEO Testing: Find Out What Wo...Martijn Scheijbeler
This document discusses SEO testing and experimentation. It provides examples of types of SEO tests that could be done, such as pre-post testing, A/B testing, and combining different measurement approaches. Specific cases are also mentioned, like testing the impact of internal linking, HTML sitemaps, and changes to page titles and meta descriptions. The document emphasizes that SEO testing takes work to set up properly and find meaningful results, but it is important for identifying what strategies are truly effective.
The NBK and the UK Distributed Print Book Collection / Rozz Evans (University...CILIP MDG
The NBK and the UK Distributed Print Book Collection / Rozz Evans (Head of Collection Strategy, University College London) and Bethan Ruddock (Senior Product Owner, Digital Resources, Jisc).
In May 2022 the Research Libraries UK (RLUK) Board of Directors endorsed a vision for a UK Distributed Print Book Collection (UK DPBC) as part of the RLUK Library Transforming Strategy of which ‘Collective Collections’ is one of the five key strands. It is envisaged that this shared print collection will extend beyond RLUK libraries, with SCONUL, national libraries, special libraries and Jisc all playing a role. This model will help to ensure preservation of and access to shared print holdings in the UK for current and future users.
A key part of the infrastructure for this work is the National Bibliographic Knowledgebase (NBK). Initial scoping work has been done using the holdings information on the NBK to provide an overview of how many print monograph titles in total are held in NBK libraries and numbers of holding libraries (where there are 10 or fewer records in the NBK). The RLUK Collections Strategy Network (CSN) is developing guidelines for libraries who wish to participate in the UK Distributed Print Book Collection (UK DPBC) and is using the NBK data analysis to help identify the minimum number of copies which should be held nationally. The intention is that participating libraries will add retention statements to the catalogue records of any books they can commit to keep in their collections, and that these retention commitments will be surfaced in the NBK.
This presentation will cover the background and aims of the UK Distributed Print Book Collection strategy; how the NBK data has been analysed so far; the role of high-quality metadata in enabling this analysis; and the future stages of the project.
Paper presented at the CILIP Metadata and Discovery Group (MDG) Conference & UKCoR RDA Day (6th - 8th Sept 2023 at IET Austin Court, Birmingham).
Yes we can! Implementing digitization requests in AlmaFrançois Renaville
Enabling patron digitization requests is one of those (new) exciting services that end users appreciate! Alma allows patrons and staff members to request the creation of a digital file for any physical or electronic item. Alma digitization request workflow is configurable according to the library’s requirements, created via Alma or Primo, partial or full, limited to some user groups, limited in numbers, with or without approval process, digitized materials added to collections and made accessible to registered and/or guest users via Primo, etc. So Alma libraries can deal with a lot of different options to satisfy their patrons' needs.
This session will present the development and launch of digital fulfillment from the perspective of 3 institutions (the University of Otago Library, the University of Liege Library, and the University of Sheffield Library) focusing on the strategic choices that were made, the implementation, and the final delivery of the services to their communities.
This document introduces linked data and discusses how publishing data as linked RDF triples on the web allows for a global linked database. It explains that linked data uses HTTP URIs to identify things and links data from different sources to be queried using SPARQL. Publishing linked data provides benefits like being able to integrate and discover related data on the web. Tools are available to convert existing data or publish new data as linked open data.
How to use the Cataloguing Code Ethics at your day job : a hands-on workshop ...CILIP MDG
In 2021, the joint USA, Canada and UK Cataloguing Ethics Steering Committee published the Cataloguing Code of Ethics. The document is a community response to clearly articulated needs for a code of ethics addressing the specific responsibilities of cataloguers and metadata managers engaged in creating, sharing, enriching, and preserving metadata in an increasingly global and technology dominated cataloguing ecosystem.
The Code was created using the work of volunteers drawn from the cataloguing community of practice in the USA, Canada, and the UK but also from Australia, Mexico, and Israel. It is a concise practical tool for reviewing cataloguing policy and practice through an ethical lens and is intended for use by practitioners (regardless of where they are on their career journey and whether they work for the cultural and heritage sector or for the companies that sell systems, content, or metadata); standards developers; students; and educators.
The ten statements of ethical principles embody a commitment to: understanding and meeting user needs; ensuring access to resources and metadata; promoting collaborative partnerships; critically applying standards; acknowledging bias; respecting the privacy and preferences of those who create and publish information resources; advocating for the value of cataloguing work and access to appropriate education and training to build a robust and sustainable workforce; promoting diversity, equity and inclusion; taking responsibility for decision-making and ensuring transparency in working practices. It is intended that the Code will provide agency and empowerment for cataloguers and metadata managers, both now and in the future, as they seek to deliver solutions to ethical dilemmas that are part and parcel of day-to-day cataloguing work.
The session will include a short introduction to the Code followed by group work that will allow attendees to learn how to use the Code to devise practical solutions. Each of the ten statements will be examined under the broader topic headings of; audiences, tasks, and hurdles; collaboration; and education, jobs and advocacy. The presenters will provide practical examples of ethical issues applying to each principle and invite attendees to contribute their own experiences and questions. Feedback will be gathered and shared for mutual benefit.
Paper presented at the CILIP Metadata and Discovery Group (MDG) Conference & UKCoR RDA Day (6th - 8th Sept 2023 at IET Austin Court, Birmingham).
Dal riconoscimento facciale al riconoscimento di frodi o difetti di fabbricazione, l'analisi di immagini e video che sfruttano tecniche di intelligenza artificiale, si stanno evolvendo e raffinando a ritmi elevati. In questo webinar esploreremo le possibilità messe a disposizione dai servizi AWS per applicare lo stato dell'arte delle tecniche di computer vision a scenari reali.
Building, Evaluating, and Optimizing your RAG App for ProductionSri Ambati
The document discusses optimizing question answering systems called RAG (Retrieve-and-Generate) stacks. It outlines challenges with naive RAG approaches and proposes solutions like improved data representations, advanced retrieval techniques, and fine-tuning large language models. Table stakes optimizations include tuning chunk sizes, prompt engineering, and customizing LLMs. More advanced techniques involve small-to-big retrieval, multi-document agents, embedding fine-tuning, and LLM fine-tuning.
BrightonSEO October 2022 - Martijn Scheybeler - SEO Testing: Find Out What Wo...Martijn Scheijbeler
This document discusses SEO testing and experimentation. It provides examples of types of SEO tests that could be done, such as pre-post testing, A/B testing, and combining different measurement approaches. Specific cases are also mentioned, like testing the impact of internal linking, HTML sitemaps, and changes to page titles and meta descriptions. The document emphasizes that SEO testing takes work to set up properly and find meaningful results, but it is important for identifying what strategies are truly effective.
The document discusses different types of taxonomies and controlled vocabularies including their definitions, purposes, and benefits. It describes term lists, synonym rings, authority files, hierarchical taxonomies, thesauri, and ontologies. The key purposes are to help people find information using different terms, retrieve relevant concepts rather than just words, and organize information into logical hierarchies or relationships to aid searching and browsing. Solo information professionals are common accidental taxonomists.
This document discusses building and using ontologies. It defines an ontology as defining a domain of interest in terms of things, attributes, and relationships. Ontologies are used to share a common understanding of a domain among people and machines. The document then discusses ontology engineering processes, examples of ontologies like DBpedia, and semantic technologies used to create intelligent applications.
Data integration, data interoperation and data quality are major challenges that continue to haunt enterprises. Every enterprise either by choice or by chance has created massive silos of data in different formats, with duplications and quality issues.
Knowledge graphs have proven to be a viable solution to address the integration and interoperation problem. Semantic technologies in particular provide an intelligent way of creating an abstract layer for the enterprise data model and mapping of siloed data to that model, allowing a smooth integration and a common view of the data.
Technologies like OWL (Web Ontology Language) and RDF (Resource Description Framework) are the back bone of semantics for knowledge graph implementation. Enterprises use OWL to build an ontology model to create a common definition for concepts and how they are connected to each other in their specific domain.
They then use RDF to create a triple format representation of their data by mapping it to the Ontology. This approach makes their data smart and machine understandable.
But how can enterprises control and validate the quality of this mapped data? Furthermore, how can they use this one abstract representation of data to meet all their different business requirements? Different departments, different LoBs and different business branches all have their own data needs, creating a new challenge to be tackled by the enterprise.
In this talk we will look at how the power of SHACL (SHAPES and Constraints Language), a W3C standard for defining constraint sets over data; complements the two core semantic technologies OWL and RDF. What are the similarities, the overlaps and the differences.
We will talk about how SHACL gives enterprises the power to reuse, customize and validate their data for various scenarios, uses cases and business requirements; making the application of semantics even more practical.
AI Writing Bots VS. Human Writers: Will AI replace Writers?Jomer Gregorio
Are AI writing bots the future of content creation? Dive into the advantages and limitations of AI and human writers in our latest article. Click now to discover and optimize now for your business!
Full blog here - https://digitalmarketingphilippines.com/ai-writing-bots-vs-human-writers-will-ai-replace-writers/
Lecture 6: Infrastructure & Tooling (Full Stack Deep Learning - Spring 2021)Sergey Karayev
The document discusses infrastructure and tooling for full stack deep learning. It provides an overview of the different components involved, including compute, data processing, experiment management, deployment, and software engineering practices. Specifically, it covers topics like GPU basics, cloud computing options, development versus training needs, popular programming languages and editors like Python and Jupyter Notebooks, and setting up development environments.
Presentation given at Macquarie University in support of the ARDC 'institutional role in the data commons' project on "Implementing FAIR: Standards in Research Data Management" https://ardc.edu.au/news/data-and-services-discovery-activities-successful-applicants/
This document summarizes key findings from surveys about researchers' data sharing practices and attitudes. It finds that while most researchers agree data should be shared, only a small percentage actually make their data openly available. Researchers typically share data through email, cloud services, or external drives rather than repositories. The document also discusses increasing emphasis on open and FAIR data in research funder policies, but notes researchers face barriers to compliance like unclear terminology, lack of skills and incentives, and confusion between open data and managed/FAIR data. It argues for engagement programs to help researchers better understand and participate in open scholarship.
This presentation contains differences between Elasticsearch and relational Databases. Along with that it also has some Glossary Of Elasticsearch and its basic operation.
Presentation given on March 12, 2013 by Marjorie M.K. Hlava of Access Innovations, Inc. as a webinar for the San Francisco chapter of the Special Libraries Association.
Understanding Semantic Search and AI Content to Drive Growth in 2023 March 2023TysonStockton1
Exploring modern search engines, semantic search, and AI technology to better understand how we can integrate into SEO strategy and content initiatives.
With the rise of ChatGPT there has been a lot of discussion around if SEO content is good or bad. To best determine how to leverage this technology in SEO workflows we must revisit how a modern search engine works and where we are at with AI technology.
Knowledge Graphs and AI to Hyper-Personalise the Fashion Retail Experience at...Connected Data World
What is the key to the holistic success of the fastest growing and most successful companies of our time globally? Well, often the key is the rapid increase in collected and analysed data. Graph databases provide a way to organise semantically by classes, not tables, are web-aware, and are superior for handling deep, complex relationships than traditional relational or NoSQL data stores.
It is these deep, complex relationships that can provide the rich context for hyper-personalising your product offering, inspiring consumers to purchase. In this talk, we describe how we are using artificial intelligence at Farfetch to not only help build a knowledge graph but also to evolve our insights with state-of-the-art graph-based AI.
The document discusses data quality success stories and provides an overview of a program on the topic. It introduces the program, which will discuss data quality as an engineering challenge, putting a price on data quality, how components of data management complement each other, savings-based and innovation-based success stories, and non-monetary success stories. The program aims to provide takeaways and allow for questions and answers.
BrightonSEO - ChatGPT-4 Localisation friend or foe FINAL.pptxJamesBall92
This document discusses the use of ChatGPT-4 for localization and translation purposes. Editors tested ChatGPT-4's ability to translate text into Spanish, French, and Chinese and found that it produced readable Spanish and French translations but struggled more with Chinese. The AI translations lacked nuance and emotion compared to human translations. A survey of 2,500 people in various markets found that around 60% felt positively about AI writing most content, but fewer had confidence in AI for long-form or important content. While AI could be useful for some applications like proofreading, the document concludes that overreliance on AI could hinder authoritative, trusted content and that human experience and expertise cannot be replaced.
Using Tags & Taxonomies to super charge your eCommerce SEOMichael King
Using tags and taxonomies can supercharge ecommerce SEO. Properly labeling products and categorizing them allows for (1) better targeting of long-tail keywords, (2) improved internal linking to distribute PageRank, and (3) helping Google's crawlers discover content more efficiently. Key recommendations include designing three-level categorization for products and using tags to link diverse products. Automating recommendations and dynamic linking based on tags and attributes can further boost performance.
This document provides an overview of Sojern's partnership and advertising services. Sojern uses traveler data to target relevant ads to in-market travelers, helping brands engage travelers efficiently. Sojern offers data monetization by earning revenue from anonymous site visitor data, data-driven marketing to target ads and increase conversions, and audience insights from its extensive data. Major travel brands partner with Sojern for incremental revenue by monetizing their data with no cost, and access audience insights to improve their own marketing. The document includes examples of Sojern's services and case studies demonstrating successful campaigns for travel, hospitality and retail brands.
Building a Data Lake for Your Enterprise, ft. Sysco (STG309) - AWS re:Invent ...Amazon Web Services
Data lakes are transforming the way enterprises store, analyze, and learn insights from their data. While data lakes are a relatively new concept, many enterprises have already generated significant business value from the insights gleaned. In this session, AWS experts and technology leaders from Sysco, a Fortune 50 company and leader in food distribution and marketing, explain why Sysco decided to evolve its data management capabilities to include data lakes and how they customized them to support diverse querying capabilities and data science use cases. They also discuss how to architect different aspects of a data lake—ingestion from disparate sources, data consumption, and usability layers—and how to track data ingestion and consumption, monitor associated costs, enforce wanted levels of user access, manage data file formats, synchronize production and non-production environments, and maintain data integrity. Services to be discussed include Amazon S3 and S3 Select, Amazon Athena, Amazon EMR, Amazon EC2, and Amazon Redshift Spectrum.
Lect6-An introduction to ontologies and ontology developmentAntonio Moreno
The document provides an overview of ontologies and ontology development:
1. It defines ontologies as explicit specifications of conceptualizations in a domain that define concepts, properties, attributes, and relationships to enable knowledge sharing.
2. Ontology components include concepts, properties, restrictions, and individuals. Ontologies can range from single large ontologies to several specialized smaller ones.
3. OWL is introduced as the standard language for representing ontologies, with features like classes, properties, restrictions, and logical operators.
4. A general methodology for ontology development is outlined, including determining scope, reusing existing ontologies, enumerating terms, and defining classes, properties, and other components in an iterative
GPT-4 is the newest version of OpenAI's language model that can understand and generate natural language. It shows improvements over GPT-3.5 in its ability to take visual inputs, be steered more precisely by the user, refuse unsafe requests, and score higher on factual benchmarks. Potential applications of GPT-4 include customer service, translation, content creation, and research. However, its adoption may displace some jobs and raises ethical issues that need addressing through education, job retraining, and responsible development of the technology.
Reference Model for an Open Archival Information Systems (OAIS): Overview and...faflrt
ALA/FAFLRT Workshop on Open Archival Information Service (OAIS). Presented by Alan Wood/A.E.Wood & Erickson/Lockheed Martin, Don Sawyer/NASA/GSFC, and Lou Reich/CSC. Sponsored by ALA Federal and Armed Forces Libraries Roundtable (FAFLRT). Presented on June 16, 2001 at the ALA Annual Conference.
OAIS and It's Applicability for Libraries, Archives, and Digital Repositories...faflrt
ALA/FAFLRT Workshop on Open Archival Information Service (OAIS). Presented by Robin Dale, RLG. Sponsored by ALA Federal and Armed Forces Libraries Roundtable (FAFLRT). Presented on June 16, 2001 at the ALA Annual Conference.
The document discusses different types of taxonomies and controlled vocabularies including their definitions, purposes, and benefits. It describes term lists, synonym rings, authority files, hierarchical taxonomies, thesauri, and ontologies. The key purposes are to help people find information using different terms, retrieve relevant concepts rather than just words, and organize information into logical hierarchies or relationships to aid searching and browsing. Solo information professionals are common accidental taxonomists.
This document discusses building and using ontologies. It defines an ontology as defining a domain of interest in terms of things, attributes, and relationships. Ontologies are used to share a common understanding of a domain among people and machines. The document then discusses ontology engineering processes, examples of ontologies like DBpedia, and semantic technologies used to create intelligent applications.
Data integration, data interoperation and data quality are major challenges that continue to haunt enterprises. Every enterprise either by choice or by chance has created massive silos of data in different formats, with duplications and quality issues.
Knowledge graphs have proven to be a viable solution to address the integration and interoperation problem. Semantic technologies in particular provide an intelligent way of creating an abstract layer for the enterprise data model and mapping of siloed data to that model, allowing a smooth integration and a common view of the data.
Technologies like OWL (Web Ontology Language) and RDF (Resource Description Framework) are the back bone of semantics for knowledge graph implementation. Enterprises use OWL to build an ontology model to create a common definition for concepts and how they are connected to each other in their specific domain.
They then use RDF to create a triple format representation of their data by mapping it to the Ontology. This approach makes their data smart and machine understandable.
But how can enterprises control and validate the quality of this mapped data? Furthermore, how can they use this one abstract representation of data to meet all their different business requirements? Different departments, different LoBs and different business branches all have their own data needs, creating a new challenge to be tackled by the enterprise.
In this talk we will look at how the power of SHACL (SHAPES and Constraints Language), a W3C standard for defining constraint sets over data; complements the two core semantic technologies OWL and RDF. What are the similarities, the overlaps and the differences.
We will talk about how SHACL gives enterprises the power to reuse, customize and validate their data for various scenarios, uses cases and business requirements; making the application of semantics even more practical.
AI Writing Bots VS. Human Writers: Will AI replace Writers?Jomer Gregorio
Are AI writing bots the future of content creation? Dive into the advantages and limitations of AI and human writers in our latest article. Click now to discover and optimize now for your business!
Full blog here - https://digitalmarketingphilippines.com/ai-writing-bots-vs-human-writers-will-ai-replace-writers/
Lecture 6: Infrastructure & Tooling (Full Stack Deep Learning - Spring 2021)Sergey Karayev
The document discusses infrastructure and tooling for full stack deep learning. It provides an overview of the different components involved, including compute, data processing, experiment management, deployment, and software engineering practices. Specifically, it covers topics like GPU basics, cloud computing options, development versus training needs, popular programming languages and editors like Python and Jupyter Notebooks, and setting up development environments.
Presentation given at Macquarie University in support of the ARDC 'institutional role in the data commons' project on "Implementing FAIR: Standards in Research Data Management" https://ardc.edu.au/news/data-and-services-discovery-activities-successful-applicants/
This document summarizes key findings from surveys about researchers' data sharing practices and attitudes. It finds that while most researchers agree data should be shared, only a small percentage actually make their data openly available. Researchers typically share data through email, cloud services, or external drives rather than repositories. The document also discusses increasing emphasis on open and FAIR data in research funder policies, but notes researchers face barriers to compliance like unclear terminology, lack of skills and incentives, and confusion between open data and managed/FAIR data. It argues for engagement programs to help researchers better understand and participate in open scholarship.
This presentation contains differences between Elasticsearch and relational Databases. Along with that it also has some Glossary Of Elasticsearch and its basic operation.
Presentation given on March 12, 2013 by Marjorie M.K. Hlava of Access Innovations, Inc. as a webinar for the San Francisco chapter of the Special Libraries Association.
Understanding Semantic Search and AI Content to Drive Growth in 2023 March 2023TysonStockton1
Exploring modern search engines, semantic search, and AI technology to better understand how we can integrate into SEO strategy and content initiatives.
With the rise of ChatGPT there has been a lot of discussion around if SEO content is good or bad. To best determine how to leverage this technology in SEO workflows we must revisit how a modern search engine works and where we are at with AI technology.
Knowledge Graphs and AI to Hyper-Personalise the Fashion Retail Experience at...Connected Data World
What is the key to the holistic success of the fastest growing and most successful companies of our time globally? Well, often the key is the rapid increase in collected and analysed data. Graph databases provide a way to organise semantically by classes, not tables, are web-aware, and are superior for handling deep, complex relationships than traditional relational or NoSQL data stores.
It is these deep, complex relationships that can provide the rich context for hyper-personalising your product offering, inspiring consumers to purchase. In this talk, we describe how we are using artificial intelligence at Farfetch to not only help build a knowledge graph but also to evolve our insights with state-of-the-art graph-based AI.
The document discusses data quality success stories and provides an overview of a program on the topic. It introduces the program, which will discuss data quality as an engineering challenge, putting a price on data quality, how components of data management complement each other, savings-based and innovation-based success stories, and non-monetary success stories. The program aims to provide takeaways and allow for questions and answers.
BrightonSEO - ChatGPT-4 Localisation friend or foe FINAL.pptxJamesBall92
This document discusses the use of ChatGPT-4 for localization and translation purposes. Editors tested ChatGPT-4's ability to translate text into Spanish, French, and Chinese and found that it produced readable Spanish and French translations but struggled more with Chinese. The AI translations lacked nuance and emotion compared to human translations. A survey of 2,500 people in various markets found that around 60% felt positively about AI writing most content, but fewer had confidence in AI for long-form or important content. While AI could be useful for some applications like proofreading, the document concludes that overreliance on AI could hinder authoritative, trusted content and that human experience and expertise cannot be replaced.
Using Tags & Taxonomies to super charge your eCommerce SEOMichael King
Using tags and taxonomies can supercharge ecommerce SEO. Properly labeling products and categorizing them allows for (1) better targeting of long-tail keywords, (2) improved internal linking to distribute PageRank, and (3) helping Google's crawlers discover content more efficiently. Key recommendations include designing three-level categorization for products and using tags to link diverse products. Automating recommendations and dynamic linking based on tags and attributes can further boost performance.
This document provides an overview of Sojern's partnership and advertising services. Sojern uses traveler data to target relevant ads to in-market travelers, helping brands engage travelers efficiently. Sojern offers data monetization by earning revenue from anonymous site visitor data, data-driven marketing to target ads and increase conversions, and audience insights from its extensive data. Major travel brands partner with Sojern for incremental revenue by monetizing their data with no cost, and access audience insights to improve their own marketing. The document includes examples of Sojern's services and case studies demonstrating successful campaigns for travel, hospitality and retail brands.
Building a Data Lake for Your Enterprise, ft. Sysco (STG309) - AWS re:Invent ...Amazon Web Services
Data lakes are transforming the way enterprises store, analyze, and learn insights from their data. While data lakes are a relatively new concept, many enterprises have already generated significant business value from the insights gleaned. In this session, AWS experts and technology leaders from Sysco, a Fortune 50 company and leader in food distribution and marketing, explain why Sysco decided to evolve its data management capabilities to include data lakes and how they customized them to support diverse querying capabilities and data science use cases. They also discuss how to architect different aspects of a data lake—ingestion from disparate sources, data consumption, and usability layers—and how to track data ingestion and consumption, monitor associated costs, enforce wanted levels of user access, manage data file formats, synchronize production and non-production environments, and maintain data integrity. Services to be discussed include Amazon S3 and S3 Select, Amazon Athena, Amazon EMR, Amazon EC2, and Amazon Redshift Spectrum.
Lect6-An introduction to ontologies and ontology developmentAntonio Moreno
The document provides an overview of ontologies and ontology development:
1. It defines ontologies as explicit specifications of conceptualizations in a domain that define concepts, properties, attributes, and relationships to enable knowledge sharing.
2. Ontology components include concepts, properties, restrictions, and individuals. Ontologies can range from single large ontologies to several specialized smaller ones.
3. OWL is introduced as the standard language for representing ontologies, with features like classes, properties, restrictions, and logical operators.
4. A general methodology for ontology development is outlined, including determining scope, reusing existing ontologies, enumerating terms, and defining classes, properties, and other components in an iterative
GPT-4 is the newest version of OpenAI's language model that can understand and generate natural language. It shows improvements over GPT-3.5 in its ability to take visual inputs, be steered more precisely by the user, refuse unsafe requests, and score higher on factual benchmarks. Potential applications of GPT-4 include customer service, translation, content creation, and research. However, its adoption may displace some jobs and raises ethical issues that need addressing through education, job retraining, and responsible development of the technology.
Reference Model for an Open Archival Information Systems (OAIS): Overview and...faflrt
ALA/FAFLRT Workshop on Open Archival Information Service (OAIS). Presented by Alan Wood/A.E.Wood & Erickson/Lockheed Martin, Don Sawyer/NASA/GSFC, and Lou Reich/CSC. Sponsored by ALA Federal and Armed Forces Libraries Roundtable (FAFLRT). Presented on June 16, 2001 at the ALA Annual Conference.
OAIS and It's Applicability for Libraries, Archives, and Digital Repositories...faflrt
ALA/FAFLRT Workshop on Open Archival Information Service (OAIS). Presented by Robin Dale, RLG. Sponsored by ALA Federal and Armed Forces Libraries Roundtable (FAFLRT). Presented on June 16, 2001 at the ALA Annual Conference.
This document provides an overview of metadata standards and sources of information about standards. It describes what metadata standards are and their purposes. Sources discussed provide basic information on metadata in general and information on specific standards. The document also describes the Encoded Archival Description (EAD) standard in detail, including its benefits and sources of information about EAD. Finally, it discusses cross-walking between standards and upcoming changes to the EAD standard.
This document discusses standards related to archival description, including EAD, DACS, and MARC. It provides an overview of each standard and their purposes. EAD is an XML standard for encoding finding aids to display them online. DACS is a content standard that does not prescribe structure, leaving that to EAD. MARC was originally created for libraries but has been adapted for archival use through standards like APPM and ACM to represent archival materials and collections.
OAIS: What is it and Where is it Going? - Don Sawyer (2002)faflrt
Open Archival Information Service (OAIS) workshop. Presented by Don Sawyer of NASA Goddard and Lou Reich, CSC contractor to NASA. Sponsored by ALA Federal and Armed Forces Libraries Roundtable (FAFLRT). Presented on June 15, 2002 at ALA Annual Conference.
Alphabet soup: CDM, VRA, CCO, METS, MODS, RDF - Why Metadata MattersNew York University
This presentation given to University of Iowa Libraries on Nov. 17, 2014, discussing 1) the alphabet soup of metadata standards, e.g. CDM, VRA, CCO, METS, MODS, RDF, including sample tagging and their applications for digital libraries, and 2) why metadata matters. It does not address metadata issues and tools for metadata creation, extraction, transformation, quality control, syndication and ingest.
A presentation given by Manjula Patel (UKOLN) at the Repository Curation Environments (RECURSE) Workshop held at the 4th International Digital Curation Conference, Edinburgh, 1st December 2008,
http://www.dcc.ac.uk/events/dcc-2008/programme/
Innovative methods for data integration: Linked Data and NLPariadnenetwork
Linked Data (LD) + Natural Language Processing (NLP)
Two technologies that open up new possibilities for semantic integration of archaeological datasets and fieldwork reports.
Overview
•Illustrative early examples
- a flavour of progress and challenges to date
•NLP of grey literature (English – Dutch)
•Mapping between multilingual vocabularies
An introduction to the Joint Information Systems Committee Resource Discovery iKit. Includes a look at controlled vocabularies declared in the Resource Discovery Framework (RDF)/Simple Knowledge Organisation System (SKOS) and wikipedia entries. Presented by Tony Ross at the CILIPS Centenary Conference Branch and Group Day which took place 5 Jun 2008.
Poster: Using Open Source Tools to Improve Access to Oral History CollectionsBecky Yoose
Presented at the Library Technology Conference 2011 in St. Paul, MN.
Program Description: Oral history collections provide a wealth of information, yet current practices in metadata creation and
access limit the amount of information within the interview transcripts that can be discovered. This
poster describes the Miami University Libraries current project of using Open Source Software in
creating enhanced access to our Oral History collection. The Oral History Project at Miami University contains over 100 interviews pertaining to experiences at the University, with transcripts for over half of
the interviews. The poster will describe the process of batch processing transcripts using OpenCalais, a web service that automates the creation of metadata for content using natural language processing and machine learning, and displaying both the transcripts and metadata in the content management system Drupal using various modules. We will discuss the results from the comparison of machine generated
and human generated metadata in this project and the benefits and concerns surrounding both methods. Future project developments will also be included.
1. The document discusses the EOSC Dataset Minimum Information (EDMI) approach for exposing research data in the European Open Science Cloud (EOSC).
2. EDMI defines a set of 12 minimum metadata properties to facilitate finding and accessing datasets without being overly descriptive.
3. The approach was developed by engaging EOSC demonstrator data repositories and repositories to propose methods for exposing metadata in a simple and sustainable way.
Metadata enriching and discovery at Solent University Library Getaneh Alemu
This document discusses metadata enriching and discovery at Solent University. It begins with introductions and context about how enriched, linked, open and filtered metadata drives resource usage. It then discusses several principles of metadata including sufficiency, necessity, user convenience, representation and standardization. The document outlines how Solent University has enriched its metadata by importing subject headings and authorities. It discusses metadata linking, openness, filtering and usage. Overall it emphasizes the importance of enriching metadata and keeping interfaces simple while maximizing resource discovery and usage.
The document discusses preservation metadata and introduces the PREMIS Data Dictionary. It defines preservation metadata as information used to support the digital preservation process. Key points include:
- Preservation metadata is necessary for understanding and reusing digital information over time.
- The PREMIS Data Dictionary defines a core set of preservation metadata elements and provides strategies for managing this metadata.
- A survey found that while cultural heritage organizations are implementing digital repositories, they have little experience with long-term digital preservation and it is unclear if the metadata being collected will be sufficient.
The document discusses preservation metadata, which is metadata that supports the long-term preservation of digital objects. It introduces the PREMIS data dictionary, which defines a core set of preservation metadata elements. The PREMIS data dictionary was developed by an international working group to provide implementable semantic units for preservation metadata. The document also discusses how other metadata standards, such as those for records management and digitization initiatives, contribute to digital preservation.
Starting from scratch – building the perfect digital repositoryVioleta Ilik
By establishing a digital repository on the Feinberg School of Medicine (FSM), Northwestern University, Chicago campus, we anticipate to gain ability to create, share, and preserve attractive, functional, and citable digital collections and exhibits. Galter Health Sciences Library did not have a repository as of November 2014. In just a few moths we formed a small team that was charged at looking to select the most suitable open source platform for our digital repository software. We followed the National Library of Medicine master evaluation criteria by looking at various factors that included: functionality, scalability, extensibility, interoperability, ease of deployment, system security, system, physical environment, platform support, demonstrated successful deployments, system support, strength of development community, stability of development organization, and strength of technology roadmap for the future. These factors are important for our case considering the desire to connect the digital repository with another platform that was an essential piece in the big FSM picture – VIVO. VIVO is a linked data platform that serves as a researchers’ hub and which provides the names of researchers from academic institutions along with their research output, affiliation, research overview, service, background, researcher’s identities, teaching, and much more.
Managing provenance in the Social Sciences: the Data Documentation Initiative...ARDC
Slides from webinar: Provenance and social science data. Presented on 15 March 2017. Presenter was Dr Steve McEachern, Director Australian Data Archive
FULL webinar recording: https://youtu.be/elPcKqWoOPg
1. Dr Steve McEachern (Director, Aust Data Archive) Data Documentation Initiative (DDI: http://www.ddialliance.org/): A free, international standard for describing data produced by surveys and other observational methods in the social, behavioral, economic, and health sciences. It can document and manage different stages in the research data lifecycle, eg conceptualization, collection, processing, distribution, discovery, and archiving. Documenting data with DDI facilitates understanding, interpretation, and use -- by people, software systems, and computer networks.
Dp Geosc Info Presentation Final Version 2Smita Chandra
Digital preservation of geoscience information is important to ensure long-term access to valuable scientific data. For example, data from the 1975 Viking Mars mission was corrupted and unusable after 20 years. The presentation discusses the importance of digital preservation, outlines different types of digital information and threats. It also describes institutional repositories, digital preservation strategies like the OAIS model, and the current scenario in India. The research aims to test implementing an OAIS-compliant preservation layer in a geoscience institutional repository.
UK Committee on RDA, RDA Day: New Tools for the Future of Cataloguing - Jenny...CILIP MDG
“The RDA Day is programmed by the UK Committee on RDA. Using activities and games throughout informative presentations, the RDA Day will inform and engage metadata practitioners and managers on a content standard which integrates well with the metadata needs of the 21st century”
Paper presented on the UKCoR RDA Day during the Metadata & Discovery Group Conference (6th - 8th Sept 2023 at IET Austin Court, Birmingham)
Challenges to implementation - Jenny WrightCILIP MDG
“The RDA Day is programmed by the UK Committee on RDA. Using activities and games throughout informative presentations, the RDA Day will inform and engage metadata practitioners and managers on a content standard which integrates well with the metadata needs of the 21st century”
Paper presented on the UKCoR RDA Day during the Metadata & Discovery Group Conference (6th - 8th Sept 2023 at IET Austin Court, Birmingham)
Application Profiles in RDA - Jenny WrightCILIP MDG
“The RDA Day is programmed by the UK Committee on RDA. Using activities and games throughout informative presentations, the RDA Day will inform and engage metadata practitioners and managers on a content standard which integrates well with the metadata needs of the 21st century”
Paper presented on the UKCoR RDA Day during the Metadata & Discovery Group Conference (6th - 8th Sept 2023 at IET Austin Court, Birmingham)
The Official RDA Toolkit - Opportunities for Efficiency - Thurstan YoungCILIP MDG
“The RDA Day is programmed by the UK Committee on RDA. Using activities and games throughout informative presentations, the RDA Day will inform and engage metadata practitioners and managers on a content standard which integrates well with the metadata needs of the 21st century”
Paper presented on the UKCoR RDA Day during the Metadata & Discovery Group Conference (6th - 8th Sept 2023 at IET Austin Court, Birmingham)
The Official RDA Toolkit - Opportunities for Enrichment - Thurstan YouingCILIP MDG
The document discusses opportunities for enriching metadata in the Official RDA Toolkit. It provides background on extension plans, representative expressions, and data provenance. An example is given of recording an extension plan and representative expression for a multi-volume work. The extension plan vocabulary and representative expression elements are shown as ways to enrich RDA descriptions through structured, encoded values.
“The RDA Day is programmed by the UK Committee on RDA. Using activities and games throughout informative presentations, the RDA Day will inform and engage metadata practitioners and managers on a content standard which integrates well with the metadata needs of the 21st century”
Presented on the UKCoR RDA Day during the Metadata & Discovery Group Conference (6th - 8th Sept 2023 at IET Austin Court, Birmingham)
RDA methods, scenarios, tools - Gordon DunsireCILIP MDG
“The RDA Day is programmed by the UK Committee on RDA. Using activities and games throughout informative presentations, the RDA Day will inform and engage metadata practitioners and managers on a content standard which integrates well with the metadata needs of the 21st century”
Paper presented on the UKCoR RDA Day during the Metadata & Discovery Group Conference (6th - 8th Sept 2023 at IET Austin Court, Birmingham)
Poster: What’s in a name? Re-Discovering cataloguing and index through metada...CILIP MDG
In 2019 CILIP’s Cataloguing and Indexing Group changed its name to the Metadata and Discovery Group. This poster will showcase the transition of the look and feel of the group’s logo and the process of designing and new one.
Poster presented at the CILIP Metadata and Discovery Group (MDG) Conference & UKCoR RDA Day (6th - 8th Sept 2023 at IET Austin Court, Birmingham).
Poster: Revamping our in-house cataloguing training / Victoria Parkinson (Kin...CILIP MDG
With hybrid working and a new LMS, we are revamping our in-house cataloguing training. We are learning from our teaching librarians and using the tools we have, such as Moodle, to create cataloguing training that allows anyone with an interest to learn the basics and making the best use of face-to-face time for putting those skills into practice. Over the past eight years we’ve adapted and updated our in-house training, and I’ll also talk about how we decide what to teach colleagues, and how we try to make the best use of staff time to keep skills up when cataloguing is one of many competing priorities and shared across several teams. Between staff turnover and COVID lockdowns and service changes, we are starting almost from scratch in building a pool of staff who can catalogue the material our suppliers can’t provide records for, which is an excellent time to take stock of what our cataloguing needs are, and advocate for the importance of creating and upgrading good quality records and why we need to build these skills in-house.
Poster presented at the Metadata & Discovery Group Conference & RDA Day (6th - 8th Sept 2023 at IET Austin Court, Birmingham)
Poster: FAST : can it lighten the load, and what is the impact? / Jenny Wrigh...CILIP MDG
This poster presents the Faceted Application of Subject Terminology, giving an overview of the scheme, its advantages and potential issues, and its practical implementation. It will demonstrate that FAST is an important development for those interested in Linked Data, and the ways in which it is a useful tool for discovery in any system.
Poster presented at the CILIP Metadata and Discovery Group (MDG) Conference & UKCoR RDA Day (6th - 8th Sept 2023 at IET Austin Court, Birmingham).
Poster: The West Midlands Evidence Repository (WMER) : a regional collaborati...CILIP MDG
The West Midlands Evidence Repository (WMER) was born from a pre-pandemic recognition by managers of Knowledge and Library Services (KLSs) of 8 NHS Trusts in the West Midlands region of the need for a repository. This was to replace existing provision, or recognition of national priorities or local needs to record, collect, and share research, as well as potential for sharing patient information leaflets or guidelines. Some managers and services had previous experience of repositories, as well as being part of a national pilot. WMER, however, represented a new start for all to work in collaboration to establish a new service. The consortium would enable sharing of both costs and experience.
Initially, different repository suppliers were investigated by the KLS that had had a long-established repository, taking on board the experience of the group from the national pilot. The Atmire Open Repository platform was chosen as it met the consortium’s needs and had a proven track record of other collaborative repositories in the NHS. Financing was taken on by one Trust and the on-boarding was led in partnership between that Trust and the Trust that had undertaken the initial investigation.
With the initial on-boarding completed and the test server set-up, the group took a step back to ensure they worked together as a collaborative going forward. Collaborative work between the KLSs was facilitated by the formal creation of two groups, a Managers Group for overall approval and financial decision making and an Operational Group handling the setup and administration of the repository for the consortium. The Operational Group is led by the service with most experience of managing repositories and the lead of it acts as liaison between the two groups, with each group having representation from the eight organisations. Learning from other regional collaborations the Future NHS site was used as a collaborative workspace and Teams as the main means of communication.
The setup of the repository was completed on time after three months. There was initially a steep learning curve for all, especially the Operational Group who undertook this process. The group identified key metadata and metadata standards for the repository, including the use of ORCIDs and the use of Wessex Classification as a controlled vocabulary. The setup process was facilitated by the collaborative nature of the project as the variety of experience in the group was a great benefit. It should be noted support from the suppliers was specifically related to technical support only.
The collaborative nature of the project also allowed work to be shared, and tasks were given to members to be undertaken independently. However, a downside of collaborative projects is that decisions can take longer to be inclusive...
Poster presented at the Metadata & Discovery Group Conference & RDA Day (6th - 8th Sept 2023 at IET Austin Court, Birmingham)
Poster: Updating the Wessex Classification Scheme for UK health libraries : a...CILIP MDG
The Wessex Classification Scheme was created by healthcare librarians in the South West of England, and was loosely based on the US National Library of Medicine classification. The scheme is widely used in healthcare libraries across the UK, both inside and outside the NHS. Although the scheme has gone through several revisions, there has been no major update since 2015, so the Wessex Classification Scheme Oversight Group was formed in September 2022 with the support of NHS England. The group aims to bring knowledge and skills from UK health library networks to improve the scheme and offers a chance for participants to develop skills in working with classification and subject indexing, and the opportunity to network widely. By forming a working group, it ensures the longevity of the scheme and shares the maintenance work more widely.
Initially, members were asked which parts of the scheme they felt needed updating the most and sub-groups were formed for LGBTQ+ issues and gender identity (the Pride sub-group), Ethnicity and Race, and Learning Disability and Neurodiversity (the LDN sub-group) as well as a smaller team working on ‘quick and simple’ updates....
Poster presented at the CILIP Metadata and Discovery Group (MDG) Conference & UKCoR RDA Day (6th - 8th Sept 2023 at IET Austin Court, Birmingham).
Revamping in-house cataloguing training / Victoria Parkinson (King's College ...CILIP MDG
With hybrid working and a new LMS, we are revamping our in-house cataloguing training. We are learning from our teaching librarians and using the tools we have, such as Moodle, to create cataloguing training that allows anyone with an interest to learn the basics and making the best use of face-to-face time for putting those skills into practice. Over the past eight years we’ve adapted and updated our in-house training, and I’ll also talk about how we decide what to teach colleagues, and how we try to make the best use of staff time to keep skills up when cataloguing is one of many competing priorities and shared across several teams. Between staff turnover and COVID lockdowns and service changes, we are starting almost from scratch in building a pool of staff who can catalogue the material our suppliers can’t provide records for, which is an excellent time to take stock of what our cataloguing needs are, and advocate for the importance of creating and upgrading good quality records and why we need to build these skills in-house.
Lightning Talk presented at the Metadata & Discovery Group Conference & RDA Day (6th - 8th Sept 2023 at IET Austin Court, Birmingham)
UK NACO funnel : progress, obstacles, and solutions / Martin Kelleher (Univer...CILIP MDG
This Lightning Talk will provide a quick update on latest progress with the now established UK NACO Funnel, which allows participating institutions to contribute to Library of Congress / PCC authority control. The presentation will include a summary of the purpose of the funnel, details of latest expansion, problems and solutions with data submission software, and further plans and collaborations.
Lightning Talk presented at the Metadata & Discovery Group Conference & RDA Day (6th - 8th Sept 2023 at IET Austin Court, Birmingham)
Ship[w]right[e]s? : the challenges of cataloguing reports from scientific exp...CILIP MDG
This document discusses the challenges of cataloguing reports from scientific expeditions, using the Challenger Reports as an example. It notes that there were 83 Challenger Reports published by 47 authors, held across 4 sites in 50 different sectional libraries, with 220 bib records created for this work. It also mentions the opportunity for the Natural History Museum to think about metadata across the entire museum collection as part of an effort to move specimens to a new location.
BFI Reuben Library : an RDA implementation story / Anastasia Kerameos (BFI Re...CILIP MDG
“From 1st January 2024, Adlib will no longer be supported or maintained by Axiell.” This statement acted as the catalyst for action, enabling the release of resources to implement significant changes to the BFI Reuben Library’s record structure, which in turn prompted a deeper look into our current cataloguing practices and future requirements.
Upgrading to Axiell Collections will allow the library to implement new RDA more fully – we had previously adopted some aspects but not all – and, importantly, it will allow us to better align our data structure with that of the organisation’s other collections, making it easier to manage and making it compatible with further planned system developments. By the time of the conference in September we will be cataloguing to an under the bonnet Work – Expression – Manifestation – Item (WEMI) record hierarchy and new cataloguing guidelines.
Having watched all the webinars available, having read every piece of documentation which seemed relevant, having spent hours reading and re-reading the contents of the RDA Toolkit we are currently working on the last stages of our application profile whilst still debating issues around putting the theory into practice, especially in the area of aggregates and diachronic works. I do not suggest I have all the answers, far from it, but by sharing the story of our journey, that of a medium sized non-academic library of specialist mostly print collections and illustrating it with practical examples I hope my presentation will be of use to others currently travelling a similar path.
Paper presented at the Metadata & Discovery Group Conference & RDA Day (6th - 8th Sept 2023 at IET Austin Court, Birmingham)
RDA implementation at the British Library / Thurstan Young (British Library)CILIP MDG
On 23rd May 2023, the RDA Board announced that the original RDA Toolkit will be removed in May 2027. All RDA users will need to be prepared for transition to the official RDA Toolkit before then. As previously announced, a Countdown Clock will start running in May 2026, a year before the sunset date.
This paper will provide an update on the British Library’s plans for implementation of the new RDA Toolkit, following completion of the RDA Toolkit Restructure and Redesign (3R) project. It will provide an overview of the timeline and scope for implementation as well as describing the training and documentation underpinning the implementation and the support available to other institutions for their implementation.
Paper presented at the Metadata & Discovery Group Conference & RDA Day (6th - 8th Sept 2023 at IET Austin Court, Birmingham)
Community forward : developing descriptive cataloguing of rare materials (RDA...CILIP MDG
Since 2013, Resource Description and Access (RDA) has been the chief cataloguing standard used in the United States. In 2019, the RDA Steering Committee previewed a new version of the RDA Toolkit, which introduced substantial changes, such as replacing instructions with a series of options, adding new concepts such as “nomens” and “diachronic works,” and replacing the prior organisation with a broader intellectual framework. This revised Toolkit became the official RDA Toolkit in December 2020, with major cataloguing bodies planning to adopt it in the coming years. Some cataloguers have expressed concerns regarding the official RDA Toolkit, particularly around cost and training required to learn the new standard.
In response to these concerns, the RBMS RDA Editorial Group, a group of volunteers from the Association of College and Research Libraries’ Rare Books and Manuscripts Section, developed a new manual, Descriptive Cataloging of Rare Materials (RDA Edition). DCRMR is informed by core principles of community and sustainability while employing open-access publication models and infrastructure. Designed in response to community feedback, it presents instructions in cataloguing workflow order using clear language while remaining aligned to the official RDA Toolkit and RDA element sets. The manual was approved in February 2022 in its first iteration and continues to be actively developed and updated. This presentation will discuss why the editorial group created an open and free manual; the process and tools for creating the manual, including the use of GitHub to publish a cataloguing standard; and outcomes to date.
Paper presented at the Metadata & Discovery Group Conference & RDA Day (6th - 8th Sept 2023 at IET Austin Court, Birmingham)
The West Midlands Evidence Repository (WMER) : a regional collaboration proje...CILIP MDG
The West Midlands Evidence Repository (WMER) was born from a pre-pandemic recognition by managers of Knowledge and Library Services (KLSs) of 8 NHS Trusts in the West Midlands region of the need for a repository. This was to replace existing provision, or recognition of national priorities or local needs to record, collect, and share research, as well as potential for sharing patient information leaflets or guidelines. Some managers and services had previous experience of repositories, as well as being part of a national pilot. WMER, however, represented a new start for all to work in collaboration to establish a new service. The consortium would enable sharing of both costs and experience.
Initially, different repository suppliers were investigated by the KLS that had had a long-established repository, taking on board the experience of the group from the national pilot. The Atmire Open Repository platform was chosen as it met the consortium’s needs and had a proven track record of other collaborative repositories in the NHS. Financing was taken on by one Trust and the on-boarding was led in partnership between that Trust and the Trust that had undertaken the initial investigation.
With the initial on-boarding completed and the test server set-up, the group took a step back to ensure they worked together as a collaborative going forward. Collaborative work between the KLSs was facilitated by the formal creation of two groups, a Managers Group for overall approval and financial decision making and an Operational Group handling the setup and administration of the repository for the consortium. The Operational Group is led by the service with most experience of managing repositories and the lead of it acts as liaison between the two groups, with each group having representation from the eight organisations. Learning from other regional collaborations the Future NHS site was used as a collaborative workspace and Teams as the main means of communication.
The setup of the repository was completed on time after three months. There was initially a steep learning curve for all, especially the Operational Group who undertook this process. The group identified key metadata and metadata standards for the repository, including the use of ORCIDs and the use of Wessex Classification as a controlled vocabulary. The setup process was facilitated by the collaborative nature of the project as the variety of experience in the group was a great benefit. It should be noted support from the suppliers was specifically related to technical support only.
The collaborative nature of the project also allowed work to be shared, and tasks were given to members to be undertaken independently. However, a downside of collaborative projects is that decisions can take longer to be inclusive...
Paper presented at the Metadata & Discovery Group Conference & RDA Day (6th - 8th Sept 2023 at IET Austin Court, Birmingham)
Authority of assertion in repository contributions to the PID graph / George ...CILIP MDG
The principles surrounding Linked Open Data and their implementation within digital libraries are well understood. Such implementations may be challenging, but successes are now well documented and continue to demonstrate the benefits of disseminating and enriching existing metadata with improved semantics and relational associations. Often facilitated in machine-readability enhancements to metadata by harnessing serializations of the Resource Description Framework (RDF) and its reliance of URIs, these LOD approaches have ensured digital libraries, and similar GLAMR initiatives elsewhere, contribute to the growing knowledge graphs associated with the wider semantic web by declaring statements of fact about web entities. Within open scholarly ecosystems a growing use of persistent identifiers (PIDs) to define and link scholarly entities has emerged, e.g., DOIs, ORCIDs, etc. The requirement for greater URI persistence has been motivated by several developments within the scholarly space; suffice to state that, when combined with appropriate structured data, PIDs can support improvements to resource discovery, as well as facilitate contributions to the ‘PID graph’ – a scholarly data graph describing and declaring associative relations between scholarly entities.
While the increased adoption of PIDs has the potential to transform scholarship, ensuring that these PIDs are used appropriately, encoded correctly within metadata, and that all relevant relational associations between scholarly entities are declared presents challenges. This is especially true within open scholarly repositories, from where many contributions to the PID graph will be made but – unlike many LOD contexts – from where the authority to assert specific relations may not always exist. Such declarations need to demonstrate reliability and provenance and are central to the interlinking of heterogeneous textual objects, datasets, software, research instruments, equipment, and the related PIDs these items may generate, such as for people, organizations, or other abstract entities.
This paper will explore the issues that arise when levels of authority to assert are lacking or are uncertain, and review results from a related study exploring the ‘PID literacy’ of scholars...
Paper presented at the Metadata & Discovery Group Conference & RDA Day (6th - 8th Sept 2023 at IET Austin Court, Birmingham)
GraphRAG for Life Science to increase LLM accuracyTomaz Bratanic
GraphRAG for life science domain, where you retriever information from biomedical knowledge graphs using LLMs to increase the accuracy and performance of generated answers
Ivanti’s Patch Tuesday breakdown goes beyond patching your applications and brings you the intelligence and guidance needed to prioritize where to focus your attention first. Catch early analysis on our Ivanti blog, then join industry expert Chris Goettl for the Patch Tuesday Webinar Event. There we’ll do a deep dive into each of the bulletins and give guidance on the risks associated with the newly-identified vulnerabilities.
Unlock the Future of Search with MongoDB Atlas_ Vector Search Unleashed.pdfMalak Abu Hammad
Discover how MongoDB Atlas and vector search technology can revolutionize your application's search capabilities. This comprehensive presentation covers:
* What is Vector Search?
* Importance and benefits of vector search
* Practical use cases across various industries
* Step-by-step implementation guide
* Live demos with code snippets
* Enhancing LLM capabilities with vector search
* Best practices and optimization strategies
Perfect for developers, AI enthusiasts, and tech leaders. Learn how to leverage MongoDB Atlas to deliver highly relevant, context-aware search results, transforming your data retrieval process. Stay ahead in tech innovation and maximize the potential of your applications.
#MongoDB #VectorSearch #AI #SemanticSearch #TechInnovation #DataScience #LLM #MachineLearning #SearchTechnology
Project Management Semester Long Project - Acuityjpupo2018
Acuity is an innovative learning app designed to transform the way you engage with knowledge. Powered by AI technology, Acuity takes complex topics and distills them into concise, interactive summaries that are easy to read & understand. Whether you're exploring the depths of quantum mechanics or seeking insight into historical events, Acuity provides the key information you need without the burden of lengthy texts.
Let's Integrate MuleSoft RPA, COMPOSER, APM with AWS IDP along with Slackshyamraj55
Discover the seamless integration of RPA (Robotic Process Automation), COMPOSER, and APM with AWS IDP enhanced with Slack notifications. Explore how these technologies converge to streamline workflows, optimize performance, and ensure secure access, all while leveraging the power of AWS IDP and real-time communication via Slack notifications.
HCL Notes und Domino Lizenzkostenreduzierung in der Welt von DLAUpanagenda
Webinar Recording: https://www.panagenda.com/webinars/hcl-notes-und-domino-lizenzkostenreduzierung-in-der-welt-von-dlau/
DLAU und die Lizenzen nach dem CCB- und CCX-Modell sind für viele in der HCL-Community seit letztem Jahr ein heißes Thema. Als Notes- oder Domino-Kunde haben Sie vielleicht mit unerwartet hohen Benutzerzahlen und Lizenzgebühren zu kämpfen. Sie fragen sich vielleicht, wie diese neue Art der Lizenzierung funktioniert und welchen Nutzen sie Ihnen bringt. Vor allem wollen Sie sicherlich Ihr Budget einhalten und Kosten sparen, wo immer möglich. Das verstehen wir und wir möchten Ihnen dabei helfen!
Wir erklären Ihnen, wie Sie häufige Konfigurationsprobleme lösen können, die dazu führen können, dass mehr Benutzer gezählt werden als nötig, und wie Sie überflüssige oder ungenutzte Konten identifizieren und entfernen können, um Geld zu sparen. Es gibt auch einige Ansätze, die zu unnötigen Ausgaben führen können, z. B. wenn ein Personendokument anstelle eines Mail-Ins für geteilte Mailboxen verwendet wird. Wir zeigen Ihnen solche Fälle und deren Lösungen. Und natürlich erklären wir Ihnen das neue Lizenzmodell.
Nehmen Sie an diesem Webinar teil, bei dem HCL-Ambassador Marc Thomas und Gastredner Franz Walder Ihnen diese neue Welt näherbringen. Es vermittelt Ihnen die Tools und das Know-how, um den Überblick zu bewahren. Sie werden in der Lage sein, Ihre Kosten durch eine optimierte Domino-Konfiguration zu reduzieren und auch in Zukunft gering zu halten.
Diese Themen werden behandelt
- Reduzierung der Lizenzkosten durch Auffinden und Beheben von Fehlkonfigurationen und überflüssigen Konten
- Wie funktionieren CCB- und CCX-Lizenzen wirklich?
- Verstehen des DLAU-Tools und wie man es am besten nutzt
- Tipps für häufige Problembereiche, wie z. B. Team-Postfächer, Funktions-/Testbenutzer usw.
- Praxisbeispiele und Best Practices zum sofortigen Umsetzen
Introduction of Cybersecurity with OSS at Code Europe 2024Hiroshi SHIBATA
I develop the Ruby programming language, RubyGems, and Bundler, which are package managers for Ruby. Today, I will introduce how to enhance the security of your application using open-source software (OSS) examples from Ruby and RubyGems.
The first topic is CVE (Common Vulnerabilities and Exposures). I have published CVEs many times. But what exactly is a CVE? I'll provide a basic understanding of CVEs and explain how to detect and handle vulnerabilities in OSS.
Next, let's discuss package managers. Package managers play a critical role in the OSS ecosystem. I'll explain how to manage library dependencies in your application.
I'll share insights into how the Ruby and RubyGems core team works to keep our ecosystem safe. By the end of this talk, you'll have a better understanding of how to safeguard your code.
In the rapidly evolving landscape of technologies, XML continues to play a vital role in structuring, storing, and transporting data across diverse systems. The recent advancements in artificial intelligence (AI) present new methodologies for enhancing XML development workflows, introducing efficiency, automation, and intelligent capabilities. This presentation will outline the scope and perspective of utilizing AI in XML development. The potential benefits and the possible pitfalls will be highlighted, providing a balanced view of the subject.
We will explore the capabilities of AI in understanding XML markup languages and autonomously creating structured XML content. Additionally, we will examine the capacity of AI to enrich plain text with appropriate XML markup. Practical examples and methodological guidelines will be provided to elucidate how AI can be effectively prompted to interpret and generate accurate XML markup.
Further emphasis will be placed on the role of AI in developing XSLT, or schemas such as XSD and Schematron. We will address the techniques and strategies adopted to create prompts for generating code, explaining code, or refactoring the code, and the results achieved.
The discussion will extend to how AI can be used to transform XML content. In particular, the focus will be on the use of AI XPath extension functions in XSLT, Schematron, Schematron Quick Fixes, or for XML content refactoring.
The presentation aims to deliver a comprehensive overview of AI usage in XML development, providing attendees with the necessary knowledge to make informed decisions. Whether you’re at the early stages of adopting AI or considering integrating it in advanced XML development, this presentation will cover all levels of expertise.
By highlighting the potential advantages and challenges of integrating AI with XML development tools and languages, the presentation seeks to inspire thoughtful conversation around the future of XML development. We’ll not only delve into the technical aspects of AI-powered XML development but also discuss practical implications and possible future directions.
Ocean lotus Threat actors project by John Sitima 2024 (1).pptxSitimaJohn
Ocean Lotus cyber threat actors represent a sophisticated, persistent, and politically motivated group that poses a significant risk to organizations and individuals in the Southeast Asian region. Their continuous evolution and adaptability underscore the need for robust cybersecurity measures and international cooperation to identify and mitigate the threats posed by such advanced persistent threat groups.
Programming Foundation Models with DSPy - Meetup SlidesZilliz
Prompting language models is hard, while programming language models is easy. In this talk, I will discuss the state-of-the-art framework DSPy for programming foundation models with its powerful optimizers and runtime constraint system.
Webinar: Designing a schema for a Data WarehouseFederico Razzoli
Are you new to data warehouses (DWH)? Do you need to check whether your data warehouse follows the best practices for a good design? In both cases, this webinar is for you.
A data warehouse is a central relational database that contains all measurements about a business or an organisation. This data comes from a variety of heterogeneous data sources, which includes databases of any type that back the applications used by the company, data files exported by some applications, or APIs provided by internal or external services.
But designing a data warehouse correctly is a hard task, which requires gathering information about the business processes that need to be analysed in the first place. These processes must be translated into so-called star schemas, which means, denormalised databases where each table represents a dimension or facts.
We will discuss these topics:
- How to gather information about a business;
- Understanding dictionaries and how to identify business entities;
- Dimensions and facts;
- Setting a table granularity;
- Types of facts;
- Types of dimensions;
- Snowflakes and how to avoid them;
- Expanding existing dimensions and facts.
Cosa hanno in comune un mattoncino Lego e la backdoor XZ?Speck&Tech
ABSTRACT: A prima vista, un mattoncino Lego e la backdoor XZ potrebbero avere in comune il fatto di essere entrambi blocchi di costruzione, o dipendenze di progetti creativi e software. La realtà è che un mattoncino Lego e il caso della backdoor XZ hanno molto di più di tutto ciò in comune.
Partecipate alla presentazione per immergervi in una storia di interoperabilità, standard e formati aperti, per poi discutere del ruolo importante che i contributori hanno in una comunità open source sostenibile.
BIO: Sostenitrice del software libero e dei formati standard e aperti. È stata un membro attivo dei progetti Fedora e openSUSE e ha co-fondato l'Associazione LibreItalia dove è stata coinvolta in diversi eventi, migrazioni e formazione relativi a LibreOffice. In precedenza ha lavorato a migrazioni e corsi di formazione su LibreOffice per diverse amministrazioni pubbliche e privati. Da gennaio 2020 lavora in SUSE come Software Release Engineer per Uyuni e SUSE Manager e quando non segue la sua passione per i computer e per Geeko coltiva la sua curiosità per l'astronomia (da cui deriva il suo nickname deneb_alpha).
Main news related to the CCS TSI 2023 (2023/1695)Jakub Marek
An English 🇬🇧 translation of a presentation to the speech I gave about the main changes brought by CCS TSI 2023 at the biggest Czech conference on Communications and signalling systems on Railways, which was held in Clarion Hotel Olomouc from 7th to 9th November 2023 (konferenceszt.cz). Attended by around 500 participants and 200 on-line followers.
The original Czech 🇨🇿 version of the presentation can be found here: https://www.slideshare.net/slideshow/hlavni-novinky-souvisejici-s-ccs-tsi-2023-2023-1695/269688092 .
The videorecording (in Czech) from the presentation is available here: https://youtu.be/WzjJWm4IyPk?si=SImb06tuXGb30BEH .
Threats to mobile devices are more prevalent and increasing in scope and complexity. Users of mobile devices desire to take full advantage of the features
available on those devices, but many of the features provide convenience and capability but sacrifice security. This best practices guide outlines steps the users can take to better protect personal devices and information.
Your One-Stop Shop for Python Success: Top 10 US Python Development Providersakankshawande
Simplify your search for a reliable Python development partner! This list presents the top 10 trusted US providers offering comprehensive Python development services, ensuring your project's success from conception to completion.
Your One-Stop Shop for Python Success: Top 10 US Python Development Providers
Metadata and archival discoverability / Getaneh Alemu (Solent University)
1. METADATA & ARCHIVAL
DISCOVERABILITY
Getaneh Alemu
Cataloguing & Metadata Librarian, Solent University
CILIP Metadata & Discovery Group Conference,
Birmingham, September 6th - 8th, 2023
2. Solent University
ARCHIVE DEFINITION
• “Any records with long-term continuing value that have been kept either
because they may be necessary for ongoing organisational purposes to their
creating body or because they have additional research value.” (Williams,
2006)
• “Records preserved for their enduring value and informative content.”
(Miller, 2017)
• Qualities of archival content: static, permanence, uniqueness and
authentic. (Miller, 2017)
3. Solent University
FROM DATA – EVIDENCE - ARCHIVE
Millar, L. (2017). Archives: Principles and practices. Facet
Publishing.
4. Solent University
Library Materials:
•Standardised genres and formats (e.g., books,
journals, etc.).
•Follows known bibliographic elements (ISBD).
•Often a copy is available in another library.
•Flat structure.
Archival Materials:
No standardised format; can include letters,
diaries, photographs, etc.
Does not follow a bibliographic standard/control.
Materials are unique and available at a single
location.
Hierarchical structure.
5. Solent University
LIBRARY VS ARCHIVAL MATERIALS
Library Books Archival Materials
Purpose Study, research, leisure, reference Research, historical sources, cultural memory
Availability Published, widely available, common holdings
Unique records, often found only at a single
place
Arrangement Classification by subject/author/title (DDC, LC) Organised by provenance or original order
Circulation/use
Generally circulated and can be checked out.
Replaceable
Special access, often locked securely.
Irreplaceable
Durability
Printed for multiple uses but may wear over
time
Continuum of care, fragile, requiring special
care and storage for long-term preservation
(Brown, 2019; Miller, 2017; Williams,
2006)
6. Solent University
Principle of Provenance (or
Respect des Fonds)
ORIGIN/SOURCE AUTHENTICITY CONTEXT OWNER/CREATOR CREDIBILITY
(Douglas, 2018; Miller, 2017; Williams, 2006)
7. Solent University
Principle of Original Order
ARRANGEMENT SEQUENCE CREATOR'S INTENT CONTEXTUAL
INFORMATION
(Miller, 2017; Williams,
2006)
8. Solent University
ARCHIVAL
METADATA
Structure and granularity; rich, dense, varied
and multi-layered (Gracy, 2014; ISO 23081-1:2006; 2017)
Accuracy and consistency (Alemu, 2022; Qin & Zeng,
2016)
Finding aids enhancement (Francisco-Revilla, et al.
2014)
User support – improve access and usage
(Williams, 2006)
Interoperability and data sharing (Qin & Zeng, 2016)
9. Solent University
METADATA FOR ARCHIVES
9
Organising and
managing
records.
1
Maintaining
original
archival order.
2
Enhancing
findability and
discoverability.
3
Validating
records’
authenticity.
4
Facilitating
long-term
preservation.
5
( ISO 23081-1:2006; 2017)
10. Solent University
ARCHIVAL METADATA STANDARDS: ISAD(G)
Source: (ICA, 2000)
As defined by ICA (2000), ISAD(G),
the General International Standard
Archival Description is a globally
recognised metadata standard.
Used in numerous countries
worldwide.
Consists of 26 main data elements
focusing on description.
Primarily utilised with finding-
tools.
12. Solent University
Descriptions
Ensure the creation of
accurate, complete and
consistent descriptions.
Finding aids
Allow users search, identify,
select, obtain and explore
archival content.
Data sharing
Facilitate the sharing of data
across various archives.
Interoperability
Allow integration of varied
descriptions into a cohesive
information system.
ISAD(G) – PURPOSES
Source: (Francisco-Revilla et al., 2014; ICA, 2000; Scifleet, 2001)
14. Solent University
Mandatory Data Elements:
Reference code: Unique identifier for the archival material.
Example: GB 3600 MAC
Title: The name given to the material or collection.
Example: The Philip Mackie Collection
Creator: The person, agency, or entity that produced the archival material.
Example: Mackie, Philip 1918-1985
Date(s): The time period during which the archival material was created or accumulated.
Example: 1947 - 1999, predominant 1950 - 1985
Extent of the unit of description: Measurement (in terms of quantity, bulk, or size) of the material.
Example: 163 boxes (12.5 metres)
Level of description: Defines the scope of the material being described, such as item, series, fonds, etc.
Example: Series
ISAD(G) – MANDATORY METADATA ELEMENTS
Source: (ICA, 2000)
15. Solent University
ISAD(G) – STRUCTURE
Source: (ICA, 2000)
CONTENT AND
STRUCTURE AREA
Scope and content
Appraisal,
destruction, and
scheduling data
Accruals
System of
arrangement
CONDITIONS OF
ACCESS AND USE
AREA
Conditions
governing access
Conditions
governing
reproduction
Language/scripts
of material
Physical and
technical specifics
Finding aids
ALLIED
MATERIALS AREA
Info on originals
Info on copies
Linked units of
description
Publication note
NOTES AREA
Notes
DESCRIPTION
CONTROL AREA
Archivist's Note
Rules or
Conventions
Date(s) of
descriptions
16. Solent University
ENCODED ARCHIVAL DESCRIPTION (EAD)
Uses XML (eXtensible Markup Language) to denote
hierarchy, elements and relationships.
Archives use EAD to present structured, both human-
readable and machine-readable descriptions; hierarchical
archival structure
Interoperable format
Compatible with most modern systems that handle XML
data
(Pitti, 1999; Scifleet, 2001; Society of American Archivists, 2002))
17. Solent University
EAD KEY FEATURES
Standardisation and consistency
EAD standardises the presentation of archival descriptions across various institutions and platforms.
Provides a consistent structure and encoding which enhances interaction with archival collections, irrespective of where
they are hosted.
Hierarchical navigation
EAD finding aids offer hierarchical navigation, enabling exploration of collections at different depths.
This hierarchy reflects the physical arrangement of collections, assisting in navigating intricate holdings.
Advantages
Major improvement in archival discoverability.
Presents descriptions in a standardised, structured and accessible online format.
Perfectly melds traditional archival practices with digital advancements.
Overall, makes archives more user-friendly and appealing to both researchers and the general public.
(Pitti, 1999; Scifleet, 2001; Society of American Archivists, 2002, Yakel & Kim, 2005)
21. Solent University
FROM ARCHIVES HUB TO PRIMO
Initial approach with OAI-PMH
Archives Hub provides content via OAI-PMH.
Initial idea: set-up an import profile using this data for automatic updates in
Primo.
Challenges: Archives Hub treats each collection as a single record.
This would result in only two collection level records in Primo due to Alma Import
Profile restrictions. This process would lose the number of records we want
(>100).
So, this solution was not considered feasible.
22. Solent University
FROM ARCHIVES HUB TO PRIMO
Second and pragmatic approach
Aim to have approximately 100 records in total – all the entries in Archives Hub.
Step 1: Download Archives Hub content as EAD (XML).
Step 2: Write a script to convert each sub-series/work into a format compatible
with Alma.
Step 3: EAD to MARC mapping, example: e.g., the first paragraph in the Scope and
Content field was intended as a summary, whereas further paragraphs would give
detailed content information. Therefore, the mapping sends the first paragraph to the
520 and subsequent paragraphs to the 505.
23. Solent University
F R O M A R C H I V E S H U B ( E A D ) TO P R I M O ( M A R C ) M A P P I N G
26. Solent University
IMPACT OF
CATALOGUING
T H E P H I L I P M A C K I E
C O L L E C T I O N
Increased
discoverability
on the Archives
Hub, Google,
Library Catalogue
Improved
access and
usage
received external
scanning and
access requests
(from the United States,
Sweden and the UK -
including one who has used
the collection extensively in
their dissertation).
27. Solent University
KEY
TAKEAWAYS
Metadata is central for archival description,
context and structure.
Metadata ensures authenticity, reliability and
integrity of records.
Accurate and consistent descriptions and finding
aids pivotal for discovery, access and usage.
ISAD(G): supports multi-level and granular
descriptions – adhering to the principles of
provenance, respect des fonds and original order.)
EAD: uses XML for encoding finding aids, an
interoperable format but technically challenging.
28. Solent University
B I B L I O G R A P H Y
Alemu, G. (2022). The future of enriched, linked, open and filtered metadata: making sense of IFLA, LRM, RDA, Linked Data and BIBFRAME. Facet Publishing.
Brown (Ed.). (2019). Archival futures. Facet Publishing.
Douglas, J. (2018). A call to rethink archival creation: Exploring types of creation in personal archives. Archival Science, 18(1), 29-49. doi:https://doi.org/10.1007/s10502-018-9285-
8
Francisco-Revilla, L., Trace, C. B., Li, H., & Buchanan, S. A. (2014). Encoded Archival Description: Data Quality and Analysis. American Society for Information Science and
Technology. Meeting. Proceedings of the ... ASIST Annual Meeting, 51(1), 1–10. https://doi.org/10.1002/meet.2014.14505101043
Gracy, K. F. (2015). Archival description and Linked Data: a preliminary study of opportunities and implementation challenges. Archival Science, 15(3), 239–294.
https://doi.org/10.1007/s10502-014-9216-2
Haynes, D. (2018) Metadata for Information Management and Retrieval: understanding metadata and its use. Facet Publishing.
Hider, P. (2018) Information Resource Description: creating and managing metadata. Facet Publishing.
ICA (2000). ISAD(G) : General International Standard Archival Description (ISAD (G), (2nd ed.). ICA. Retrieved from
https://repositories.lib.utexas.edu/bitstream/handle/2152/14140/ISAD%28G%29.pdf?sequence=2&isAllowed=y
Millar, L. (2017). Archives: Principles and practices. Facet Publishing.
Pitti D. V. (1999). Encoded archival description an introduction and overview. D-Lib Magazine. Retrieved from https://www.dlib.org/dlib/november99/11pitti.html
Scifleet, P. (2001). International Standard Archival Description ISAD(G). Archival Science, 1(4), 397–402. A book review. https://doi.org/10.1023/A:1016058816263
Society of American Archivists. (2002). Encoded Archival Description tag library version 2002 ; prepared by the Encoded Archival Description Working Group of the Society Of
American Archivists and the Network Development And Marc Standards Office of The Library of Congress. Retrieved from
https://www2.archivists.org/sites/all/files/EAD2002TL_5-03-V2.pdf
Williams. (2006). Managing archives : foundations, principles and practice. Chandos.
Yakel, E. and Kim, J. (2005), Adoption and diffusion of Encoded Archival Description. J. Am. Soc. Inf. Sci., 56: 1427-1437. https://doi.org/10.1002/asi.20236. Chandos.
Zeng, M. L., & Qin, J. (2016). Metadata (2nd ed.). Facet Publishing.
Metadata plays a key role in describing and enhancing archival collections. This presentation covers the standards, metadata elements and tools chosen to describe two unique archival collections at Solent University Library. Using the International Standard Archival Description (General) (ISAD (G)), as a framework for the creation and management of archival descriptions, EAD as a serialisation format and the Jisc’s Archives Hub Editor as an application, the archival works of Philip Mackie and Ken Russell were described.
A decision was made to comply with ISAD(G)’s mandatory elements such as Reference code, Title, Name of Creator, Dates of Creation, Extent of the Unit of Description and Level of description. Other metadata elements were also identified to record context, provenance, and access points. Because of this work, the library saw noticeable increase in the interest of its archival collections both within and outside the university.
This presentation includes a step-by-step explanation of how the Archives Hub Editor was used to describe the works of Philip Mackie and the works of Ken Russell which includes scripts, photographs, scripts, correspondence, and other documents.
Williams (2006) “Arranging and describing archives is quite different from classifying and cataloguing books. Books are single items that do not depend on other items for their meaning. They are catalogued as discrete units, often by subject. However, each archival collection is an aggregate: it comprises a sequence of interrelated documents. It has a collective significance, and significance is lost if documents are treated as single items. Much of the meaning of archives is derived from their context: from an understanding of the origin, functions, administrative or business process that generated the information they contain. Because each collection of archives is unique, a unique arrangement will need to be developed for each. Classification schemes do not apply to archives in the same way as to books: and any standardising systems that are developed (e.g. for the arrangement of business archives or records of non-conformist churches) must remain flexible. Archives should never be arranged by subject unless this coincides with their contextual or functional origins.
At the highest level, all arrangement is based on the principle of provenance."
Miller (2017)
“Respect des fonds
Archivists combine the principles of provenance and original order into the overarching principle of respect des fonds. Defined fully, respect des fonds is
an overarching concept, of which provenance and original order are parts, that means that in order to protect the integrity of archives, all archives from one particular creator or source (provenance) must be kept together as a unified whole, not separated into artificial groups or intermingled with archives from another source, and that all archives within that unified whole should be preserved in the order in which they were made and used (original order). ”
Williams (2006) states states that “The principle of original order requires that records should be maintained in the same order as they were in while in active use: a kind of internal provenance. It states that even if the original order appears to be chaotic it should still be retained as evidence of how the individual or organisation operated and is part of its provenance.”
Ensure the creation of accurate, complete and consistent descriptions.
Identifying archival content, description, access, preservation, search, retrieval, legal compliance, curation, research, interoperability, accessibility
Aid in retrieval and data sharing about archival content.
Facilitate the sharing of authority data.
Allow integration of varied descriptions into a cohesive information system.
The Society of American Archivists (SAA) defines a finding aid as “a single document that places the materials in context by consolidating information about the collection, such as acquisition and processing; provenance, including administrative history or biographical note; scope of the collection, including size, subjects, media; organization and arrangement; and an inventory of the series and the folders.”
Enhances findability, discoverability, preservation.
Validates records' integrity, authenticity, and regulatory adherence (ISO 23081-1:2006).
Supports records' lifecycle from creation to disposal (ICA, 2016).
Findability: It's what you enter into a search engine (Zeng & Qin, 2016).
Discoverability: Metadata acts as a language for finding and rediscovering.
Preservation: Integral to maintaining records' authenticity and integrity over time (ISO 23081-1: 2017-10).
ISAD(G) Guidelines: Metadata fields outlined by International Council on Archives, 2000.
Archival History & Requisites: Metadata captures history, access conditions, and technical needs (ISO 15489-1:2006).
Lifecycle: Stands at the core of record evolution and provenance (ISO 23081-2:2009).
“Finding aids capture information and metadata about the provenance and original order of the records; contexts which are crucial to archival users in understanding the significance of the material.” (Francisco-Revilla et al., 2014)
ISAD (G): https://repositories.lib.utexas.edu/bitstream/handle/2152/14140/ISAD%28G%29.pdf?sequence=2&isAllowed=y
Subjects / access points - supports subject and name authorities (LCSH, LCNAF, VIAF, The UNESCO Thesaurus, The UK Archival Thesaurus (UKAT), ISAAR (CPF): International Standard Archival Authority Record for Corporate Bodies, Persons and Families)
ISAD (G): https://repositories.lib.utexas.edu/bitstream/handle/2152/14140/ISAD%28G%29.pdf?sequence=2&isAllowed=y
Downloaded XML doesn't auto-update Primo.
Minimal effort for Archive Hub edits:
Download XML.
Run the script.
Load MARC file via Alma import profile.
Infrequent updates to Archives Hub make this approach viable
https://archiveshub.jisc.ac.uk/data/gb3600-mac
Gracy (2014): “Archival description is often rich and densely layered. It serves three primary purposes: as an instrument for documenting record provenance, as an aid to information retrieval, and as a tool for management of physical and electronic collections.”