The document outlines recommendations to improve the search experience on a library website. It discusses analyzing log data from the library search engine and EZProxy to help enhance the user experience through a new recommender system. The project will collect and analyze activity data to provide personalized recommendations and improve search capabilities. Privacy and ethics are also addressed.
The document discusses recommendations to improve the search experience on the Open University's library search system. It outlines evaluating activity data from the library search logs to power a recommender system that would provide personalized recommendations to students. It describes the scope of collecting this activity data, challenges with data quality, and initial positive feedback from student surveys and interviews on recommendations related to their course of study.
PhD Viva - Disambiguating Identity Web References using Social DataMatthew Rowe
This document summarizes Matthew Rowe's PhD thesis on automatically disambiguating identity web references. The thesis claims that automated techniques can replace humans in performing disambiguation at scale and with high accuracy by leveraging seed data from social web platforms. It outlines three disambiguation techniques evaluated in the thesis: 1) inference rules, 2) random walks on graphs, and 3) self-training classifiers. Evaluation results show the self-training approach achieves recall comparable to humans while the rule-based approach has the highest precision. The document also discusses generating metadata models from web resources and requirements for seed data and disambiguation techniques.
This document provides an overview of research basics and library resources from a presentation by Laksamee Putnam, a research librarian. It discusses developing keywords for search topics, evaluating online information sources, searching the library catalog and databases for books and articles, and asking librarians for research help. Evaluation criteria like currency, reliability, authority and purpose are presented for assessing information quality. Boolean operators and truncation are also covered for effective searching.
The Buzz About BIBFRAME, by Angela KroegerAngela Kroeger
Overview of the Library of Congress's BIBFRAME initiative, including an overview of the BIBFRAME model and core classes, the necessity of replacing MARC, BIBFRAME alignment with RDA and FRBR, and the kinds of bibliographic functionality BIBFRAME linked data might enable in the future. Presentation by Angela Kroeger of the Criss Library at the University of Nebraska at Omaha, presented at the Amigos Online Conference, "Is RDA on Your RaDAr?" February 20, 2014. Full presenter notes/script and bibliography available upon request. Contact angelajkroeger [at] gmail [dot] com.
VIVO is an open-source semantic web application and information model that enables discovery of research across disciplines at institutions. It harvests data from verified sources to create detailed profiles of faculty and researchers. The structured linked data in VIVO allows for relationships and connections between researchers, publications, grants, and more to be visualized. Libraries can play important roles in implementing and supporting VIVO through activities like outreach, training, ontology development, and technical support.
Slides for a presentation with Sarah Laleman and Margaret Bausman at the 2014 CUNY assessment conference: REINVENTING LIBRARIES, REINVENTING ASSESSMENT. Surveys of faculty satisfaction with library resources and services are a common benchmark of library performance, but what does satisfaction really mean and is it really all that anyone wants out of their library?
This presentation was given by Arta Kabashi of AMIGOS for Day 2 of the Small Library Management Training Program's IV: Reference workshop in Spring of 2010. This presentation covers open source tools for the ref desk, ready reference sources and introduces the TexShare databases.
This document summarizes a library instruction session on evaluating online information sources. It provides the instructor's contact information and an agenda that includes discussing an article on evaluating online information, background reading and searching strategies, developing keywords, and evaluating resources. It then covers developing effective search strategies using keywords, Boolean operators, truncation and phrase searching. It discusses evaluating websites using the CRAP test to assess currency, reliability, authority and purpose/point of view. Students are instructed to analyze example websites and fill out a worksheet to begin their own research.
The document discusses recommendations to improve the search experience on the Open University's library search system. It outlines evaluating activity data from the library search logs to power a recommender system that would provide personalized recommendations to students. It describes the scope of collecting this activity data, challenges with data quality, and initial positive feedback from student surveys and interviews on recommendations related to their course of study.
PhD Viva - Disambiguating Identity Web References using Social DataMatthew Rowe
This document summarizes Matthew Rowe's PhD thesis on automatically disambiguating identity web references. The thesis claims that automated techniques can replace humans in performing disambiguation at scale and with high accuracy by leveraging seed data from social web platforms. It outlines three disambiguation techniques evaluated in the thesis: 1) inference rules, 2) random walks on graphs, and 3) self-training classifiers. Evaluation results show the self-training approach achieves recall comparable to humans while the rule-based approach has the highest precision. The document also discusses generating metadata models from web resources and requirements for seed data and disambiguation techniques.
This document provides an overview of research basics and library resources from a presentation by Laksamee Putnam, a research librarian. It discusses developing keywords for search topics, evaluating online information sources, searching the library catalog and databases for books and articles, and asking librarians for research help. Evaluation criteria like currency, reliability, authority and purpose are presented for assessing information quality. Boolean operators and truncation are also covered for effective searching.
The Buzz About BIBFRAME, by Angela KroegerAngela Kroeger
Overview of the Library of Congress's BIBFRAME initiative, including an overview of the BIBFRAME model and core classes, the necessity of replacing MARC, BIBFRAME alignment with RDA and FRBR, and the kinds of bibliographic functionality BIBFRAME linked data might enable in the future. Presentation by Angela Kroeger of the Criss Library at the University of Nebraska at Omaha, presented at the Amigos Online Conference, "Is RDA on Your RaDAr?" February 20, 2014. Full presenter notes/script and bibliography available upon request. Contact angelajkroeger [at] gmail [dot] com.
VIVO is an open-source semantic web application and information model that enables discovery of research across disciplines at institutions. It harvests data from verified sources to create detailed profiles of faculty and researchers. The structured linked data in VIVO allows for relationships and connections between researchers, publications, grants, and more to be visualized. Libraries can play important roles in implementing and supporting VIVO through activities like outreach, training, ontology development, and technical support.
Slides for a presentation with Sarah Laleman and Margaret Bausman at the 2014 CUNY assessment conference: REINVENTING LIBRARIES, REINVENTING ASSESSMENT. Surveys of faculty satisfaction with library resources and services are a common benchmark of library performance, but what does satisfaction really mean and is it really all that anyone wants out of their library?
This presentation was given by Arta Kabashi of AMIGOS for Day 2 of the Small Library Management Training Program's IV: Reference workshop in Spring of 2010. This presentation covers open source tools for the ref desk, ready reference sources and introduces the TexShare databases.
This document summarizes a library instruction session on evaluating online information sources. It provides the instructor's contact information and an agenda that includes discussing an article on evaluating online information, background reading and searching strategies, developing keywords, and evaluating resources. It then covers developing effective search strategies using keywords, Boolean operators, truncation and phrase searching. It discusses evaluating websites using the CRAP test to assess currency, reliability, authority and purpose/point of view. Students are instructed to analyze example websites and fill out a worksheet to begin their own research.
This document summarizes a library instruction session on evaluating online resources. It introduces the instructor, Laksamee Putnam, and provides her contact information. The agenda includes discussing an article on evaluating online information, background reading and searching strategies, developing keywords, and evaluating resources. Tips are provided on developing search strategies using keywords, Boolean operators, truncation and phrase searching. Students are guided through evaluating websites using the CRAP test to assess currency, reliability, authority and purpose/point of view. Next steps include completing a worksheet and preparing for discussion by finding additional resources.
"Leaders and Laggards in the preservation of raw biomedical research data" presented at NEDCC 2010, The Tectonics of Digital Curation
A Symposium on the Shifting Preservation and Access Landscape
This document provides an overview of online visualization tools and resources for visualizing concepts, data, and other information. It discusses tools for visual searches, word clouds, mind mapping, timelines, data visualization, and mashups. Examples of specific tools are given for each category. The document also lists additional resources for visualization trends, techniques, and staying up to date on new developments.
Facilitating Open Science and Research Discovery via VIVO and the Semantic WebKristi Holmes
Kristi Holmes discusses how VIVO, an open-source semantic web application, enables the discovery of research and scholarship across disciplines by creating profiles for faculty and researchers that display their publications, teaching, service, and affiliations. VIVO harvests data from internal sources like HR directories and external sources like PubMed to populate these profiles. It stores this information as linked data using semantic web standards, allowing the data to be integrated into the larger Linked Open Data cloud and consumed by other applications and systems both within and outside an institution.
Scholarly Identity 2.0: What does the Web say about your research?Michael Habib
Congress Center Hotel Zira
Belgrade, Serbia – October 30, 2009
Hosted by University of Belgrade...
Blog post describing presentation and proposed concept model:
http://mchabib.com/2009/11/04/scholarly-identity-2-0-matrix-concept-model-and-presentation/
A video of the presentation is located here:
http://bit.ly/6VpsbX
This document summarizes a library instruction session on researching for academic papers. It provides an overview of the research process and discusses evaluating online sources, developing effective keywords and search strings, finding books and scholarly articles, and citing sources. Students are guided through hands-on exercises to practice researching their topics and evaluating websites. Contact information is also provided for the librarian presenting and for general research help.
User-centered research for developing programs & articulating value.Lynn Connaway
Connaway, L. S. (2019). User-centered research for developing programs & articulating value. Presented at the University of Adelaide, February 18, 2019, Adelaide, Australia.
From Academic Library 2.0 to (Literature) Research 2.0Michael Habib
The document discusses the transition from Library 1.0 to Library 2.0 and Research 2.0, which are influenced by concepts of Web 2.0 like user-generated content, long tails, and collective intelligence. It provides examples of how libraries can apply these concepts through social networking, bookmarking, citation analysis tools, and APIs to engage users and meet their evolving needs. The document also shares results from a survey that found researchers are increasingly using social media and see it becoming more influential in their work in the next 5 years.
Scholarly Reputation Management Online: The Challenges and Opportunities of ...Michael Habib
Session 6: Wissenschaftskommunikation 2.0 – Social Software @ WorkSchloss Mickeln, Düsseldorf, 29. September 2009 Abstract: Social media provides scholars with unprecedented opportunities to promote their accomplishments and expertise. Conversely, social media creates more identity information to for scholars to manage. Different facets of scholar identity online will be introduced. Within this framework, new types of identity content produced by social software and the challenges this creates will be discussed. Lastly, opportunities for using social software to manage scholarly reputation will be explored.
The document discusses the relationship between web science and artificial intelligence (AI). It makes three key points:
1) AI can be used for the web, such as using connectionist AI for clustering, recommending, classifying and indexing web data, and symbolic AI for structured querying, reasoning and interoperating.
2) The web can be used for AI, such as providing large datasets and corpus for connectionist AI training, and standards for symbolic AI knowledge representation.
3) There is an intersection between web science and AI, as the web and AI influence each other, which is an important topic for web science.
Wiser Pku Lecture@Life Science School Pkuguest8ed46d
The document discusses various online tools for effective literature management and reference searching. It introduces popular tools like Mendeley, EndNote and Zotero for building local reference databases and sharing references online. Social bookmarking and networking sites like Diigo, SlideShare and Wikipedia are described as useful resources for searching references in a social way through tags and user connections.
This document discusses semantic web technologies including linked data and how they can help solve information needs on the current web. It provides an overview of semantic web standards like RDF and SPARQL and examples of datasets like DBPedia that publish structured data on the web in RDF format and link to each other. It also demonstrates tools for extracting structured data from websites and a mobile app built using open linked data.
The document discusses how libraries need to adapt to changing user behaviors and technologies. It outlines trends like social media, mobile access, cloud computing, and virtual research environments. The library needs to collaborate more closely with researchers, be involved in digital research outputs, and redefine its services and systems to improve visibility, findability, searchability and accessibility on the web. Libraries should support learning and research through communication, cooperation and facilitating the creation, management and sharing of information.
Making the Web Searchable - Keynote ICWE 2015Peter Mika
This document discusses making the web more searchable through semantic technologies. It begins with an overview of how web search currently works and its limitations, and then discusses how the semantic web aims to address these issues by adding explicit meaning and relationships between data on the web. It describes early skepticism of the semantic web from the information retrieval community and how it has become more practical over time. It also outlines research into semantic search done at Yahoo, including developing a knowledge graph and using semantic information to enhance search results. Finally, it discusses how semantic technologies are now being adopted more widely through efforts like schema.org.
Advanced internet searching to find OER is a resource which will assist students to find useful resources which can be reused, remixed and redistributed under the Creative Commons banner. Nicole Southgate is an OER practitioner and Technical Assistant at the Education Development Unit at the Faculty of Health Sciences, University of Cape Town.
This document discusses data citation and using identifiers to cite datasets. It explains that identifiers provide exposure, transparency, citation tracking and verification for datasets. Identifiers associate an alphanumeric string with the location of an object, like a dataset, and can include optional metadata. Common identifier systems like DOIs provide a precise way to identify and cite datasets. Services like EZID make it easy to create and manage identifiers for datasets. The document encourages attendees to get started with data citation by creating test identifiers and discussing options with librarians.
Adaptability, aboutness, and authenticity: Towards discovery platforms with n...blisspix
Presentation for Internet Librarian International 2009. Full paper available to participants, will be uploaded to an open access repository following the conference.
Realizing the Benefits for Scholars and the Digital Library Community Through...guest6a8e0d
The document summarizes the DLF Aquifer initiative's American Social History Online project. It discusses the project's goals of aggregating digital collections to make them easier for scholars to access as a single collection. It provides an overview of the project accomplishments to date, as well as next steps, and introduces the core team members and technologies used.
Seminar 1 from ReachOut to Research (R2R)
Small seminar about library services supporting research & technology
Reachout to Research : library support services.
See also Seminar 2: https://www.slideshare.net/digicmb/the-user-the-technology-the-library-and-why-to-go-in-between
http://lanyrd.com/2013/r2ruit/
This document summarizes a library instruction session on evaluating online resources. It introduces the instructor, Laksamee Putnam, and provides her contact information. The agenda includes discussing an article on evaluating online information, background reading and searching strategies, developing keywords, and evaluating resources. Tips are provided on developing search strategies using keywords, Boolean operators, truncation and phrase searching. Students are guided through evaluating websites using the CRAP test to assess currency, reliability, authority and purpose/point of view. Next steps include completing a worksheet and preparing for discussion by finding additional resources.
"Leaders and Laggards in the preservation of raw biomedical research data" presented at NEDCC 2010, The Tectonics of Digital Curation
A Symposium on the Shifting Preservation and Access Landscape
This document provides an overview of online visualization tools and resources for visualizing concepts, data, and other information. It discusses tools for visual searches, word clouds, mind mapping, timelines, data visualization, and mashups. Examples of specific tools are given for each category. The document also lists additional resources for visualization trends, techniques, and staying up to date on new developments.
Facilitating Open Science and Research Discovery via VIVO and the Semantic WebKristi Holmes
Kristi Holmes discusses how VIVO, an open-source semantic web application, enables the discovery of research and scholarship across disciplines by creating profiles for faculty and researchers that display their publications, teaching, service, and affiliations. VIVO harvests data from internal sources like HR directories and external sources like PubMed to populate these profiles. It stores this information as linked data using semantic web standards, allowing the data to be integrated into the larger Linked Open Data cloud and consumed by other applications and systems both within and outside an institution.
Scholarly Identity 2.0: What does the Web say about your research?Michael Habib
Congress Center Hotel Zira
Belgrade, Serbia – October 30, 2009
Hosted by University of Belgrade...
Blog post describing presentation and proposed concept model:
http://mchabib.com/2009/11/04/scholarly-identity-2-0-matrix-concept-model-and-presentation/
A video of the presentation is located here:
http://bit.ly/6VpsbX
This document summarizes a library instruction session on researching for academic papers. It provides an overview of the research process and discusses evaluating online sources, developing effective keywords and search strings, finding books and scholarly articles, and citing sources. Students are guided through hands-on exercises to practice researching their topics and evaluating websites. Contact information is also provided for the librarian presenting and for general research help.
User-centered research for developing programs & articulating value.Lynn Connaway
Connaway, L. S. (2019). User-centered research for developing programs & articulating value. Presented at the University of Adelaide, February 18, 2019, Adelaide, Australia.
From Academic Library 2.0 to (Literature) Research 2.0Michael Habib
The document discusses the transition from Library 1.0 to Library 2.0 and Research 2.0, which are influenced by concepts of Web 2.0 like user-generated content, long tails, and collective intelligence. It provides examples of how libraries can apply these concepts through social networking, bookmarking, citation analysis tools, and APIs to engage users and meet their evolving needs. The document also shares results from a survey that found researchers are increasingly using social media and see it becoming more influential in their work in the next 5 years.
Scholarly Reputation Management Online: The Challenges and Opportunities of ...Michael Habib
Session 6: Wissenschaftskommunikation 2.0 – Social Software @ WorkSchloss Mickeln, Düsseldorf, 29. September 2009 Abstract: Social media provides scholars with unprecedented opportunities to promote their accomplishments and expertise. Conversely, social media creates more identity information to for scholars to manage. Different facets of scholar identity online will be introduced. Within this framework, new types of identity content produced by social software and the challenges this creates will be discussed. Lastly, opportunities for using social software to manage scholarly reputation will be explored.
The document discusses the relationship between web science and artificial intelligence (AI). It makes three key points:
1) AI can be used for the web, such as using connectionist AI for clustering, recommending, classifying and indexing web data, and symbolic AI for structured querying, reasoning and interoperating.
2) The web can be used for AI, such as providing large datasets and corpus for connectionist AI training, and standards for symbolic AI knowledge representation.
3) There is an intersection between web science and AI, as the web and AI influence each other, which is an important topic for web science.
Wiser Pku Lecture@Life Science School Pkuguest8ed46d
The document discusses various online tools for effective literature management and reference searching. It introduces popular tools like Mendeley, EndNote and Zotero for building local reference databases and sharing references online. Social bookmarking and networking sites like Diigo, SlideShare and Wikipedia are described as useful resources for searching references in a social way through tags and user connections.
This document discusses semantic web technologies including linked data and how they can help solve information needs on the current web. It provides an overview of semantic web standards like RDF and SPARQL and examples of datasets like DBPedia that publish structured data on the web in RDF format and link to each other. It also demonstrates tools for extracting structured data from websites and a mobile app built using open linked data.
The document discusses how libraries need to adapt to changing user behaviors and technologies. It outlines trends like social media, mobile access, cloud computing, and virtual research environments. The library needs to collaborate more closely with researchers, be involved in digital research outputs, and redefine its services and systems to improve visibility, findability, searchability and accessibility on the web. Libraries should support learning and research through communication, cooperation and facilitating the creation, management and sharing of information.
Making the Web Searchable - Keynote ICWE 2015Peter Mika
This document discusses making the web more searchable through semantic technologies. It begins with an overview of how web search currently works and its limitations, and then discusses how the semantic web aims to address these issues by adding explicit meaning and relationships between data on the web. It describes early skepticism of the semantic web from the information retrieval community and how it has become more practical over time. It also outlines research into semantic search done at Yahoo, including developing a knowledge graph and using semantic information to enhance search results. Finally, it discusses how semantic technologies are now being adopted more widely through efforts like schema.org.
Advanced internet searching to find OER is a resource which will assist students to find useful resources which can be reused, remixed and redistributed under the Creative Commons banner. Nicole Southgate is an OER practitioner and Technical Assistant at the Education Development Unit at the Faculty of Health Sciences, University of Cape Town.
This document discusses data citation and using identifiers to cite datasets. It explains that identifiers provide exposure, transparency, citation tracking and verification for datasets. Identifiers associate an alphanumeric string with the location of an object, like a dataset, and can include optional metadata. Common identifier systems like DOIs provide a precise way to identify and cite datasets. Services like EZID make it easy to create and manage identifiers for datasets. The document encourages attendees to get started with data citation by creating test identifiers and discussing options with librarians.
Adaptability, aboutness, and authenticity: Towards discovery platforms with n...blisspix
Presentation for Internet Librarian International 2009. Full paper available to participants, will be uploaded to an open access repository following the conference.
Realizing the Benefits for Scholars and the Digital Library Community Through...guest6a8e0d
The document summarizes the DLF Aquifer initiative's American Social History Online project. It discusses the project's goals of aggregating digital collections to make them easier for scholars to access as a single collection. It provides an overview of the project accomplishments to date, as well as next steps, and introduces the core team members and technologies used.
Seminar 1 from ReachOut to Research (R2R)
Small seminar about library services supporting research & technology
Reachout to Research : library support services.
See also Seminar 2: https://www.slideshare.net/digicmb/the-user-the-technology-the-library-and-why-to-go-in-between
http://lanyrd.com/2013/r2ruit/
This document discusses Research Objects (RO), which provide a framework for bundling, exchanging, and linking resources related to experiments in order to improve reproducibility. The RO framework uses unique identifiers, aggregation, and metadata to group related resources. Real-world examples of ROs include reviewed scientific papers, workflow runs, and Docker images. ROs can help make research fully FAIR (Findable, Accessible, Interoperable, Reusable). Tools and platforms like FAIRDOM, SEEK, and Figshare support the use of ROs.
"Data in Context" IG sessions @ RDA 3rd PlenaryBrigitte Jörg
The Data in Context Interest Group at the 3rd RDA Plenary in Dublin discussed developing standards and requirements for representing data context through the data lifecycle. They reviewed several existing data lifecycle and metadata models, as well as relevant standards organizations. Their work plan involves creating an overview of context-aware standards by month 6 and a prioritized requirements list by month 12. The long-term goal is to establish a Working Group to implement standardized profiles and enable automated transformation between standards to represent data context.
Data in Context Interest Group Sessions @ RDA 3rd Plenary, Dublin (March 26-2...Brigitte Jörg
The Data in Context Interest Group at the 3rd RDA Plenary in Dublin discussed developing a common understanding of data context and lifecycles. They reviewed several existing data lifecycle models and standards that address contextual metadata. Their goals are to provide an overview of relevant standardization work, prioritize requirements, and establish a working group to further develop standardized profiles and facilitate transformation between standards to represent data context. The group plans initial deliverables reviewing contextual standards work and prioritizing requirements, to inform establishing a working group.
Rise presentation for jisc online mtg 2011 06-02Richard Nurse
The document discusses collecting activity data from online resources like EZProxy, CIRCE, and RISE to provide recommendations to users about courses, searches, and relationships. It asks questions about what resources and attributes the data involves, how the data is extracted, what insights the data can provide, whether aggregating this data with other activity data would be useful, and addresses any potential legal or privacy issues.
NSF Workshop Data and Software Citation, 6-7 June 2016, Boston USA, Software Panel
FIndable, Accessible, Interoperable, Reusable Software and Data Citation: Europe, Research Objects, and BioSchemas.org
VIVO is a semantic web application that enables discovery of research across disciplines in an institution. It allows granular editing of profiles while also ingesting data automatically from sources like HR systems. The presenter discussed VIVO's history and architecture, how it exposes linked open data through SPARQL queries and RDF views. Visualizations like co-authorship networks and implementations at various universities were also covered.
The document provides an overview of the International Image Interoperability Framework (IIIF). It discusses the IIIF image API, presentation API, search API, authentication API, and annotations. It demonstrates examples of manifests, sequences, and using IIIF with annotations. The document also discusses future APIs, additional resources for learning about IIIF, and encourages participation in the IIIF community.
ALIAOnline Practical Linked (Open) Data for Libraries, Archives & MuseumsJon Voss
This document discusses practical applications of Linked Open Data (LOD) for libraries, archives, and museums. It describes how LOD allows these institutions to publish structured data on the web in ways that are interoperable and can be connected to other open datasets. Examples are given of how LOD is being used by various institutions to share metadata, images, and other cultural heritage assets on the web in open, machine-readable formats. The presenter argues that LOD represents a new paradigm that these cultural organizations should embrace to make their collections more accessible and useful on the web.
Library discovery: past, present and some futureslisld
A presentation at the NISO virtual conference on Webscale Discovery Services, 20 November 2013.
Considers some of the issues that have led to the adoption of these services, and some future directions.
Distinguishes between discovery (providing a library destination) and discoverability (making stuff discoverable elsewhere).
The Progress of BIBFRAME, by Angela KroegerAngela Kroeger
Presentation given at the OLAC-MOUG 2014 conference. Abstract: BIBFRAME is the Library of Congress's current effort to develop a linked data replacement for MARC. BIBFRAME is a work in progress, not yet ready for implementation. In this two-hour session, we will examine how BIBFRAME works, what it is intended to accomplish, and the progress that has been made toward that goal. We'll take a look at the BIBFRAME tools that are under development, including the prototype editor for creating new records. And we'll share a glimpse of what the future holds for library catalogs and cataloging. NOTE: SlideShare seems to have garbled the formatting of some of my slides. To receive a clean copy via email, contact me at angelajkroeger [at] gmail [dot] com.
Talk given at Open Knowledge Foundation 'Opening Up Metadata: Challenges, Standards and Tools' Workshop, Queen Mary University of London, 13th June 2012.
Info on the event at http://openglam.org/2012/05/31/last-places-left-for-opening-up-metadata-challenges-standards-and-tools/
Linked Data Love: research representation, discovery, and assessment
#ALAAC15
The explosion of linked data platforms and data stores over the last five years has been profound – both in terms of quantity of data as well as its potential impact. Research information systems such as VIVO (www.vivoweb.org) play a significant role in enabling this work. VIVO is an open source, Semantic Web-based application that provides an integrated, searchable view of the scholarly activities of an organization. The uniform semantic structure of VIVO-ISF data enables a new class of tools to advance science. This presentation will provide a brief introduction and update to VIVO and present ways that this semantically-rich data can enable visualizations, reporting and assessment, next-generation collaboration and team building, and enhanced multi-site search. Libraries are uniquely positioned to facilitate the open representation of research information and its subsequent use to spur collaboration, discovery, and assessment. The talk will conclude with a description of ways librarians are engaged in this work – including visioning, metadata and ontology creation, policy creation, data curation and management, technical, and engagement activities.
Kristi Holmes, PhD
Director, Galter Health Sciences Library
Director of Evaluation, NUCATS
Associate Professor, Preventive Medicine-Health and Biomedical Informatics
Northwestern University Feinberg School of Medicine
Mobile Strategies for Libraries by Amy Yorkayork1120
This document discusses strategies for libraries to provide mobile access to their resources. It recommends having a mobile-optimized website, responsive design, or mobile applications. Native applications provide benefits like offline access and device capabilities but have higher development costs. The document explores options for mobile sites like responsive frameworks and redirects. It also provides examples of libraries that implemented mobile strategies successfully and discusses promoting mobile services through QR codes, SMS, and research applications.
Towards Cognitive Agents for BigData DiscoveryJack Park
1) The document discusses developing cognitive agents to improve deep question answering and discovery by coupling two platforms: the Berkeley Data Analytics Stack (BDAS) for big data analysis and SolrSherlock for literature-based discovery.
2) It describes how these agents could harvest and represent patterns, contexts, and relations from literature to discover new processes and connections between concepts.
3) The goal is to augment existing methods by allowing hypothesis formation and evidence gathering across both structured data and unstructured literature at scale.
Sherif presentation richard nurse new templateRichard Nurse
1) The study analyzed usage data from eResources at the Open University, a distance learning institution, to determine if there was a relationship between library resource usage and student success, as has been found at traditional universities.
2) The results showed that students who passed modules accessed library eResources twice as often as students who failed. Further statistical analysis found a significant association between higher levels of eResource access and better student results.
3) The findings suggest that even in a digital library environment, higher student engagement with library resources is still correlated with academic success, as seen in other university studies.
Rise presentation for programme meeting 2011 07-05Richard Nurse
This document summarizes a presentation on improving search experiences through recommendations. It discusses collecting user data on searches, resources viewed, and user ratings to generate course, search, and relationship recommendations. A recommender database was created from databases like EZProxy, CIRCE, and Crossref. Student surveys found that recommendations were useful and improved search results. Analyzing user data can provide recommendations to help users discover additional relevant resources.
Rise presentation for jisc online mtg 2011 06-02Richard Nurse
The document discusses an online event about using activity data from online resources to improve search experiences. It lists various online resources like EZProxy, CIRCE, and RISE that collect usage data and the attributes available from each resource. The document also questions how the data could be used, such as seeing which courses students accessing certain resources are studying, and whether the data can be aggregated or has any legal or privacy implications.
The document discusses gaining business intelligence from user activity data in libraries and higher education institutions. It outlines challenges in collecting and analyzing comprehensive user data from different systems. The Open University perspective is that most students do not visit physical libraries and sign up for individual courses rather than degrees. However, the university has significant online user traffic that could provide insights if integrated across various learning and library systems. Overcoming cultural, technical, and data challenges will be key to developing a comprehensive view of user activity data.
Generative AI Deep Dive: Advancing from Proof of Concept to ProductionAggregage
Join Maher Hanafi, VP of Engineering at Betterworks, in this new session where he'll share a practical framework to transform Gen AI prototypes into impactful products! He'll delve into the complexities of data collection and management, model selection and optimization, and ensuring security, scalability, and responsible use.
Full-RAG: A modern architecture for hyper-personalizationZilliz
Mike Del Balso, CEO & Co-Founder at Tecton, presents "Full RAG," a novel approach to AI recommendation systems, aiming to push beyond the limitations of traditional models through a deep integration of contextual insights and real-time data, leveraging the Retrieval-Augmented Generation architecture. This talk will outline Full RAG's potential to significantly enhance personalization, address engineering challenges such as data management and model training, and introduce data enrichment with reranking as a key solution. Attendees will gain crucial insights into the importance of hyperpersonalization in AI, the capabilities of Full RAG for advanced personalization, and strategies for managing complex data integrations for deploying cutting-edge AI solutions.
GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...Neo4j
Leonard Jayamohan, Partner & Generative AI Lead, Deloitte
This keynote will reveal how Deloitte leverages Neo4j’s graph power for groundbreaking digital twin solutions, achieving a staggering 100x performance boost. Discover the essential role knowledge graphs play in successful generative AI implementations. Plus, get an exclusive look at an innovative Neo4j + Generative AI solution Deloitte is developing in-house.
TrustArc Webinar - 2024 Global Privacy SurveyTrustArc
How does your privacy program stack up against your peers? What challenges are privacy teams tackling and prioritizing in 2024?
In the fifth annual Global Privacy Benchmarks Survey, we asked over 1,800 global privacy professionals and business executives to share their perspectives on the current state of privacy inside and outside of their organizations. This year’s report focused on emerging areas of importance for privacy and compliance professionals, including considerations and implications of Artificial Intelligence (AI) technologies, building brand trust, and different approaches for achieving higher privacy competence scores.
See how organizational priorities and strategic approaches to data security and privacy are evolving around the globe.
This webinar will review:
- The top 10 privacy insights from the fifth annual Global Privacy Benchmarks Survey
- The top challenges for privacy leaders, practitioners, and organizations in 2024
- Key themes to consider in developing and maintaining your privacy program
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdfPaige Cruz
Monitoring and observability aren’t traditionally found in software curriculums and many of us cobble this knowledge together from whatever vendor or ecosystem we were first introduced to and whatever is a part of your current company’s observability stack.
While the dev and ops silo continues to crumble….many organizations still relegate monitoring & observability as the purview of ops, infra and SRE teams. This is a mistake - achieving a highly observable system requires collaboration up and down the stack.
I, a former op, would like to extend an invitation to all application developers to join the observability party will share these foundational concepts to build on:
Threats to mobile devices are more prevalent and increasing in scope and complexity. Users of mobile devices desire to take full advantage of the features
available on those devices, but many of the features provide convenience and capability but sacrifice security. This best practices guide outlines steps the users can take to better protect personal devices and information.
Securing your Kubernetes cluster_ a step-by-step guide to success !KatiaHIMEUR1
Today, after several years of existence, an extremely active community and an ultra-dynamic ecosystem, Kubernetes has established itself as the de facto standard in container orchestration. Thanks to a wide range of managed services, it has never been so easy to set up a ready-to-use Kubernetes cluster.
However, this ease of use means that the subject of security in Kubernetes is often left for later, or even neglected. This exposes companies to significant risks.
In this talk, I'll show you step-by-step how to secure your Kubernetes cluster for greater peace of mind and reliability.
UiPath Test Automation using UiPath Test Suite series, part 6DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 6. In this session, we will cover Test Automation with generative AI and Open AI.
UiPath Test Automation with generative AI and Open AI webinar offers an in-depth exploration of leveraging cutting-edge technologies for test automation within the UiPath platform. Attendees will delve into the integration of generative AI, a test automation solution, with Open AI advanced natural language processing capabilities.
Throughout the session, participants will discover how this synergy empowers testers to automate repetitive tasks, enhance testing accuracy, and expedite the software testing life cycle. Topics covered include the seamless integration process, practical use cases, and the benefits of harnessing AI-driven automation for UiPath testing initiatives. By attending this webinar, testers, and automation professionals can gain valuable insights into harnessing the power of AI to optimize their test automation workflows within the UiPath ecosystem, ultimately driving efficiency and quality in software development processes.
What will you get from this session?
1. Insights into integrating generative AI.
2. Understanding how this integration enhances test automation within the UiPath platform
3. Practical demonstrations
4. Exploration of real-world use cases illustrating the benefits of AI-driven test automation for UiPath
Topics covered:
What is generative AI
Test Automation with generative AI and Open AI.
UiPath integration with generative AI
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Alt. GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using ...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
Building RAG with self-deployed Milvus vector database and Snowpark Container...Zilliz
This talk will give hands-on advice on building RAG applications with an open-source Milvus database deployed as a docker container. We will also introduce the integration of Milvus with Snowpark Container Services.
“An Outlook of the Ongoing and Future Relationship between Blockchain Technologies and Process-aware Information Systems.” Invited talk at the joint workshop on Blockchain for Information Systems (BC4IS) and Blockchain for Trusted Data Sharing (B4TDS), co-located with with the 36th International Conference on Advanced Information Systems Engineering (CAiSE), 3 June 2024, Limassol, Cyprus.
How to Get CNIC Information System with Paksim Ga.pptxdanishmna97
Pakdata Cf is a groundbreaking system designed to streamline and facilitate access to CNIC information. This innovative platform leverages advanced technology to provide users with efficient and secure access to their CNIC details.
In the rapidly evolving landscape of technologies, XML continues to play a vital role in structuring, storing, and transporting data across diverse systems. The recent advancements in artificial intelligence (AI) present new methodologies for enhancing XML development workflows, introducing efficiency, automation, and intelligent capabilities. This presentation will outline the scope and perspective of utilizing AI in XML development. The potential benefits and the possible pitfalls will be highlighted, providing a balanced view of the subject.
We will explore the capabilities of AI in understanding XML markup languages and autonomously creating structured XML content. Additionally, we will examine the capacity of AI to enrich plain text with appropriate XML markup. Practical examples and methodological guidelines will be provided to elucidate how AI can be effectively prompted to interpret and generate accurate XML markup.
Further emphasis will be placed on the role of AI in developing XSLT, or schemas such as XSD and Schematron. We will address the techniques and strategies adopted to create prompts for generating code, explaining code, or refactoring the code, and the results achieved.
The discussion will extend to how AI can be used to transform XML content. In particular, the focus will be on the use of AI XPath extension functions in XSLT, Schematron, Schematron Quick Fixes, or for XML content refactoring.
The presentation aims to deliver a comprehensive overview of AI usage in XML development, providing attendees with the necessary knowledge to make informed decisions. Whether you’re at the early stages of adopting AI or considering integrating it in advanced XML development, this presentation will cover all levels of expertise.
By highlighting the potential advantages and challenges of integrating AI with XML development tools and languages, the presentation seeks to inspire thoughtful conversation around the future of XML development. We’ll not only delve into the technical aspects of AI-powered XML development but also discuss practical implications and possible future directions.
Pushing the limits of ePRTC: 100ns holdover for 100 daysAdtran
At WSTS 2024, Alon Stern explored the topic of parametric holdover and explained how recent research findings can be implemented in real-world PNT networks to achieve 100 nanoseconds of accuracy for up to 100 days.
20 Comprehensive Checklist of Designing and Developing a WebsitePixlogix Infotech
Dive into the world of Website Designing and Developing with Pixlogix! Looking to create a stunning online presence? Look no further! Our comprehensive checklist covers everything you need to know to craft a website that stands out. From user-friendly design to seamless functionality, we've got you covered. Don't miss out on this invaluable resource! Check out our checklist now at Pixlogix and start your journey towards a captivating online presence today.
UiPath Test Automation using UiPath Test Suite series, part 5DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 5. In this session, we will cover CI/CD with devops.
Topics covered:
CI/CD with in UiPath
End-to-end overview of CI/CD pipeline with Azure devops
Speaker:
Lyndsey Byblow, Test Suite Sales Engineer @ UiPath, Inc.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/06/building-and-scaling-ai-applications-with-the-nx-ai-manager-a-presentation-from-network-optix/
Robin van Emden, Senior Director of Data Science at Network Optix, presents the “Building and Scaling AI Applications with the Nx AI Manager,” tutorial at the May 2024 Embedded Vision Summit.
In this presentation, van Emden covers the basics of scaling edge AI solutions using the Nx tool kit. He emphasizes the process of developing AI models and deploying them globally. He also showcases the conversion of AI models and the creation of effective edge AI pipelines, with a focus on pre-processing, model conversion, selecting the appropriate inference engine for the target hardware and post-processing.
van Emden shows how Nx can simplify the developer’s life and facilitate a rapid transition from concept to production-ready applications.He provides valuable insights into developing scalable and efficient edge AI solutions, with a strong focus on practical implementation.
2. Why?
“The search
engine on the
library is not very
user friendly. I
had to find a
specific article
recommended in
the text and it “The search
took several facility is poor
attempts to and doesn’t
locate it.” find stuff that is
supposed to
http://www.flickr.com/photos/james_lumb/3921968993/sizes/z/in/photostream
be there”
3. New search system
New generation
Discovery
System from
EBSCO
http://www.flickr.com/photos/jiscimages/435135071/sizes/m/in/photostream/
4. Could we do more?
http://www.flickr.com/photos/davepattern/5808712333/sizes/z/in/photostream/
5. Recommendations Improve the
Search Experience?
“That recommender systems
can enhance the student
experience in new generation
e-resource discovery services”
6. Recommendations Improve the
Search Experience?
Can you use search data to
make recommendations?
Are recommendations useful
in Discovery systems?
http://www.flickr.com/photos/davepattern/3473326634/sizes/z/in/photostream/
7. JISC Activity Data Programme
JISC funded project
February – July 2011
One of eight projects [list at http://bit.ly/gwCmNS]
http://www.open.ac.uk/blogs/rise
8. Why activity data?
"Every day I wake up and ask,
'how can I flow data better,
manage data better, analyse
data better?"
Rollin Ford, the CIO of Wal-Mart
http://www.flickr.com/photos/zerimski/5215633183/sizes/z/in/photostream/
11. Library systems environment
Athens DA authentication built into local (SAMS) login system
EZProxy remote resource access
SFX knowledge base and OpenURL link resolver
Ebsco Discovery Solution
12. Scope of the project
Algorithms &
Activity data recommender Search
code interface
14. So what is in the EZProxy logs?
• Remote host
• Date/Time
• Oucu
• Request
• Status
• Size of response
• Referrer
• User agent
• Session
http://www.flickr.com/photos/vixon/116447718/sizes/m/in/photostream/
15. So what is in the EZProxy logs?
"0"|||"137.108.143.168"|||20110115235421|||“nn12
34"|||"GET http://libezproxy.open.ac.uk:80/connect?
Session=st3ShtizgtrS7tU5&url=
http://search.ebscohost.com/login.aspx?direct=true&
site=edslive&scope=site&type=0&cli0=FT&clv0=Y&c
li1=FT1&clv1=Y&authtype=ip&group=VCStud&bquer
y=War%20Against%20the%20Panthers
HTTP/1.1“|||302|||0|||http://library.open.ac.uk/
|||"Mozilla/5.0 (X11; U; Linux i686; en-US;
rv:1.9.2.13) Gecko/20101206 Ubuntu/10.10
(maverick) Firefox/3.6.13"|||"t3ShtizgtrS7tU5"
16. So what is in the EZProxy logs?
"0"|||"137.108.143.168"|||20110115235421|||“nn12
34"|||"GET http://libezproxy.open.ac.uk:80/connect?
date and time
Session=st3ShtizgtrS7tU5&url=
http://search.ebscohost.com/login.aspx?direct=true&
site=edslive&scope=site&type=0&cli0=FT&clv0=Y&c
li1=FT1&clv1=Y&authtype=ip&group=VCStud&bquer
y=War%20Against%20the%20Panthers
HTTP/1.1“|||302|||0|||http://library.open.ac.uk/
|||"Mozilla/5.0 (X11; U; Linux i686; en-US;
rv:1.9.2.13) Gecko/20101206 Ubuntu/10.10
(maverick) Firefox/3.6.13"|||"t3ShtizgtrS7tU5"
17. So what is in the EZProxy logs?
"0"|||"137.108.143.168"|||20110115235421|||“nn12
34"|||"GET http://libezproxy.open.ac.uk:80/connect?
User
Session=st3ShtizgtrS7tU5&url=
name
http://search.ebscohost.com/login.aspx?direct=true&
site=edslive&scope=site&type=0&cli0=FT&clv0=Y&c
li1=FT1&clv1=Y&authtype=ip&group=VCStud&bquer
y=War%20Against%20the%20Panthers
HTTP/1.1“|||302|||0|||http://library.open.ac.uk/
|||"Mozilla/5.0 (X11; U; Linux i686; en-US;
rv:1.9.2.13) Gecko/20101206 Ubuntu/10.10
(maverick) Firefox/3.6.13"|||"t3ShtizgtrS7tU5"
18. So what is in the EZProxy logs?
"0"|||"137.108.143.168"|||20110115235421|||“nn12
34"|||"GET http://libezproxy.open.ac.uk:80/connect?
Session=st3ShtizgtrS7tU5&url=
http://search.ebscohost.com/login.aspx?direct=true&
site=edslive&scope=site&type=0&cli0=FT&clv0=Y&c
li1=FT1&clv1=Y&authtype=ip&group=VCStud&bquer
y=War%20Against%20the%20Panthers
HTTP/1.1“|||302|||0|||http://library.open.ac.uk/
|||"Mozilla/5.0 (X11; Request
U; Linux i686; en-US;
rv:1.9.2.13) Gecko/20101206 Ubuntu/10.10
(maverick) Firefox/3.6.13"|||"t3ShtizgtrS7tU5"
22. What can the data tell us?
People who looked at resource ‘C’ also
People on course ‘A’ viewed resource ‘B’
looked at resource ‘D’
Which are the most popular resources
This resource is being used by people studying this course
23. But what isn’t there?
ISSNs DOI
Article Subject
information terms
http://www.flickr.com/photos/kevharb/5466661946/sizes/z/in/photostream/
24. So how do you improve your data?
Remote host | Date/Time | Oucu | request | status EZProxy
| size of response | referrer | user agent | session
user type | course code(s) CIRCE
EDS
Bibliographic data matching
Crossref
25. So what about collecting more
data?
http://library.open.ac.uk/rise
www.open.ac.uk/libraryservices/rise/
http://www.open.ac.uk/blogs/rise
30. So how do you improve your data?
Remote host | Date/Time | Oucu | request | status | size of EZProxy
response | referrer | user agent | session
user type | course code(s) CIRCE
EDS
Bibliographic data matching
Crossref
RISE Searches in RISE
32. What can the data tell us?
People on course ‘A’ viewed People who looked at resource People who searched for subject
resource ‘B’ ‘C’ also looked at resource ‘D’ ‘E’ looked at resource ‘F’
People are looking at resources on this subject
This resource is being used by people studying this course
34. Getting a recommendation
User A Views Resource B Views +1 Resource B
Module A123 RV=14 RV=15
User C Recommended Resource B Views +1 Resource B
Module A123 RV=15 RV=16
User C Rate Useful +1 Resource B
Module A123 RV=17
User C Rate Not Useful Resource B
Module A123 -2 RV=14
35. Data Protection and privacy
Added a privacy policy to RISE,
EDS and SFX interfaces
Provided an opt-out feature
Privacy and opt-out URL
http://library.open.ac.uk/rise/?p
age=privacy
If we go back to 2009, it became obvious that library search simply didn’t work as well as users expected it towe were getting the sort of comments you see on screen which showed that library users were struggling with the federated search system that we were usingSo the library embarked on some work to improve search, with a new discovery search system and other changes
we changed the search system to a new generation of library search system from EBSCO. Instead of searching library resources individually and telling you how many results are in each database it now searches one index and shows the results in a single list
We started thinking whether there was more that we could do to improve the user experience. For a while we’d been following with interest some JISC work looking at whether activity data could be used by libraries to improve services, in projects such as TILE and MOSAIC. So we started to think whether there was an opportunity to look at whether using activity data could improve the user experience of library search
So when we knew that JISC were going to be funding some more work on activity data, we thought about what we’d want to do, and came up with this hypothesis
The project we came up with was RISE – Recommendations Improve the Search ExperienceWe set out to test two thingsCan you use search data to make recommendationsAre recommendations useful for these new systems.
RISE was funded as part of the Activity Data strand of the JISC Infrastructure for Education and Research programmeIt was a very short project, just six months, with a small team – developer, project managerAnd there were seven other projects in the programme. Some of which were working with libraries such as SALT and LIDP, others of which are looking at activity data in a range of other areas from VLEs, through repositories, to student systems to video-conferencing data, and including the UCIAD project in Kmi looking at a user-centred approach to web clickstream data.
The business sector, particularly companies such as Tesco, Amazon and Wal-Mart exploit the data they have about customer activities to support decision making.Some early research by JISC, in the TILE and MOSAIC projects identified that the HE sector also had extensive user data and there was some potential to make use of it, but it was greatly underused. So this JISC programme has set out to explore this area in more detail. Across the sector we are being told to be more business-like and the use of customer data is one of the areas that businesses seem to be exploiting far more than we do
For a traditional ‘bricks and mortar’ university these are some of the ways that you’d typically interact with your customers.Well, for the OU things are a bit different
We don’t really loan many books to students or have many accessing the library. All our students are distance learners so they interact with us online and use our resources electronically. And with more than 450,000 unique users of our website and over 100,000 unique users of our e-resources each year then there’s a fair amount of activity data for us to use
So, if we are concentrating on our e-resources then the systems we use are SAMS single sign on. The EZProxy system from OCLC which allows students to access our resources as if they were locally within the library We are using SFX from ExLibris as our resources knowledge base and as the OpenURL link resolver and then finally the Ebsco Discovery Solution in place of an older federated search system
The stages of the project were to build the database fill it with activity data, write some software to create the recommendations create a search interface to show the recommendations test it with some users
We push as much as possible through ezproxy, so we use it for access through our discovery solution, for links from SFX, for links placed in our VLE. So it seemed the obvious choice as the place to start to look at e-resource activity data. We didn’t have access to the Ebsco Discovery log files and we hadn’t been using that system for long whereas we did have a few months of log files from EZProxySo we started with the EZProxy log files as the core dataset
So when we start to look in detail at what data is contained within the log files you’ve got some useful data and other data that isn’t so useful for activity data purposes.We know the user name – that’s the oucu the Open University Computer User account name. You know the request, that is the website that is being accessedSo when you look at the detail of the record what you get is…
Something that looks like this (we’ve anonymised the oucu for obvious reasons).this is one record out of tens of thousands of rows but with a bit of work you can break it down
So you’ve got the date and time – useful to be able to know when something happened
And the oucu of the user
And the request that has been made – in this case an ebsco host search
So our database starts to build up with details of userand resources
So we can get data about the courses that students were studying from our internal student information system
So that added a bit more to the mix
So, the data we have so far can tell us which courses people are on, so we can make recommendations based on that, i.e. these are the most popular resources that people on your course are looking at. We can also start to say that if you looked at resource C and then straightaway looked at resource D that there is a likelihood that there is some relationship between resource C and resource D.And we can also say which overall are the most popular articles or journals.But there are limitations to the ezproxy data, we don’t have the search terms that are used to find these resources.
But there are limitations. From the logs you don’t always know what search terms were used or have much information about the item that is being accessedAnd if you want to make a recommendation you don’t even have an article or journal title to show as the recommendationSo looked at how we could improve the data. At the moment we use another EDS API call to extract bibliographic details that are used to extract data from Crossref that we can store in the database
So we decided that we could use the EDS API to retrieve some bibliographic data.Originally we’d hoped that we would be able to store basic metadata from EBSCO in the system but after discussion with them we realised that the license terms wouldn’t let us do that.So we had to look for other metadata sources that we could use. So we set the system up to retrieve data keys from EBSCO and use them to search Crossref. The Crossref data license allows you to store that data locally.
We created a test search interface to test recommendations with users using the Ebsco Discovery Solution API.
And when you get your search results, you also get recommendations based the articles viewed by people who used similar search terms
If you view one of the recommended resources it will open the record in another window and you are given the chance to rate the usefulness of the recommendation.
We also built a second interface – this one is a Google Gadget version with pretty much the same functions as the main interface.
Log in sorted out by working with SocialLearn team
We also then started to capture search terms used in the RISE interface
Now we can add search terms that are being used
So we’ve ended up with a set of data that can give us a range of different types of recommendationsFrom ‘people on your course are looking at these articles’ through ‘people who looked at this article also looked at this article’ and ‘to people using this search term looked at these resources’And we are sure that you could put the data to other types of use.
When we were looking at recommendations we thought that the simplest approach was just to start with something very basicWhat drives the recommendations is a set of relationship values. Values are assigned based on resource views and subsequent ratings by usersThe relationships are ranked according to value so the top ones get shown as recommendations.
Each relationship starts as value 0 +1 each time the resource is viewed +1 each time the recommendation is viewed +1 each time the recommendation is rated as ‘Useful’ -2 each time the recommendation is rated as ‘Not Useful’Recommendations are displayed in value order
Any system that deals with personal data has to be mindful of privacy and data protection requirements. After discussion within the Activity Data programme and some helpful information particularly from EDINA’sOpenURL project we put together a specific privacy policy and discussed it with our data protection people at the University. The policy explicitly covered activity data and we have linked to it from the RISE interfaces, from our main EDS page and from SFX. The policy gives people an opt-out to have their data removed from the recommendations, even though they aren’t identified personally in any of the recommendations.With the new EU ‘cookies’ legislation we are doing some more work to ensure that we are legally compliant. Ideally we would want any institutional ‘cookie’ policy and agreement to cover permission to use data for this type of activity.
The original plan with the project was to be able to release an open data set of search data. And we spent quite a lot of time looking at methods of anonymising the data, by removing oucus, genericising courses to broad subjects and looking at whether there was a threshold of students that we needed on a course to be able to release any data from that course.We faced a major challenge because the activity data we had was fairly meaningless without some article metadata and at the time we could only find data we could use ourselves and nothing we could make available in an open data set.So unfortunately it wasn’t possible to release the data. But others at EDINA, LIDP and SALT were able to do so.
Google Gadget will go into list of tools for students alongside those being developed by DOULSWe are migrating the database so we can use it for more mainstream use. We plan to use it for the new MACON mobiles search project. And we’re interested in how this data could be used by Learning Analytics
We are also looking at how we can use these approaches to provide personalised services to users through the library website, so have been looking at being able to show people what articles are being looked at and have been developing some beta services to demonstrate this
EZProxy data – on its own it there are limits to the recommendations you can make, they would mostly be about which are the most popular resourcesOur main issue is to get access to bibliographic data about the articles being accessed and recommended.You need to combine the ezproxy data with other stuff, such CIRCE dataThe more data you can get the better. The more data you get hold of the better you can make the recommendationsLicense restrictions on article level metadata limit what you can store in your database
I’m now going to hand over to Liz who will take you through the findings of the testing with users