PATHSenrich: A Web Service Prototype for Automatic Cultural Heritage Item Enrichment, Eneko Agirre, Ander Barrena, Kike Fernandez, Esther Miranda, Arantxa Otegi, and Aitor Soroa, paper presented the international conference on Theory and Practice in Digital Libraries, TPDL 2013
Large amounts of cultural heritage material are nowadays available through online digital library portals. Most of these cultural items have short descriptions and lack rich contextual information. The PATHS project has developed experimental enrichment services. As a proof of concept, this paper presents a web service prototype which allows independent content providers to enrich cultural heritage items with a subset of the full functionality: links to related items in the collection and links to related Wikipedia articles. In the future we plan to provide more advanced functionality, as available offline for PATHS.
Implementing Recommendations in the PATHS system, SUEDL 2013pathsproject
Implementing Recommendations in the PATHS system, Paul Clough, Arantxa Otegi, Eneko Agirre and Mark Hall, paper presented at the Supporting Users Exploration of Digital Libraries, SUEDL 2013 workshop, during TPDL 2013 in Valetta, Malta
In this paper we describe the design and implementation of nonpersonalized recommendations in the PATHS system. This system allows users to explore items from Europeana in new ways. Recommendations of the type “people who viewed this item also viewed this item” are powered by pairs of viewed items mined from Europeana. However, due to limited usage data only 10.3% of items in the PATHS dataset have recommendations (4.3% of item pairs visited more than once). Therefore, “related items”, a form of contentbased recommendation, are offered to users based on identifying similar items. We discuss some of the problems with implementing recommendations and highlight areas for future work in the PATHS project.
User-Centred Design to Support Exploration and Path Creation in Cultural Her...pathsproject
User-Centred Design to Support Exploration and Path Creation in Cultural Heritage Collections, Paula Goodale, Paul Clough, Nigel Ford, Mark Hall, Mark Stevenson, Samuel Fernando, Nikolaos Aletras, Kate Fernie, Phil Archer, Andrea De Polo, Euro HCIR 2012. Pre-print of paper.
In this paper we present the results of the user requirements and interface design phase for a prototype system, designed to enhance interaction with cultural heritage collections
online through means of a pathway metaphor. We present a single user interaction model that supports various work and
information seeking tasks undertaken by both expert and non
-expert users within the context of collection exploration and path creation. The user interaction model is shown to enable seamless movement between interaction modes, with the potential over time to encourage deeper engagement and learning.
Roadmap from ESEPaths to EDMPaths: a note on representing annotations resulti...pathsproject
Roadmap from ESEPaths to EDMPaths: a note on representing annotations resulting from automatic enrichment - Aitor Soroa, Eneko Agirre, Arantxa Otegi and Antoine Isaac
This document is a case study on using the Europeana Data Model (EDM) [Doerr et al., 2010] for representing annotations of Cultural Heritage Objects (CHO). One of the main goals of
the PATHS project is to augment CHOs (items) with information that will enrich the user’s experience. The additional information includes links between items in cultural collections and from items to external sources like Wikipedia. With this goal, the PATHS project has applied Natural Language Processing (NLP) techniques on a subset of the items in Europeana.
Aletras, Nikolaos and Stevenson, Mark (2013) "Evaluating Topic Coherence Us...pathsproject
Aletras, Nikolaos and Stevenson, Mark (2013) "Evaluating Topic Coherence Using Distributional Semantics” Proceedings of the 10th International Conference on Computational Semantics (IWCS 2013) -- Long Papers, Potsdam, Germany
Semantic Enrichment of Cultural Heritage content in PATHSpathsproject
Semantic Enrichment of Cultural Heritage content in PATHS, report by Mark Stevenson and Arantxa Otegi with Eneko Agirre, Nikos Aletras, Paul Clough, Samuel Fernando and Aitor Saroa.
The aim of the PATHS project is to enable exploration and discovery within cultural heritage collections. In order to support this the project developed a range of enrichment techniques which augmented these collections with additional information to enhance the users’ browsing experience. One of the demonstration systems developed in PATHS makes use of content from Europeana. This document summarises the semantic enrichment techniques developed in PATHS, with particular reference to their application to the Europeana data.
Implementing Recommendations in the PATHS system, SUEDL 2013pathsproject
Implementing Recommendations in the PATHS system, Paul Clough, Arantxa Otegi, Eneko Agirre and Mark Hall, paper presented at the Supporting Users Exploration of Digital Libraries, SUEDL 2013 workshop, during TPDL 2013 in Valetta, Malta
In this paper we describe the design and implementation of nonpersonalized recommendations in the PATHS system. This system allows users to explore items from Europeana in new ways. Recommendations of the type “people who viewed this item also viewed this item” are powered by pairs of viewed items mined from Europeana. However, due to limited usage data only 10.3% of items in the PATHS dataset have recommendations (4.3% of item pairs visited more than once). Therefore, “related items”, a form of contentbased recommendation, are offered to users based on identifying similar items. We discuss some of the problems with implementing recommendations and highlight areas for future work in the PATHS project.
User-Centred Design to Support Exploration and Path Creation in Cultural Her...pathsproject
User-Centred Design to Support Exploration and Path Creation in Cultural Heritage Collections, Paula Goodale, Paul Clough, Nigel Ford, Mark Hall, Mark Stevenson, Samuel Fernando, Nikolaos Aletras, Kate Fernie, Phil Archer, Andrea De Polo, Euro HCIR 2012. Pre-print of paper.
In this paper we present the results of the user requirements and interface design phase for a prototype system, designed to enhance interaction with cultural heritage collections
online through means of a pathway metaphor. We present a single user interaction model that supports various work and
information seeking tasks undertaken by both expert and non
-expert users within the context of collection exploration and path creation. The user interaction model is shown to enable seamless movement between interaction modes, with the potential over time to encourage deeper engagement and learning.
Roadmap from ESEPaths to EDMPaths: a note on representing annotations resulti...pathsproject
Roadmap from ESEPaths to EDMPaths: a note on representing annotations resulting from automatic enrichment - Aitor Soroa, Eneko Agirre, Arantxa Otegi and Antoine Isaac
This document is a case study on using the Europeana Data Model (EDM) [Doerr et al., 2010] for representing annotations of Cultural Heritage Objects (CHO). One of the main goals of
the PATHS project is to augment CHOs (items) with information that will enrich the user’s experience. The additional information includes links between items in cultural collections and from items to external sources like Wikipedia. With this goal, the PATHS project has applied Natural Language Processing (NLP) techniques on a subset of the items in Europeana.
Aletras, Nikolaos and Stevenson, Mark (2013) "Evaluating Topic Coherence Us...pathsproject
Aletras, Nikolaos and Stevenson, Mark (2013) "Evaluating Topic Coherence Using Distributional Semantics” Proceedings of the 10th International Conference on Computational Semantics (IWCS 2013) -- Long Papers, Potsdam, Germany
Semantic Enrichment of Cultural Heritage content in PATHSpathsproject
Semantic Enrichment of Cultural Heritage content in PATHS, report by Mark Stevenson and Arantxa Otegi with Eneko Agirre, Nikos Aletras, Paul Clough, Samuel Fernando and Aitor Saroa.
The aim of the PATHS project is to enable exploration and discovery within cultural heritage collections. In order to support this the project developed a range of enrichment techniques which augmented these collections with additional information to enhance the users’ browsing experience. One of the demonstration systems developed in PATHS makes use of content from Europeana. This document summarises the semantic enrichment techniques developed in PATHS, with particular reference to their application to the Europeana data.
Cluster Based Web Search Using Support Vector MachineCSCJournals
Now days, searches for the web pages of a person with a given name constitute a notable fraction of queries to Web search engines. This method exploits a variety of semantic information extracted from web pages. The rapid growth of the Internet has made the Web a popular place for collecting information. Today, Internet user access billions of web pages online using search engines. Information in the Web comes from many sources, including websites of companies, organizations, communications and personal homepages, etc. Effective representation of Web search results remains an open problem in the Information Retrieval community. For ambiguous queries, a traditional approach is to organize search results into groups (clusters), one for each meaning of the query. These groups are usually constructed according to the topical similarity of the retrieved documents, but it is possible for documents to be totally dissimilar and still correspond to the same meaning of the query. To overcome this problem, the relevant Web pages are often located close to each other in the Web graph of hyperlinks. It presents a graphical approach for entity resolution & complements the traditional methodology with the analysis of the entity-relationship (ER) graph constructed for the dataset being analyzed. It also demonstrates a technique that measures the degree of interconnectedness between various pairs of nodes in the graph. It can significantly improve the quality of entity resolution. Using Support vector machines (SVMs) which are a set of related Supervised learning methods used for classification of load of user queries to the sever machine to different client machines so that system will be stable. clusters web pages based on their capacities stores whole database on server machine. Keywords: SVM, cluster; ER.
Tutorial at OAI5 (cern.ch/oai5). Abstract: This tutorial will provide a practical overview of current practices in modelling complex or compound digital objects. It will examine some of the key scenarios around creating complex objects and will explore a number of approaches to packaging and transport. Taking research papers, or scholarly works, as an example, the tutorial will explore the different ways in which these, and their descriptive metadata, can be treated as complex objects. Relevant application profiles and metadata formats will be introduced and compared, such as Dublin Core, in particular the DCMI Abstract Model, and MODS, alongside content packaging standards, such as METS MPEG 21 DIDL and IMS CP. Finally, we will consider some future issues and activities that are seeking to address these. The tutorial will be of interest to librarians and technical staff with an interest in metadata or complex objects, their creation, management and re-use.
Usage of Linked Data: Introduction and Application ScenariosEUCLID project
This presentation introduces the main principles of Linked Data, the underlying technologies and background standards. It provides basic knowledge for how data can be published over the Web, how it can be queried, and what are the possible use cases and benefits. As an example, we use the development of a music portal (based on the MusicBrainz dataset), which facilitates access to a wide range of information and multimedia resources relating to music.
Metadata enriching and filtering for enhanced collection discoverability Getaneh Alemu
The return on investment for academic libraries is chiefly tied to access, usage and impact. Without accurate, consistent and quality metadata on the one hand, and an easy-to-use and effective discovery service on the other, these valuable resources may remain invisible and inaccessible to users. In this talk, Getaneh aims to present four overarching metadata principles, namely: metadata enriching, linking, openness and filtering. And how these ideas help shape the metadata creation and discovery services at Solent University – focusing on the implementation of RDA and FRBR as well as the use of subject headings and authority controls.
Big Linked Data - Creating Training CurriculaEUCLID project
This presentation includes an overview of the basic rules to follow when developing training and education curricula for Linked Data and Big Linked Data
Metadata enriching and discovery at Solent University Library Getaneh Alemu
Metadata enriching and discovery at Solent University Library / Getaneh Alemu
In today’s challenging financial environment, we at Solent are convinced that access, usage, impact and return on investment in library resources are ever more important. At Solent, 90% of the resources budget goes onto digital resources (e-books, e-journals and databases), and usage drives everything. We place particular emphasis on the function of the cataloguing and metadata part of our work. Metadata plays an important role to ensuring print and electronic resources are discoverable and usable by users as well as supporting the circulation, acquisition and interlibrary loan functions of a library (Alemu & Stevens, 2015; Haynes, 2018; Gartner, 2016; Zeng & Qin, 2016). “Metadata plays a critical role in the function of any discovery service. Search, relevancy ranking, faceted refinement, and recording grouping function (FRBR) all respond to the metadata present” (Han & Weathers, 2016, p, 276 in Varnum, 2016).
Metadata aides acquisition – the more we know about a resource, the more likely we acquire the right one. Metadata also aides usage – as undiscoverable resources do not get used. To this end, we are early adopters of RDA and FRBR. Currently, we are also part of the Jisc Hub Discovery where we share our bibliographic data. This short paper argues that metadata that is enriched, linked, open and filtered drives usage of resource (Alemu, 2014).
Presented at CIG (MDG) Conference 2020
Metadata and Discovery - Online, 7-11 September 2020
This presentation focuses on providing means for exploring Linked Data. In particular, it gives an overview of current visualization tools and techniques, looking at semantic browsers and applications for presenting the data to the end used. We also describe existing search options, including faceted search, concept-based search and hybrid search, based on a mix of using semantic information and text processing. Finally, we conclude with approaches for Linked Data analysis, describing how available data can be synthesized and processed in order to draw conclusions.
This presentation gives details on technologies and approaches towards exploiting Linked Data by building LD applications. In particular, it gives an overview of popular existing applications and introduces the main technologies that support implementation and development. Furthermore, it illustrates how data exposed through common Web APIs can be integrated with Linked Data in order to create mashups.
What Are Links in Linked Open Data? A Characterization and Evaluation of Link...Armin Haller
Linked Open Data promises to provide guiding principles to publish interlinked knowledge graphs on the Web in the form of findable, accessible, interoperable, and reusable datasets. In this talk I argue that while as such, Linked Data may be viewed as a basis for instantiating the FAIR principles, there are still a number of open issues that cause significant data quality issues even when knowledge graphs are published as Linked Data. In this talk I will first define the boundaries of what constitutes a single coherent knowledge graph within Linked Data, i.e., present a principled notion of what a dataset is and what links within and between datasets are. I will also define different link types for data in Linked datasets and present the results of our empirical analysis of linkage among the datasets of the Linked Open Data cloud. Recent results from our analysis of Wikidata, which has not been part of the Linked Open Data Cloud, will also be presented.
Cluster Based Web Search Using Support Vector MachineCSCJournals
Now days, searches for the web pages of a person with a given name constitute a notable fraction of queries to Web search engines. This method exploits a variety of semantic information extracted from web pages. The rapid growth of the Internet has made the Web a popular place for collecting information. Today, Internet user access billions of web pages online using search engines. Information in the Web comes from many sources, including websites of companies, organizations, communications and personal homepages, etc. Effective representation of Web search results remains an open problem in the Information Retrieval community. For ambiguous queries, a traditional approach is to organize search results into groups (clusters), one for each meaning of the query. These groups are usually constructed according to the topical similarity of the retrieved documents, but it is possible for documents to be totally dissimilar and still correspond to the same meaning of the query. To overcome this problem, the relevant Web pages are often located close to each other in the Web graph of hyperlinks. It presents a graphical approach for entity resolution & complements the traditional methodology with the analysis of the entity-relationship (ER) graph constructed for the dataset being analyzed. It also demonstrates a technique that measures the degree of interconnectedness between various pairs of nodes in the graph. It can significantly improve the quality of entity resolution. Using Support vector machines (SVMs) which are a set of related Supervised learning methods used for classification of load of user queries to the sever machine to different client machines so that system will be stable. clusters web pages based on their capacities stores whole database on server machine. Keywords: SVM, cluster; ER.
Tutorial at OAI5 (cern.ch/oai5). Abstract: This tutorial will provide a practical overview of current practices in modelling complex or compound digital objects. It will examine some of the key scenarios around creating complex objects and will explore a number of approaches to packaging and transport. Taking research papers, or scholarly works, as an example, the tutorial will explore the different ways in which these, and their descriptive metadata, can be treated as complex objects. Relevant application profiles and metadata formats will be introduced and compared, such as Dublin Core, in particular the DCMI Abstract Model, and MODS, alongside content packaging standards, such as METS MPEG 21 DIDL and IMS CP. Finally, we will consider some future issues and activities that are seeking to address these. The tutorial will be of interest to librarians and technical staff with an interest in metadata or complex objects, their creation, management and re-use.
Usage of Linked Data: Introduction and Application ScenariosEUCLID project
This presentation introduces the main principles of Linked Data, the underlying technologies and background standards. It provides basic knowledge for how data can be published over the Web, how it can be queried, and what are the possible use cases and benefits. As an example, we use the development of a music portal (based on the MusicBrainz dataset), which facilitates access to a wide range of information and multimedia resources relating to music.
Metadata enriching and filtering for enhanced collection discoverability Getaneh Alemu
The return on investment for academic libraries is chiefly tied to access, usage and impact. Without accurate, consistent and quality metadata on the one hand, and an easy-to-use and effective discovery service on the other, these valuable resources may remain invisible and inaccessible to users. In this talk, Getaneh aims to present four overarching metadata principles, namely: metadata enriching, linking, openness and filtering. And how these ideas help shape the metadata creation and discovery services at Solent University – focusing on the implementation of RDA and FRBR as well as the use of subject headings and authority controls.
Big Linked Data - Creating Training CurriculaEUCLID project
This presentation includes an overview of the basic rules to follow when developing training and education curricula for Linked Data and Big Linked Data
Metadata enriching and discovery at Solent University Library Getaneh Alemu
Metadata enriching and discovery at Solent University Library / Getaneh Alemu
In today’s challenging financial environment, we at Solent are convinced that access, usage, impact and return on investment in library resources are ever more important. At Solent, 90% of the resources budget goes onto digital resources (e-books, e-journals and databases), and usage drives everything. We place particular emphasis on the function of the cataloguing and metadata part of our work. Metadata plays an important role to ensuring print and electronic resources are discoverable and usable by users as well as supporting the circulation, acquisition and interlibrary loan functions of a library (Alemu & Stevens, 2015; Haynes, 2018; Gartner, 2016; Zeng & Qin, 2016). “Metadata plays a critical role in the function of any discovery service. Search, relevancy ranking, faceted refinement, and recording grouping function (FRBR) all respond to the metadata present” (Han & Weathers, 2016, p, 276 in Varnum, 2016).
Metadata aides acquisition – the more we know about a resource, the more likely we acquire the right one. Metadata also aides usage – as undiscoverable resources do not get used. To this end, we are early adopters of RDA and FRBR. Currently, we are also part of the Jisc Hub Discovery where we share our bibliographic data. This short paper argues that metadata that is enriched, linked, open and filtered drives usage of resource (Alemu, 2014).
Presented at CIG (MDG) Conference 2020
Metadata and Discovery - Online, 7-11 September 2020
This presentation focuses on providing means for exploring Linked Data. In particular, it gives an overview of current visualization tools and techniques, looking at semantic browsers and applications for presenting the data to the end used. We also describe existing search options, including faceted search, concept-based search and hybrid search, based on a mix of using semantic information and text processing. Finally, we conclude with approaches for Linked Data analysis, describing how available data can be synthesized and processed in order to draw conclusions.
This presentation gives details on technologies and approaches towards exploiting Linked Data by building LD applications. In particular, it gives an overview of popular existing applications and introduces the main technologies that support implementation and development. Furthermore, it illustrates how data exposed through common Web APIs can be integrated with Linked Data in order to create mashups.
What Are Links in Linked Open Data? A Characterization and Evaluation of Link...Armin Haller
Linked Open Data promises to provide guiding principles to publish interlinked knowledge graphs on the Web in the form of findable, accessible, interoperable, and reusable datasets. In this talk I argue that while as such, Linked Data may be viewed as a basis for instantiating the FAIR principles, there are still a number of open issues that cause significant data quality issues even when knowledge graphs are published as Linked Data. In this talk I will first define the boundaries of what constitutes a single coherent knowledge graph within Linked Data, i.e., present a principled notion of what a dataset is and what links within and between datasets are. I will also define different link types for data in Linked datasets and present the results of our empirical analysis of linkage among the datasets of the Linked Open Data cloud. Recent results from our analysis of Wikidata, which has not been part of the Linked Open Data Cloud, will also be presented.
Exchange In-Place eDiscovery & Hold | Introduction | 5#7Eyal Doron
Exchange In-Place eDiscovery & Hold | Introduction | 5#7
http://o365info.com/exchange-in-place-ediscovery-hold-introduction-part-5-7
The Exchange In-Place Hold & eDiscovery, is a very powerful tool that can help us to accomplish three main tasks.
1. Search for information (mail items) in single or multiple mailboxes
2. Put specific information on “hold” (enable to save the information for an unlimited time period)
3. Recover deleted mail items
In this article, we will review the logic and the concepts of the Exchange In-Place Hold & eDiscovery toll.
In the next article xx, we will demonstrate how to use the Exchange In-Place Hold & eDiscovery toll for recovering deleted mail items.
Eyal Doron | o365info.com
The old exchange environment versus modern exchange environment part 02#36Eyal Doron
The old Exchange environment versus “modern” Exchange environment | Part 02#36
The Autodiscover as a solution for the modern Exchange environment versus, the older Exchange server architecture that did not have the Autodiscover infrastructure.
http://o365info.com/the-old-exchange-environment-versus-modern-exchange-environment-part-02-of-36
Eyal Doron | o365info.com
My E-mail appears as spam - Troubleshooting path | Part 11#17Eyal Doron
My E-mail appears as spam - Troubleshooting path | Part 11#17
http://o365info.com/my-e-mail-appears-as-spam-troubleshooting-path-part-11-17
Troubleshooting scenario of internal \ outbound spam in Office 365 and Exchange Online environment.
Verifying if our domain name is blacklisted, verifying if the problem is related to E-mail content, verifying if the problem is related to specific organization user E-mail address, Moving the troubleshooting process to the “other side.
Eyal Doron | o365info.com
How does sender verification work how we identify spoof mail) spf, dkim dmar...Eyal Doron
The process of “sender verification”, enables us to distinguish between a legitimate sender versus an attacker who spoof his identity.
In the current article, we will review in details the five available methods that we can use for fighting the phenomena of Spoof mail attack.
http://o365info.com/how-does-sender-verification-work-how-we-identify-spoof-mail-the-five-heros-spf-dkim-dmarc-exchange-and-exchange-online-protection-part-9-of-9
Recommendations for the automatic enrichment of digital library content using...pathsproject
Recommendations for the enrichment of digital library content using open source software, PATHS report by Eneko Agirre and Arantxa Otegi
The goal of this document is to present an overall set of recommendations for the automatic enrichment of Digital Library content using open source software. It is intended to be useful for third-parties who would like to offer enrichment services. Note that this is not a step-by-step guide for reimplementation, but an overall view of the software required and the programming effort involved.
Nikola Ikonomov, Boyan Simeonov, Jana Parvanova and Vladimir Alexiev. In Digital Presentation and Preservation of Cultural and Scientific Heritage (DiPP 2013), Veliko Tarnovo, Bulgaria, Sep 2013
This is one of 7 reports provided in work package 3: Micro services for small and medium institutions.
Authors:
Aitor Soroa, Arantxa Otegi(, Eneko Agirre(, Rodrigo Agerri
Universidad del País Vasco
(University of the Basque Country)
(EHU)
Case Study: Europeana API Implementation in Polish Digital LibrariesNeil Bates
In 2009, PSNC, the operator of Polish Digital Libraries Federation, began working with Europeana and exploiting the potential of promoting Polish cultural heritage content across Europe with the aim of creating wider access and facilitating re-use of their content. Europeana’s API has provided a good platform for this, both enriching existing services and providing a wider breadth of content to users of the DLW and DLF services. Through a number of widgets, DLW and DLF have utilised Europeana’s API to bring in content from Europeana that is related to search queries on their existing websites. This provides end-users with more choice as results are returned from hundreds of other trusted heritage institutions across Europe via the Europeana API.
Institutional Services and Tools for Content, Metadata and IPR ManagementPaolo Nesi
Ingest a large range metadata formats (XML based or Dublin Core, METS, MPEG-21, etc.) coming from different channels (http, ftp, oai-pmh, etc.) and content files >500 ff.
Perform human content enrichment, translations, validation; comments; social media, rating; promoting; publication; corrections; assessment, etc.
Perform automated activities, technical parameters (duration, size, etc.), descriptors, indexing, translations, VIP names, geonames, LOD, assessment, IPR , verification
IPR modelling, assignment and verification.
Harmonising the activities of human and automated processing
Scale up of the back office architecture to cope with a large number of transactions
Support and model one or more workflows
Presentation at the Education Session of the American Art Collaborative (AAC) Linked Open Data Initiative, 31 March 2015. http://americanartcollaborative.org/
Similar to PATHSenrich: A Web Service Prototype for Automatic Cultural Heritage Item Enrichment, @TPDL 2013 (20)
Generating Paths through Cultural Heritage Collections Latech2013 paperpathsproject
Generating Paths through Cultural Heritage Collections, Samuel Fernando, Paula Goodale, Paul Clough, Mark Stevenson, Mark Hall and Eneko Agirre. Paper presented at Latech 2013
Cultural heritage collections usually organise sets of items into exhibitions or guided tours. These items are often accompanied by text that describes the theme and topic of the exhibition and provides background context and details of connections with other items. The PATHS project brings the idea of guided tours to digital library collections where a tool to create virtual paths are used to assist with navigation and provide guides on particular subjects and topics. In this paper we characterise and analyse paths of items created by users of our online system.
Supporting User's Exploration of Digital Libraries, Suedl 2012 workshop proce...pathsproject
Workshop proceedings from the International workshop on Supporting Users Exploration of Digital Libraries, SUEDL 2012 which was held at TPDL 2012 (the international conference on Theory and Practice in Digital Libraries), Paphos, Cyprus, September 2012.
The aim of the workshop was to stimulate collaboration from experts and stakeholders in Digital Libraries, Cultural Heritage, Natural Language Processing and Information Retrieval in order to explore methods and strategies to support exploration of Digital Libraries, beyond the white box paradigm of search and click.
The proceedings includes:
"Browsing Europeana - Opportunities and Challenges', David Haskiya
"Query re-writing using shallow language processing effects', Anna Mastora and Sarantos Kapidakis
"Visualising Television Heritage" Johan Ooman et al,
"Providing suitable information access for new users of Digital Libraries", Rike Brecht et al
"Exploring Pelagios: a Visual Browser for Geo-tagged datasets" Rainer Simon et al
PATHS state of the art monitoring reportpathsproject
This document provides an update to an Initial State of the Art Monitoring report delivered by the project. The report covers the areas of Educational Informatics, Information Retrieval and Semantic Similarity relatedness.
Generating Paths through Cultural Heritage Collections, LATECH 2013 paperpathsproject
Generating Paths through Cultural Heritage Collections Samuel Fernando, Paula Goodale, Paul Clough, Mark Stevenson, Mark Hall and Eneko Agirre.
The PATHS project brings the idea of guided tours to digital library collections where a tool to create virtual paths are used to assist with navigation and provide guides on particular subjects and topics. In this paper we characterise and analyse paths of items created by users of our online system.
Generating PATHS through Cultural Heritage Collections, Samuel Fernando, Paula Goodale, Paul Clough,
Mark Stevenson, Mark Hall, Eneko Agirre. Presentation given at LaTeCH 2013, ACL Workshop, Sofia, Bulgaria.
Presentation given by Kate Fernie about the PATHS project and the second PATHS prototype at the European Association of Archaeologists conference, Pilsen, September 2013
Let's dive deeper into the world of ODC! Ricardo Alves (OutSystems) will join us to tell all about the new Data Fabric. After that, Sezen de Bruijn (OutSystems) will get into the details on how to best design a sturdy architecture within ODC.
Slack (or Teams) Automation for Bonterra Impact Management (fka Social Soluti...Jeffrey Haguewood
Sidekick Solutions uses Bonterra Impact Management (fka Social Solutions Apricot) and automation solutions to integrate data for business workflows.
We believe integration and automation are essential to user experience and the promise of efficient work through technology. Automation is the critical ingredient to realizing that full vision. We develop integration products and services for Bonterra Case Management software to support the deployment of automations for a variety of use cases.
This video focuses on the notifications, alerts, and approval requests using Slack for Bonterra Impact Management. The solutions covered in this webinar can also be deployed for Microsoft Teams.
Interested in deploying notification automations for Bonterra Impact Management? Contact us at sales@sidekicksolutionsllc.com to discuss next steps.
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
Essentials of Automations: Optimizing FME Workflows with ParametersSafe Software
Are you looking to streamline your workflows and boost your projects’ efficiency? Do you find yourself searching for ways to add flexibility and control over your FME workflows? If so, you’re in the right place.
Join us for an insightful dive into the world of FME parameters, a critical element in optimizing workflow efficiency. This webinar marks the beginning of our three-part “Essentials of Automation” series. This first webinar is designed to equip you with the knowledge and skills to utilize parameters effectively: enhancing the flexibility, maintainability, and user control of your FME projects.
Here’s what you’ll gain:
- Essentials of FME Parameters: Understand the pivotal role of parameters, including Reader/Writer, Transformer, User, and FME Flow categories. Discover how they are the key to unlocking automation and optimization within your workflows.
- Practical Applications in FME Form: Delve into key user parameter types including choice, connections, and file URLs. Allow users to control how a workflow runs, making your workflows more reusable. Learn to import values and deliver the best user experience for your workflows while enhancing accuracy.
- Optimization Strategies in FME Flow: Explore the creation and strategic deployment of parameters in FME Flow, including the use of deployment and geometry parameters, to maximize workflow efficiency.
- Pro Tips for Success: Gain insights on parameterizing connections and leveraging new features like Conditional Visibility for clarity and simplicity.
We’ll wrap up with a glimpse into future webinars, followed by a Q&A session to address your specific questions surrounding this topic.
Don’t miss this opportunity to elevate your FME expertise and drive your projects to new heights of efficiency.
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
Software Delivery At the Speed of AI: Inflectra Invests In AI-Powered QualityInflectra
In this insightful webinar, Inflectra explores how artificial intelligence (AI) is transforming software development and testing. Discover how AI-powered tools are revolutionizing every stage of the software development lifecycle (SDLC), from design and prototyping to testing, deployment, and monitoring.
Learn about:
• The Future of Testing: How AI is shifting testing towards verification, analysis, and higher-level skills, while reducing repetitive tasks.
• Test Automation: How AI-powered test case generation, optimization, and self-healing tests are making testing more efficient and effective.
• Visual Testing: Explore the emerging capabilities of AI in visual testing and how it's set to revolutionize UI verification.
• Inflectra's AI Solutions: See demonstrations of Inflectra's cutting-edge AI tools like the ChatGPT plugin and Azure Open AI platform, designed to streamline your testing process.
Whether you're a developer, tester, or QA professional, this webinar will give you valuable insights into how AI is shaping the future of software delivery.
Kubernetes & AI - Beauty and the Beast !?! @KCD Istanbul 2024Tobias Schneck
As AI technology is pushing into IT I was wondering myself, as an “infrastructure container kubernetes guy”, how get this fancy AI technology get managed from an infrastructure operational view? Is it possible to apply our lovely cloud native principals as well? What benefit’s both technologies could bring to each other?
Let me take this questions and provide you a short journey through existing deployment models and use cases for AI software. On practical examples, we discuss what cloud/on-premise strategy we may need for applying it to our own infrastructure to get it to work from an enterprise perspective. I want to give an overview about infrastructure requirements and technologies, what could be beneficial or limiting your AI use cases in an enterprise environment. An interactive Demo will give you some insides, what approaches I got already working for real.
Accelerate your Kubernetes clusters with Varnish CachingThijs Feryn
A presentation about the usage and availability of Varnish on Kubernetes. This talk explores the capabilities of Varnish caching and shows how to use the Varnish Helm chart to deploy it to Kubernetes.
This presentation was delivered at K8SUG Singapore. See https://feryn.eu/presentations/accelerate-your-kubernetes-clusters-with-varnish-caching-k8sug-singapore-28-2024 for more details.
Connector Corner: Automate dynamic content and events by pushing a buttonDianaGray10
Here is something new! In our next Connector Corner webinar, we will demonstrate how you can use a single workflow to:
Create a campaign using Mailchimp with merge tags/fields
Send an interactive Slack channel message (using buttons)
Have the message received by managers and peers along with a test email for review
But there’s more:
In a second workflow supporting the same use case, you’ll see:
Your campaign sent to target colleagues for approval
If the “Approve” button is clicked, a Jira/Zendesk ticket is created for the marketing design team
But—if the “Reject” button is pushed, colleagues will be alerted via Slack message
Join us to learn more about this new, human-in-the-loop capability, brought to you by Integration Service connectors.
And...
Speakers:
Akshay Agnihotri, Product Manager
Charlie Greenberg, Host
Empowering NextGen Mobility via Large Action Model Infrastructure (LAMI): pav...
PATHSenrich: A Web Service Prototype for Automatic Cultural Heritage Item Enrichment, @TPDL 2013
1. PATHSenrich: A Web Service Prototype
for Automatic Cultural Heritage Item
Enrichment
Eneko Agirre, Ander Barrena, Kike Fernandez, Esther Miranda,
Arantxa Otegi, and Aitor Soroa
IXA NLP Group, University of the Basque Country UPV/EHU
arantza.otegi@ehu.es
Abstract. Large amounts of cultural heritage material are nowadays
available through online digital library portals. Most of these cultural
items have short descriptions and lack rich contextual information. The
PATHS project has developed experimental enrichment services. As a
proof of concept, this paper presents a web service prototype which allows
independent content providers to enrich cultural heritage items with a
subset of the full functionality: links to related items in the collection
and links to related Wikipedia articles. In the future we plan to provide
more advanced functionality, as available offline for PATHS.
1
Introduction
Large amounts of cultural heritage (CH) material are now available through
online digital library portals, such as Europeana1. Europeana hosts millions of
books, paintings, films, museum objects and archival records that have been digitised throughout Europe. Europeana collects contextual information or metadata
about different types of content, which the users can use for their searches.
The main strength of Europeana lays in the vast number of items it contains.
Sometimes, though, this quantity comes at the cost of a restricted amount of
metadata, with many items having very short descriptions and a lack of rich
contextual information. One of the goals of the PATHS project2 is precisely to
enrich CH items, using a selected subset of Europeana as a testbed[1].
Whithin the project, this enrichment will make possible to create a system
that acts as an interactive personalised tour guide through Europeana collections, offering suggestions about items to look at and assist in their interpretation by providing relevant contextual information from related items within
Europeana and items from external sources like Wikipedia. Users of such digital
libraries may require information for purposes such as learning and seeking answers to questions. This additional information supports users in fulfilling their
information need, as the evaluation of the first PATHS prototype shows [2].
In this paper we present a web service prototype which allows independent
content providers to enrich CH items. Specifically, the service enriches the items
1
2
http://www.europeana.eu/portal/
http://www.paths-project.eu
T. Aalberg et al. (Eds.): TPDL 2013, LNCS 8092, pp. 462–465, 2013.
c Springer-Verlag Berlin Heidelberg 2013
2. PATHSenrich: A Web Service Prototype for Automatic CH Item Enrichment
463
with two types of information. On the one hand, the item will be linked to
similar items within the collection. On the other hand, the item will be linked
to Wikipedia articles which are related to it.
There have been many attempts to automatically enrich cultural heritage
metadata. Some projects (for instance, MIMO-DB3 or MERLIN4 ) relate CH
objects with terms of an external authority or vocabulary. Some others (like
MACE5 or YUMA 6 ) adopt a collaborative annotation paradigm for metadata
enrichment. To our knowledge, PATHS is the first project using semantic NLP
processing to link CH items to similar items or external Wikipedia articles.
The current service has limited bandwidth, and provides a selected subset
of the enrichment functionality available internally in the PATHS project. The
quality of the links produce is also slightly lower, although we plan to improve it
in the short future. However, we think that the prototype is useful to demonstrate
the potential to construct a web service for automatically enriching CH items
with high quality information.
2
Demo Description
The web service takes as input one CH item represented following the Europeana
Data Model (EDM) in JSON format, as exported by the Europeana API v2.07 (a
sample record is provided in the interface). The web service returns the following:
– A list of 10 closely related items within the collection.
– A list of Wikipedia pages which are related to the target item.
Figure 1 shows a snapshot of the web service. The service is publicly accessible
following the URL http://ixa2.si.ehu.es/paths_wp2/paths_wp2.pl.
The enrichment is performed by analyzing the metadata associated with the
item, i.e., the title of the item, its description, etc. The next sections briefly
describe how this enrichment is performed.
2.1
Related Items within the Collection
The list of related items is obtained by first creating a query with the content
of the title, subject and description fields (stopwords are removed). The query
is then posted to a SOLR search engine8 . The SOLR search engine accesses an
index created with the subset of Europeana items already enriched offline within
the PATHS project. In that way, the most related Europeana items in the subset
are obtained, and the identifiers of those related items are listed. Note that the
related items used internally in the PATHS project are produced using more
sophisticated methods. Please refer to [1] for further details.
3
4
5
6
7
8
http://www.mimo-international.com
http://www.ucl.ac.uk/ls/merlin
http://www.mace-project.eu
http://dme.ait.ac.at/annotation
http://preview.europeana.eu/portal/api-introduction.html
http://lucene.apache.org/solr/
3. 464
E. Agirre et al.
Fig. 1. Web service interface. It consists of a text area to introduce the input item
in JSON format (top). The “Get EDM JSON example” button can be used to get an
input example. Once a JSON record is typed, click “Process” button to get the output.
The output (bottom) consists on a list of related items and background links.
2.2
Related Wikipedia Articles
For linking the items to Wikipedia articles we follow an implementation similar
to the method described in [3]. This method creates a dictionary, an association
between string mentions with all possible articles the mention can refer to. Our
dictionary is constructed using the title of the Wikipedia article, the redirect
pages, the disambiguation pages and the anchor texts from Wikipedia links.
Mentions are lower-cased and all text between parenthesis is removed. If the
mention links to a disambiguation page, it is associated with all possible articles
the disambiguation page points to. Besides, each association between a mention
and article is scored with the prior probability, estimated as the number of
times that the mention occurs in the anchor text of an article. Note that such
dictionaries can disambiguate any mention, just returning the highest-scoring
article for this particular mention.
Once the dictionary is built, the web service analyzes the title, subject and
description fields of the CH item and matches the longest substring within those
fields with entries in the dictionary. When a match is found, the Wikipedia article
with highest score for this entry is returned. Note that the links to Wikipedia
in the PATHS project are produced using more sophisticated methods. Please
refer to [1] for further details.
4. PATHSenrich: A Web Service Prototype for Automatic CH Item Enrichment
3
465
Conclusions and Future Work
This paper presents a web service prototype which automatically enriches CH
items with metadata. The web service is inspired in the enrichment work carried
out in the PATHS project, but, contrary to the batch methodology used in the
project, this enrichment is performed online. The prototype has been designed
for demonstration purposes, to showcase the feasibility of providing full-fledged
automatic enrichment.
Our plans for the future include moving the offline enrichment services which
are currently being evaluated in the PATHS project to the web service. In the
case of related Wikipedia articles, we will take into account the context of the
matched entities, which improves the quality of the links [4], and we will include
a filtering algorithm to discard entities that are not relevant. Regarding related
items, we will classify them according to the type of relation [5]. In addition we
plan to automatically organize the items hierarchically, according to a Wikipediabased vocabulary [6].
Acknowledgements. The research leading to these results was carried out as
part of the PATHS project (http://www.paths-project.eu) funded by European Communitys Seventh Framework Programme (FP7/2007- 2013) under
grant agreement no. 270082. The work has been also funded by the Basque
Government (project IBILBIDE, SAIOTEK S-PE12UN089).
References
1. Otegi, A., Agirre, E., Soroa, A., Aletras, N., Chandrinos, C., Fernando, S., GonzalezAgirre, A.: Report accompanying D2.2: Processing and Representation of Content
for Second Prototype. PATHS Project Deliverable (2012),
http://www.paths-project.eu/eng/content/download/2489/18113/version/2/
file/D2.2.Content+Processing-2nd+Prototype-revised.v2.pdf
2. Griffiths, J., Goodale, P., Minelli, S., de Polo, A., Agerri, R., Soroa, A., Hall, M.,
Bergheim, S.R., Chandrinos, K., Chryssochoidis, G., Fernie, K., Usher, T.: D5.1:
Evaluation of the first PATHS prototype. PATHS Project Deliverable (2012),
http://www.paths-project.eu/eng/Resources/
D5.1-Evaluation-of-the-1st-PATHS-Prototype
3. Chang, A.X., Spitkovsky, V.I., Yeh, E., Agirre, E., Manning, C.D.: Stanford-UBC
entity linking at TAC-KBP. In: Proceedings of TAC 2010, Gaithersburg, Maryland,
USA (2010)
4. Han, X., Sun, L.: A Generative Entity-Mention Model for Linking Entities with
Knowledge Base. In: Proceedings of the ACL, Portland, Oregon, USA (2011)
5. Agirre, E., Aletras, N., Gonzalez-Agirre, A., Rigau, G., Stevenson, M.: UBC UOSTYPED: Regression for typed-similarity. In: Second Joint Conference on Lexical
and Computational Semantics (*SEM), Atlanta, Georgia, USA (2013)
6. Fernando, S., Hall, M., Agirre, E., Soroa, A., Clough, P., Stevenson, M.: Comparing Taxonomies for Organising Collections of Documents. In: Proceedings of
COLING 2012, Mumbai, India (2013)