Presentation at ISWC2018: http://iswc2018.semanticweb.org/sessions/the-rijksmuseum-collection-as-linked-data/ of our paper published originally in the Semantic Web Journal: http://www.semantic-web-journal.net/content/rijksmuseum-collection-linked-data-2
Many museums are currently providing online access to their collections. The state of the art research in the last decade shows that it is beneficial for institutions to provide their datasets as Linked Data in order to achieve easy cross-referencing, interlinking and integration. In this paper, we present the Rijksmuseum linked dataset (accessible at http://datahub.io/dataset/rijksmuseum), along with collection and vocabulary statistics, as well as lessons learned from the process of converting the collection to Linked Data. The version of March 2016 contains over 350,000 objects, including detailed descriptions and high-quality images released under a public domain license.
The Vienna History Wiki – a Collaborative Knowledge Platform for the City of...Bernhard Krabina
The Vienna City Archive and the Vienna City Library have joined forces with several other institutions in Vienna, Austria to create the “Wien Geschichte Wiki” (Vienna History Wiki), a knowledge platform for the history of Vienna with more than 34,000 articles and 120,000 visits per month. The wiki is powered by Semantic MediaWiki and serves not only as an online encyclopedia for everybody to use and contribute to, but also as a central knowledge base for several administrative departments of the city administration. In a peer-review process, wiki edits are checked before they become visible. A usage log analysis and an online survey have been carried out to gain first insights after six months of operation.
“Archäologische Informationen” and Open Journal Systems. Chances and Possibil...ariadnenetwork
Presentation by Alexandra Büttner, Heidelberg University Library, Germany
EAA 2014 session: Open Access and Open Data in Archaeology
Istanbul, Turkey
13 September 2013
Research partnerships, user participation, extended outreach – some of ETH L...ETH-Bibliothek
IFLA Satellite Meeting 2017: Digital Humanities, Berlin, August 2017
> From "boutique" to mass digitization
> (Cooperative) online platforms for digitized content
> Research Partnerships
> User Participation
> Outreach
The Vienna History Wiki – a Collaborative Knowledge Platform for the City of...Bernhard Krabina
The Vienna City Archive and the Vienna City Library have joined forces with several other institutions in Vienna, Austria to create the “Wien Geschichte Wiki” (Vienna History Wiki), a knowledge platform for the history of Vienna with more than 34,000 articles and 120,000 visits per month. The wiki is powered by Semantic MediaWiki and serves not only as an online encyclopedia for everybody to use and contribute to, but also as a central knowledge base for several administrative departments of the city administration. In a peer-review process, wiki edits are checked before they become visible. A usage log analysis and an online survey have been carried out to gain first insights after six months of operation.
“Archäologische Informationen” and Open Journal Systems. Chances and Possibil...ariadnenetwork
Presentation by Alexandra Büttner, Heidelberg University Library, Germany
EAA 2014 session: Open Access and Open Data in Archaeology
Istanbul, Turkey
13 September 2013
Research partnerships, user participation, extended outreach – some of ETH L...ETH-Bibliothek
IFLA Satellite Meeting 2017: Digital Humanities, Berlin, August 2017
> From "boutique" to mass digitization
> (Cooperative) online platforms for digitized content
> Research Partnerships
> User Participation
> Outreach
A Corpus of Chinese Comic Books: Database, Metadata, and Visual Object Recogn...Matthias Arnold
Chinese comics mostly appeared in the form of small rectangular pocket-books (ca. 13 x 9 cm) with stories containing a picture on the top and a short narrative on the bottom. They are called “chain(ed) pictures” lianhuanhua 连环画 or, alternatively, “small people’s books” xiaorenshu 小人书. These have been the dominant form of comic popular in China since the 1910s and, at least until the late 1980s They introduced generations of Chinese to literature as well as propaganda texts. Especially popular in the 1960s and 1970s, they were the dominating popular medium in everyday life and helped to form a cultural memory. Since the 1990s they have almost completely disappeared and can only be found in private collections today. Academic interest only gradually evolves with researcher focusing on individual artists or publishing systematic genre overviews. The Heidelberg collection is an attempt to join some of the scattered holdings in Germany to form a basis for future research. It comprises the scans of over 1200 comic books, which are combined with systematic metadata recordings and a basic contents analysis through keywords. The focus of our endeavor was on comics from the second half of the Cultural Revolution and immediately thereafter.
In my presentation I will introduce the medium lianhuanhua, discuss aspects of metadata recording and database development, as well as report on an experimental approach to automatically detect visual objects in the images.
Workshop "Comics Annotation: Designing Common Frameworks for Empirical Research", Potsdam 2018-06-18/19
By invitation of the Center for Public Innovation in Bucharest, Romania, made possible by the Dutch embassy in Romania on the occasion of the Culture Hack program, a presentation to see what Romania and The Netherlands can collaborate upon in the context of European Digital Culture with a focus on open data.
Integration and Exploration of Financial Data using Semantics and OntologiesRoberto García
Keynote at the Eurofiling XBRL Week, Academic Track, 6-9 June 2017. Hosted by the European Central Bank, Frankfurt, Germany. The keynote reported about one the first attempts to move a significant amount of XBRL to the Semantic Web, modelling XRML XML with RDF and XBRL Taxonomies with OWL.
Tanya Szrajber, The British Museum Collection DatabaseAndrew Prescott
'The British Museum Collection Database: How to Create and Manage over 2,000,000 Records': seminar by Tanya Szrajber, Head of Documentation, The British Museum, to Department of Digital Humanities, King's College London, 20 November 2012
A Corpus of Chinese Comic Books: Database, Metadata, and Visual Object Recogn...Matthias Arnold
Chinese comics mostly appeared in the form of small rectangular pocket-books (ca. 13 x 9 cm) with stories containing a picture on the top and a short narrative on the bottom. They are called “chain(ed) pictures” lianhuanhua 连环画 or, alternatively, “small people’s books” xiaorenshu 小人书. These have been the dominant form of comic popular in China since the 1910s and, at least until the late 1980s They introduced generations of Chinese to literature as well as propaganda texts. Especially popular in the 1960s and 1970s, they were the dominating popular medium in everyday life and helped to form a cultural memory. Since the 1990s they have almost completely disappeared and can only be found in private collections today. Academic interest only gradually evolves with researcher focusing on individual artists or publishing systematic genre overviews. The Heidelberg collection is an attempt to join some of the scattered holdings in Germany to form a basis for future research. It comprises the scans of over 1200 comic books, which are combined with systematic metadata recordings and a basic contents analysis through keywords. The focus of our endeavor was on comics from the second half of the Cultural Revolution and immediately thereafter.
In my presentation I will introduce the medium lianhuanhua, discuss aspects of metadata recording and database development, as well as report on an experimental approach to automatically detect visual objects in the images.
Workshop "Comics Annotation: Designing Common Frameworks for Empirical Research", Potsdam 2018-06-18/19
By invitation of the Center for Public Innovation in Bucharest, Romania, made possible by the Dutch embassy in Romania on the occasion of the Culture Hack program, a presentation to see what Romania and The Netherlands can collaborate upon in the context of European Digital Culture with a focus on open data.
Integration and Exploration of Financial Data using Semantics and OntologiesRoberto García
Keynote at the Eurofiling XBRL Week, Academic Track, 6-9 June 2017. Hosted by the European Central Bank, Frankfurt, Germany. The keynote reported about one the first attempts to move a significant amount of XBRL to the Semantic Web, modelling XRML XML with RDF and XBRL Taxonomies with OWL.
Tanya Szrajber, The British Museum Collection DatabaseAndrew Prescott
'The British Museum Collection Database: How to Create and Manage over 2,000,000 Records': seminar by Tanya Szrajber, Head of Documentation, The British Museum, to Department of Digital Humanities, King's College London, 20 November 2012
Integrating archaeological data: The ARIADNE Infrastructure, Achille Felicett...ariadnenetwork
This presentation by Achille Felicetti of PIN (Università degli Studi di Firenze, Prato) on the work by the ARIADNE infrastructure to integrating archaeological data was given as part of a workshop organised by Digital Humanities Austria. The workshop focussed on the pressing question of long-term preservation of digital data from various angles, central being user needs specific to the different fields of the Humanities. Felicetti introduced the ARIADNE research infrastructure, which has been funded by the EC's FP7 programme, to integrate archaeological research datasets from across Europe and support their uses by researchers.
Keynote presentation for CSWS 2013 Conference in Shanghai, China.
Some slides borrowed from Jan Wielemaker, Guus Schreiber, Jacco van Ossenbruggen, Niels Ockeloen, Antske Fokkens, Serge ter Braake.
A Framework for Improved Access to Museum Databases in the Semantic WebMariana Damova, Ph.D
This paper presents a framework for processing Museum databases according to a set of interlinked ontologies, including CIDOC-CRM, and loading them in a reason-able view of the web of data, providing additional links to datasets from the LOD cloud. The infrastructure allows accessing the data via SPARQL queries and to verbalize the query results in natural language, the GF formalism, which allows access to 18 natural languages.
Linked Open Data Publications through Wikidata & Persistent Identification...PACKED vzw
In order for museums to truly reap the benefits of publishing their collections online in a sustainable way, PACKED vzw presents the results of its Linked open data project as a best practice guide for the Flemish heritage sector.
In order for museums to truly reap the benefits of publishing their collections online in a sustainable way, PACKED vzw presents the results of its Linked open data project as a best practice guide for the Flemish heritage sector.
Managing the Digitization of Large Press ArchivesDLFCLIR
From the 2014 DLF Forum in Atlanta, GA.
Session Leaders
Bassem Elsayed, Bibliotheca Alexandrina
Ahmed Samir, Bibliotheca Alexandrina
Managing the digitization of press material is quite a challenge; not only in terms of quantity, but also in terms of text and material quality, designing the workflow system which organizes the operations, and handling the metadata. This challenge has been the focus of the Bibliotheca Alexandrina’s digitization work during the past year in the course of its partnership with the Center for Economic, Judicial, and Social Study and Documentation (CEDEJ). Having more than 800,000 pages of press articles to be digitally preserved and publicly accessed, triggered an inevitable need to design a workflow that can manage such a massive collection and handle its attributes proficiently. The deployment of this endeavor required simultaneous intervention of four main aspects; data analysis of the collection, developing a digitization workflow for the collection at hand, implementing and installing the necessary software tools for metadata entry, and finally, publishing the digital archive online for researchers and public access.
The presentation will demonstrate the workflow system which is being implemented to manage this massive press collection, which has yielded to date more than 400,000 pages. It will shed some light on the BA’s Digital Assets Factory (DAF), which is the nucleus upon which the digitization process of CEDEJ collection has been built. Additionally, the presentation will discuss the tools implemented for ingesting data into the digitization process starting form indexing until the creation of batches that are ingested into the system. The outflow will also be discussed in terms of organizing and grouping multipart press clips, in addition to the reviewing, validation and correction of the output. Light will also be shed on the challenges encountered to associate the accessible online archive with a powerful search engine supporting multidimensional search while maintaining a user-friendly navigation experience.
Joseph Padfield and Rupert Shepherd, The National Gallery, and Rob Tice, Knowledge Integration
How can information be opened up within an organisation? The National Gallery was faced with a series of different systems, all holding data related to the collection - but speaking to each other only intermittently. This issue was solved with the installation of a middleware system to combine and deliver data from these eight different data sources as a seamless whole.
Our paper will look at the implications this has had for how we work with our data, and as an organisation. We will also touch upon the benefits of opening information up within our organisation, and some projects that are currently using - or are planning to use - our data, which will be delivered using established, open standards.
Cultural Heritage Information DashboardsRichard Urban
Large-scale aggregations of digital collections from libraries, archives and museums offer users unprecedented access to cultural heritage materials. But they also have failed to incorporate important contextual information that allows users to develop an understanding of the significant features of purpose-built collections. This paper explores the development of information dashboard prototypes that provide users a high-level overview of cultural heritage collections. Two case studies using rapid-prototyping methodologies are presented.
FAIRview: Responsible Video Summarization @NYCML'18Lora Aroyo
Presentation at the NYC Media Lab (NYCML2018). There is a growing demand for news videos online, with more consumers preferring to watch the news than read or listen to it. On the publisher side, there is a growing effort to use video summarization technology in order to create easy-to-consume previews (trailers) for different types of broadcast programs. How can we measure the quality of video summaries and their potential to misinform? This workshop will inform participants about automatic video summarization algorithms and how to produce more “representative” video summaries. The research presented is from the FAIRview project and is supported by the Digital News Innovation Fund (DNI Fund), which is part of the Google News Initiative.
DH Benelux 2017 Panel: A Pragmatic Approach to Understanding and Utilising Ev...Lora Aroyo
Lora Aroyo, Chiel van den Akker, Marnix van Berchum, Lodewijk
Petram, Gerard Kuys, Tommaso Caselli, Jacco van Ossenbruggen, Victor de Boer, Sabrina Sauer, Berber Hagedoorn
Crowdsourcing ambiguity aware ground truth - collective intelligence 2017Lora Aroyo
The process of gathering ground truth data through human annotation is a major bottleneck in the use of information extraction methods. Crowdsourcing-based approaches are gaining popularity in the attempt to solve the issues related to the volume of data and lack of annotators. Typically these practices use inter-annotator agreement as a measure of quality. However, this assumption often creates issues in practice. Previous experiments we performed found that inter-annotator disagreement is usually never captured, either because the number of annotators is too small to capture the full diversity of opinion, or because the crowd data is aggregated with metrics that enforce consensus, such as majority vote. These practices create artificial data that is neither general nor reflects the ambiguity inherent in the data.
To address these issues, we proposed the method for crowdsourcing ground truth by harnessing inter-annotator disagreement. We present an alternative approach for crowdsourcing ground truth data that, instead of enforcing an agreement between annotators, captures the ambiguity inherent in semantic annotation through the use of disagreement-aware metrics for aggregating crowdsourcing responses. Based on this principle, we have implemented the CrowdTruth framework for machine-human computation, that first introduced the disagreement-aware metrics and built a pipeline to process crowdsourcing data with these metrics.
In this paper, we apply the CrowdTruth methodology to collect data over a set of diverse tasks: medical relation extraction, Twitter event identification, news event extraction and sound interpretation. We prove that capturing disagreement is essential for acquiring a high-quality ground truth. We achieve this by comparing the quality of the data aggregated with CrowdTruth metrics with a majority vote, a method which enforces consensus among annotators. By applying our analysis over a set of diverse tasks we show that, even though ambiguity manifests differently depending on the task, our theory of inter-annotator disagreement as a property of ambiguity is generalizable.
My ESWC 2017 keynote: Disrupting the Semantic Comfort ZoneLora Aroyo
Ambiguity in interpreting signs is not a new idea, yet the vast majority of research in machine interpretation of signals such as speech, language, images, video, audio, etc., tend to ignore ambiguity. This is evidenced by the fact that metrics for quality of machine understanding rely on a ground truth, in which each instance (a sentence, a photo, a sound clip, etc) is assigned a discrete label, or set of labels, and the machine’s prediction for that instance is compared to the label to determine if it is correct. This determination yields the familiar precision, recall, accuracy, and f-measure metrics, but clearly presupposes that this determination can be made. CrowdTruth is a form of collective intelligence based on a vector representation that accommodates diverse interpretation perspectives and encourages human annotators to disagree with each other, in order to expose latent elements such as ambiguity and worker quality. In other words, CrowdTruth assumes that when annotators disagree on how to label an example, it is because the example is ambiguous, the worker isn’t doing the right thing, or the task itself is not clear. In previous work on CrowdTruth, the focus was on how the disagreement signals from low quality workers and from unclear tasks can be isolated. Recently, we observed that disagreement can also signal ambiguity. The basic hypothesis is that, if workers disagree on the correct label for an example, then it will be more difficult for a machine to classify that example. The elaborate data analysis to determine if the source of the disagreement is ambiguity supports our intuition that low clarity signals ambiguity, while high clarity sentences quite obviously express one or more of the target relations. In this talk I will share the experiences and lessons learned on the path to understanding diversity in human interpretation and the ways to capture it as ground truth to enable machines to deal with such diversity.
Data Science with Human in the Loop @Faculty of Science #Leiden UniversityLora Aroyo
Software systems are becoming ever more intelligent and more useful, but the way we interact with these machines too often reveals that they don’t actually understand people. Knowledge Representation and Semantic Web focus on the scientific challenges involved in providing human knowledge in machine-readable form. However, we observe that various types of human knowledge cannot yet be captured by machines, especially when dealing with wide ranges of real-world tasks and contexts. The key scientific challenge is to provide an approach to capturing human knowledge in a way that is scalable and adequate to real-world needs. Human Computation has begun to scientifically study how human intelligence at scale can be used to methodologically improve machine-based knowledge and data management. My research is focusing on understanding human computation for improving how machine-based systems can acquire, capture and harness human knowledge and thus become even more intelligent. In this talk I will show how the CrowdTruth framework (http://crowdtruth.org) facilitates data collection, processing and analytics of human computation knowledge.
Some project links:
- http://controcurator.org/
- http://crowdtruth.org/
- http://diveproject.beeldengeluid.nl/
- http://vu-amsterdam-web-media-group.github.io/linkflows/
State of ICS and IoT Cyber Threat Landscape Report 2024 previewPrayukth K V
The IoT and OT threat landscape report has been prepared by the Threat Research Team at Sectrio using data from Sectrio, cyber threat intelligence farming facilities spread across over 85 cities around the world. In addition, Sectrio also runs AI-based advanced threat and payload engagement facilities that serve as sinks to attract and engage sophisticated threat actors, and newer malware including new variants and latent threats that are at an earlier stage of development.
The latest edition of the OT/ICS and IoT security Threat Landscape Report 2024 also covers:
State of global ICS asset and network exposure
Sectoral targets and attacks as well as the cost of ransom
Global APT activity, AI usage, actor and tactic profiles, and implications
Rise in volumes of AI-powered cyberattacks
Major cyber events in 2024
Malware and malicious payload trends
Cyberattack types and targets
Vulnerability exploit attempts on CVEs
Attacks on counties – USA
Expansion of bot farms – how, where, and why
In-depth analysis of the cyber threat landscape across North America, South America, Europe, APAC, and the Middle East
Why are attacks on smart factories rising?
Cyber risk predictions
Axis of attacks – Europe
Systemic attacks in the Middle East
Download the full report from here:
https://sectrio.com/resources/ot-threat-landscape-reports/sectrio-releases-ot-ics-and-iot-security-threat-landscape-report-2024/
Software Delivery At the Speed of AI: Inflectra Invests In AI-Powered QualityInflectra
In this insightful webinar, Inflectra explores how artificial intelligence (AI) is transforming software development and testing. Discover how AI-powered tools are revolutionizing every stage of the software development lifecycle (SDLC), from design and prototyping to testing, deployment, and monitoring.
Learn about:
• The Future of Testing: How AI is shifting testing towards verification, analysis, and higher-level skills, while reducing repetitive tasks.
• Test Automation: How AI-powered test case generation, optimization, and self-healing tests are making testing more efficient and effective.
• Visual Testing: Explore the emerging capabilities of AI in visual testing and how it's set to revolutionize UI verification.
• Inflectra's AI Solutions: See demonstrations of Inflectra's cutting-edge AI tools like the ChatGPT plugin and Azure Open AI platform, designed to streamline your testing process.
Whether you're a developer, tester, or QA professional, this webinar will give you valuable insights into how AI is shaping the future of software delivery.
Securing your Kubernetes cluster_ a step-by-step guide to success !KatiaHIMEUR1
Today, after several years of existence, an extremely active community and an ultra-dynamic ecosystem, Kubernetes has established itself as the de facto standard in container orchestration. Thanks to a wide range of managed services, it has never been so easy to set up a ready-to-use Kubernetes cluster.
However, this ease of use means that the subject of security in Kubernetes is often left for later, or even neglected. This exposes companies to significant risks.
In this talk, I'll show you step-by-step how to secure your Kubernetes cluster for greater peace of mind and reliability.
Key Trends Shaping the Future of Infrastructure.pdfCheryl Hung
Keynote at DIGIT West Expo, Glasgow on 29 May 2024.
Cheryl Hung, ochery.com
Sr Director, Infrastructure Ecosystem, Arm.
The key trends across hardware, cloud and open-source; exploring how these areas are likely to mature and develop over the short and long-term, and then considering how organisations can position themselves to adapt and thrive.
Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...UiPathCommunity
💥 Speed, accuracy, and scaling – discover the superpowers of GenAI in action with UiPath Document Understanding and Communications Mining™:
See how to accelerate model training and optimize model performance with active learning
Learn about the latest enhancements to out-of-the-box document processing – with little to no training required
Get an exclusive demo of the new family of UiPath LLMs – GenAI models specialized for processing different types of documents and messages
This is a hands-on session specifically designed for automation developers and AI enthusiasts seeking to enhance their knowledge in leveraging the latest intelligent document processing capabilities offered by UiPath.
Speakers:
👨🏫 Andras Palfi, Senior Product Manager, UiPath
👩🏫 Lenka Dulovicova, Product Program Manager, UiPath
GraphRAG is All You need? LLM & Knowledge GraphGuy Korland
Guy Korland, CEO and Co-founder of FalkorDB, will review two articles on the integration of language models with knowledge graphs.
1. Unifying Large Language Models and Knowledge Graphs: A Roadmap.
https://arxiv.org/abs/2306.08302
2. Microsoft Research's GraphRAG paper and a review paper on various uses of knowledge graphs:
https://www.microsoft.com/en-us/research/blog/graphrag-unlocking-llm-discovery-on-narrative-private-data/
Neuro-symbolic is not enough, we need neuro-*semantic*Frank van Harmelen
Neuro-symbolic (NeSy) AI is on the rise. However, simply machine learning on just any symbolic structure is not sufficient to really harvest the gains of NeSy. These will only be gained when the symbolic structures have an actual semantics. I give an operational definition of semantics as “predictable inference”.
All of this illustrated with link prediction over knowledge graphs, but the argument is general.
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...DanBrown980551
Do you want to learn how to model and simulate an electrical network from scratch in under an hour?
Then welcome to this PowSyBl workshop, hosted by Rte, the French Transmission System Operator (TSO)!
During the webinar, you will discover the PowSyBl ecosystem as well as handle and study an electrical network through an interactive Python notebook.
PowSyBl is an open source project hosted by LF Energy, which offers a comprehensive set of features for electrical grid modelling and simulation. Among other advanced features, PowSyBl provides:
- A fully editable and extendable library for grid component modelling;
- Visualization tools to display your network;
- Grid simulation tools, such as power flows, security analyses (with or without remedial actions) and sensitivity analyses;
The framework is mostly written in Java, with a Python binding so that Python developers can access PowSyBl functionalities as well.
What you will learn during the webinar:
- For beginners: discover PowSyBl's functionalities through a quick general presentation and the notebook, without needing any expert coding skills;
- For advanced developers: master the skills to efficiently apply PowSyBl functionalities to your real-world scenarios.
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
Accelerate your Kubernetes clusters with Varnish CachingThijs Feryn
A presentation about the usage and availability of Varnish on Kubernetes. This talk explores the capabilities of Varnish caching and shows how to use the Varnish Helm chart to deploy it to Kubernetes.
This presentation was delivered at K8SUG Singapore. See https://feryn.eu/presentations/accelerate-your-kubernetes-clusters-with-varnish-caching-k8sug-singapore-28-2024 for more details.
JMeter webinar - integration with InfluxDB and GrafanaRTTS
Watch this recorded webinar about real-time monitoring of application performance. See how to integrate Apache JMeter, the open-source leader in performance testing, with InfluxDB, the open-source time-series database, and Grafana, the open-source analytics and visualization application.
In this webinar, we will review the benefits of leveraging InfluxDB and Grafana when executing load tests and demonstrate how these tools are used to visualize performance metrics.
Length: 30 minutes
Session Overview
-------------------------------------------
During this webinar, we will cover the following topics while demonstrating the integrations of JMeter, InfluxDB and Grafana:
- What out-of-the-box solutions are available for real-time monitoring JMeter tests?
- What are the benefits of integrating InfluxDB and Grafana into the load testing stack?
- Which features are provided by Grafana?
- Demonstration of InfluxDB and Grafana using a practice web application
To view the webinar recording, go to:
https://www.rttsweb.com/jmeter-integration-webinar
Builder.ai Founder Sachin Dev Duggal's Strategic Approach to Create an Innova...Ramesh Iyer
In today's fast-changing business world, Companies that adapt and embrace new ideas often need help to keep up with the competition. However, fostering a culture of innovation takes much work. It takes vision, leadership and willingness to take risks in the right proportion. Sachin Dev Duggal, co-founder of Builder.ai, has perfected the art of this balance, creating a company culture where creativity and growth are nurtured at each stage.
Kubernetes & AI - Beauty and the Beast !?! @KCD Istanbul 2024Tobias Schneck
As AI technology is pushing into IT I was wondering myself, as an “infrastructure container kubernetes guy”, how get this fancy AI technology get managed from an infrastructure operational view? Is it possible to apply our lovely cloud native principals as well? What benefit’s both technologies could bring to each other?
Let me take this questions and provide you a short journey through existing deployment models and use cases for AI software. On practical examples, we discuss what cloud/on-premise strategy we may need for applying it to our own infrastructure to get it to work from an enterprise perspective. I want to give an overview about infrastructure requirements and technologies, what could be beneficial or limiting your AI use cases in an enterprise environment. An interactive Demo will give you some insides, what approaches I got already working for real.
1. The Rijksmuseum Collection
as Linked Data
Chris Dijkshoorn , Lora Aroyo,
Jacco van Ossenbruggen,
Guus Schreiber, Wesley ter Weele,
Jan Wielemaker
Lizzy Jongma
http://www.semantic-web-journal.net/content/rijksmuseum-collection-linked-data-2
@laroyo
@LizzyJongma
@rasvaan
2. Open up data silos
‣ Improve reusability data
‣ Support integration collections
‣ Identifiers for things
‣ Cross-referencing
‣ Lins across collections
‣ Shared views & context of objects
‣ Data models for interoperability
Researchers & Collection Managers
using it for deep analysis of
objects and collections as a whole
Linked Data in
Cultural Heritage
3. Collection
‣ Collection of ~1,000,000 objects
‣ Artworks on display ~8.000
‣ Dutch Masters like Rembrandt
Online Collection
‣ Accessible through API
‣ 597,193 object records
‣ 207,441 works have CC0 image
Images are released in the public
domain for users & developers
https://www.rijksmuseum.nl/en/api
Rijksmuseum Amsterdam
4. Professional catalogers and
photographers
‣ Register artworks
‣ Provide annotations
‣ Digitise artworks
‣ Publish them online
~40,000 new object records a year
time consuming & costly
endeavour
Versioning of data
Digitisation projects
5. Collection Management System
Rijksmuseum
Content Management
System 597 fields
Rijksmuseum
Collection Data
597,193 objects
Rijksmuseum
API
XSLT
exporting
XML
XML
identifying
fields
Data from collection management is harvested daily &
loaded in a database serving the website
6. Website
Website 245 fields
Website Data
597,193 objects
Rijksmuseum
Content Management
System 597 fields
Rijksmuseum
Collection Data
597,193 objects
Rijksmuseum Regular user
daily
JSONrequest
API
JSON
export
XSLT
exporting
XML
Only CC0
Developer
API
XSLT
exporting
XML
XML
identifying
fields
• A subset of 245 metadata fields (597 in total) are included in the output
of collection management
• Fields no longer used or contain sensitive data, e.g. insurance values
are excluded
• The selected fields are transformed to form field names which better
reflect their content, omit empty values and generate links to other
databases maintained by the Rijksmuseum (XSLT)
7. Conversion to Linked Data
Website 245 fields
Website Data
597,193 objects
Rijksmuseum
Content Management
System 597 fields
Rijksmuseum
Collection Data
597,193 objects
Rijksmuseum Regular user
daily
JSONrequest
request
API
JSON
export
XSLT
exporting
XML
Only CC0
Developer
Triple Store 15 fields
Researcher
RDF EDM
15 fields
API
XSLT
exporting
XML
XML
identifying
fields
Rijksmuseum
Linked Data
351,814 objects
Relevant metadata fields of a collection object are
mapped to the Europeana Data Model that most
closely resembles the values of the field.
The output of the API is used to obtain a complete
harvest of the data, which is in turn loaded into a
triple store (run on a monthly basis with links to
downloads of older versioned datadumps)
8. Conversion to Linked Data
Website 245 fields
Website Data
597,193 objects
Rijksmuseum
Content Management
System 597 fields
Rijksmuseum
Collection Data
597,193 objects
Rijksmuseum Regular user
daily
JSONrequest
request
API
JSON
export
XSLT
exporting
XML
Only CC0
Developer
Triple Store 15 fields
Researcher
RDF EDM
15 fields
API
XSLT
exporting
XML
XML
identifying
fields
Rijksmuseum
Linked Data
351,814 objects
modelling the complete collection &
integrating it with other collections from
other institutions required the ability to
model different (potentially conflicting)
metadata records from different sources
describing the same artwork
9. Europeana Data Model
ProvidedCHO
SK-A-3276
"Jeremiah Lamenting the
Destruction of Jerusalem"@en
"Rembrandt
Harmensz.
van Rijn"
title
aggregated
CHO
creator
aggregation
COL.5242
Agent
PEOPLE.5706
isShownBy
pref
Label
"Rijksmuseum"
data
Provider
WebResource
The Rijksmuseum dataset was one of the first entries in the Europeana Thought Lab
Images converted to comply with the VRA data model, 46K
The data model is designed with reuse of existing classes and properties in mind. It includes
elements from the Dublin Core metadata initiative and the Object Reuse and Exchange
definition of the Open Archives Initiative.
three core classes:
• edm:ProvidedCHO for
cultural heritage objects
• edm:WebResource for
web resources
• ore:Aggregation for
aggregations of
resources
properties:
• dc:creator
• dc:title
• dc:format
• dc:subject
10. Iconclass
‣ Concepts about subjects,
themes and motifs in Western art
‣ Links artworks to subject
Art & Architecture Thesaurus (AAT)
‣ Concepts about art styles,
materials and agents
‣ Links artworks to type and format
Short-Title catalogue Netherlands
(STCN)
‣ retrospective national bibliography of the
Netherlands maintained by the National
Library of the Netherlands.
‣ books that are the source of objects in the
print collection of the Rijksmuseum
Links to external datasets
11. Links to external datasets
"Rijksmuseum"
ProvidedCHO
SK-A-3276
Concept
71O77
"Jeremiah Lamenting the
Destruction of Jerusalem"@en
prefLabel
"Jeremiah lamenting over the
destruction of Jerusalem"@en
broader
Concept
300015050
prefLabel
concept
1000014078-en
"Rembrandt Harmensz. van Rijn"
Vocabularies
title
aggregated
CHO
creator
aggregation
COL.5242
Agent
PEOPLE.5706
isShownBy
format
Concept
71
prefLabel
"Old
Testament"@en
prefLabel
term
"oil paint"@en
dataProvider
WebResource
subject
12. Dataset
stats
22,846,996 triples
describing 351,814 objects
207,441 with graphical depiction
Ten sub-collections are maintained:
• sculptures (29,782 objects)
• historical items (19,936 objects)
• paintings (3,949 objects)
• Asian art (3,722 objects)
• prints, drawings & photos (280,047 objects)
13. Frequency distributions of the top 50 concepts of
AAT & Iconclass in Rijksmuseum collection
A small subset of concepts is often used:
• 305 distinct formats
• 124 distinct types
• prints (183,916)
• stereoscopic photographs (3,480)
• plates (1,617)
• art styles are often debatable
Many concepts are often used (ave ~ 27 times):
• 39,578 concepts in the vocabulary
• 10,434 are used to add information to an object
• 351,814 collection objects
• 172,059 have one or more Iconclass annotations
14. Focus on art-historical
information
Occasional lack of expertise
regarding subject matter
annotations
This print is described as:
‣ “Bird with blue head”
‣ “Branch with red leaves”
Annotating Artworks
15. Create links using
Accurator annotation tool
http://annotation.accurator.nl/
Organise annotation events
‣ Bird watching event
‣ Fashion event
Experts are adding
information
16. Publishing data widens the type
of users involved
Engage in a dialogue
‣ What information is needed?
‣ Which vocabularies to use?
‣ Which fields can be used to
describe the objects?
Dialogue about data
21. Many prints originate from books
‣ References to these books are added as
curators comments
Short-Title catalogue Netherlands
‣ Retrospective national bibliography in
the period 1540-1800
‣ Includes 139,817 publications
Linking books to prints
‣ Scan for curator comments containing
Title, Author and Year
‣ 3598 links from prints to 501
publications
Linking to the National Library
25. All at once
monthly datadumps
https://datahub.io/dataset/rijksmuseum
Request based
OAI API
https://www.rijksmuseum.nl/en/api/
rijksmuseum-oai-api-instructions-for-use
Queries
SPARQL Endpoint
https://datahub.io/dataset/rijksmuseum
How to use the data
26. The Rijksmuseum Collection
as Linked Data
http://www.semantic-web-journal.net/content/rijksmuseum-collection-linked-data-2
Chris Dijkshoorn , Lora Aroyo,
Jacco van Ossenbruggen,
Guus Schreiber, Wesley ter Weele,
Jan Wielemaker
Lizzy Jongma
@laroyo
@LizzyJongma
@rasvaan