Triplify is a tool that publishes semantic data from relational databases on the web as Linked Data. It works by mapping SQL queries to RDF representations. The SQL queries select structured data from databases behind existing web applications. Triplify then converts the query results into RDF triples. This exposes the semantics behind web applications and makes the data accessible to semantic search engines and applications. Triplify aims to overcome the lack of semantic data on the web by leveraging existing relational data sources.
This session will first explain the index related options that are available when developing a data model and how these choices affect indexing and searching. We will cover Alfresco FTS in detail, and compare SQL 92 with the CMIS QL. We will also consider sorting and other ways to control the results returned, and how query performance may be affected by ACL evaluation.
Features of Alfresco Search Services.
Features of Alfresco Search & Insight Engine.
Future plans for the product
---
DEMO GUIDE
[1] Queries: Share > Node Browser
ASPECT:'cm:titled' AND cm:title:'*Sample*' AND TEXT:'code'
SELECT * FROM cm:titled WHERE cm:title like '%Sample%' AND CONTAINS('code')
[2] Queries: Share > JS Console
var ctxt = Packages.org.springframework.web.context.ContextLoader.getCurrentWebApplicationContext();
var searchService = ctxt.getBean('SearchService', org.alfresco.service.cmr.search.SearchService);
var StoreRef = Packages.org.alfresco.service.cmr.repository.StoreRef;
var SearchService = Packages.org.alfresco.service.cmr.search.SearchService;
var ResultSet = Packages.org.alfresco.repo.search.impl.lucene.SolrJSONResultSet;
ResultSet =
searchService.query(
StoreRef.STORE_REF_WORKSPACE_SPACESSTORE,
SearchService.LANGUAGE_FTS_ALFRESCO,
"ASPECT:'cm:titled' AND cm:title:'*Sample*' AND TEXT:'code'");
logger.log(ResultSet.getNodeRefs());
---
var ctxt = Packages.org.springframework.web.context.ContextLoader.getCurrentWebApplicationContext();
var searchService = ctxt.getBean('SearchService', org.alfresco.service.cmr.search.SearchService);
var StoreRef = Packages.org.alfresco.service.cmr.repository.StoreRef;
var SearchService = Packages.org.alfresco.service.cmr.search.SearchService;
var ResultSet = Packages.org.alfresco.repo.search.impl.lucene.SolrJSONResultSet;
ResultSet =
searchService.query(
StoreRef.STORE_REF_WORKSPACE_SPACESSTORE,
SearchService.LANGUAGE_CMIS_ALFRESCO,
"SELECT * FROM cm:titled WHERE cm:title like '%Sample%' AND CONTAINS('code')");
logger.log(ResultSet.getNodeRefs());
---
var def =
{
query: "ASPECT:'cm:titled' AND cm:title:'*Sample*' AND TEXT:'code'",
language: "fts-alfresco"
};
var results = search.query(def);
logger.log(results);
[3] Queries: api-explorer
{
"query": {
"language": "afts",
"query": "ASPECT:\"cm:titled\" AND cm:title:\"*Sample\" AND TEXT:\"code\""
}
}
---
{
"query": {
"language": "cmis",
"query": "SELECT * FROM cm:titled WHERE cm:title like '%Sample%' AND CONTAINS('code')"
}
}
[4] Queries: CMIS Workbench > Groovy Console
rs = session.query("SELECT * FROM cm:titled WHERE cm:title like '%Sample%' AND CONTAINS('code')", false)
for (res in rs) {
println(res.getPropertyValueById('cmis:objectId'))
}
[5] Queries: SOLR Web Console > (alfresco) > Query
/afts
ASPECT:'cm:titled' AND cm:title:'*Sample*' AND TEXT:'code'
---
/cmis
SELECT * FROM cm:titled WHERE cm:title like '%Sample%' AND CONTAINS('code')
---
This session will first explain the index related options that are available when developing a data model and how these choices affect indexing and searching. We will cover Alfresco FTS in detail, and compare SQL 92 with the CMIS QL. We will also consider sorting and other ways to control the results returned, and how query performance may be affected by ACL evaluation.
Features of Alfresco Search Services.
Features of Alfresco Search & Insight Engine.
Future plans for the product
---
DEMO GUIDE
[1] Queries: Share > Node Browser
ASPECT:'cm:titled' AND cm:title:'*Sample*' AND TEXT:'code'
SELECT * FROM cm:titled WHERE cm:title like '%Sample%' AND CONTAINS('code')
[2] Queries: Share > JS Console
var ctxt = Packages.org.springframework.web.context.ContextLoader.getCurrentWebApplicationContext();
var searchService = ctxt.getBean('SearchService', org.alfresco.service.cmr.search.SearchService);
var StoreRef = Packages.org.alfresco.service.cmr.repository.StoreRef;
var SearchService = Packages.org.alfresco.service.cmr.search.SearchService;
var ResultSet = Packages.org.alfresco.repo.search.impl.lucene.SolrJSONResultSet;
ResultSet =
searchService.query(
StoreRef.STORE_REF_WORKSPACE_SPACESSTORE,
SearchService.LANGUAGE_FTS_ALFRESCO,
"ASPECT:'cm:titled' AND cm:title:'*Sample*' AND TEXT:'code'");
logger.log(ResultSet.getNodeRefs());
---
var ctxt = Packages.org.springframework.web.context.ContextLoader.getCurrentWebApplicationContext();
var searchService = ctxt.getBean('SearchService', org.alfresco.service.cmr.search.SearchService);
var StoreRef = Packages.org.alfresco.service.cmr.repository.StoreRef;
var SearchService = Packages.org.alfresco.service.cmr.search.SearchService;
var ResultSet = Packages.org.alfresco.repo.search.impl.lucene.SolrJSONResultSet;
ResultSet =
searchService.query(
StoreRef.STORE_REF_WORKSPACE_SPACESSTORE,
SearchService.LANGUAGE_CMIS_ALFRESCO,
"SELECT * FROM cm:titled WHERE cm:title like '%Sample%' AND CONTAINS('code')");
logger.log(ResultSet.getNodeRefs());
---
var def =
{
query: "ASPECT:'cm:titled' AND cm:title:'*Sample*' AND TEXT:'code'",
language: "fts-alfresco"
};
var results = search.query(def);
logger.log(results);
[3] Queries: api-explorer
{
"query": {
"language": "afts",
"query": "ASPECT:\"cm:titled\" AND cm:title:\"*Sample\" AND TEXT:\"code\""
}
}
---
{
"query": {
"language": "cmis",
"query": "SELECT * FROM cm:titled WHERE cm:title like '%Sample%' AND CONTAINS('code')"
}
}
[4] Queries: CMIS Workbench > Groovy Console
rs = session.query("SELECT * FROM cm:titled WHERE cm:title like '%Sample%' AND CONTAINS('code')", false)
for (res in rs) {
println(res.getPropertyValueById('cmis:objectId'))
}
[5] Queries: SOLR Web Console > (alfresco) > Query
/afts
ASPECT:'cm:titled' AND cm:title:'*Sample*' AND TEXT:'code'
---
/cmis
SELECT * FROM cm:titled WHERE cm:title like '%Sample%' AND CONTAINS('code')
---
http://tinyurl.com/4u5awh <-- View recorded webinar. This presentation covers how Alfresco can help users easily find information managed within and across Alfresco repositories, 3rd party applications and the Web.
Ian Robinson, Engineer at Neo4j, talks about how you productionize your Neo4j-based application. In this talk from GraphConnect San Francisco 2015, he looks at some of the most important considerations around designing, building and operating a Neo4j app.
Topics include:
* Where Neo4j fits in your application architecture, in both its embedded and server modes
* How to configure its clustering for high availability and high read throughput
* Backup strategies
* The new monitoring capabilities in Neo4j 2.3
Deep Dive: Alfresco Core Repository (... embedded in a micro-services style a...J V
Alfresco Summit 2014 (London)
Though best practice is to leverage Alfresco through the well defined API's, it can be useful to understand the internals of the repository so that your development efforts are the most effective. A deep understanding of the repository will help you to evaluate performance bottlenecks, look for bugs, or make contributions. This session provides an overview of the repository internals, including the major components, the key services, subsystems, and database. We then provide an example where we leverage the repository in a micro-service architecture while building Alfresco's future cloud products and show how the different parts of the repository interact to fulfill requests.
http://summit.alfresco.com/london/sessions/diving-deep-alfresco-repository
https://www.youtube.com/watch?v=TAE9UjC0xxc
Cross Datacenter Replication aka CDCR has been a long requested feature in Apache Solr. In this talk, we will discuss CDCR as released in Apache Solr 6.0 and beyond to understand its use-cases, limitations, setup and performance. We will also take a quick look at the future enhancements that can further simplify and scale this feature.
OSGi for real in the enterprise: Apache Karaf - NLJUG J-FALL 2010Adrian Trenaman
Want to know how to design, implement and deploy modular enterprise integration solutions using OSGi? The Apache Karaf OSGi shell, used by Apache Felix and Apache ServiceMix, enhances core OSGi implementations like Felix or Equinox with an easy to use, extendible command shell, providing logging, hot deployment, configuration, container administration, clustering, high availability and easy 'feature-based' dependency management In this session, you'll learn how Karaf works, and how you can leverage Karaf either on its own or embedded within ServiceMix to deploy business logic, RESTful services, EIP-based integration flows and web services. You'll learn how to extend the command shell with your own commands, and, use Spring-DM *or* OSGi BluePrint Services to make using OSGi a walk in the park.
Kerberos is the system which underpins the vast majority of strong authentication across the Apache HBase/Hadoop application stack. Kerberos errors have brought many to their knees and it is often referred to as “black magic” or “the dark arts”; a long-standing joke that there are so few who understand how it works. This talk will cover the types of problems that Kerberos solves and doesn’t solve for HBase, decrypt some jargon on related libraries and technology that enable Kerberos authentication in HBase and Hadoop, and distill some basic takeaways designed to ease users in developing an application that can securely communicate with a “kerberized” HBase installation.
Apache is the most popular web server software. The course is geared to make you a Apache pro. Once you digest all basics, the course will help you create your very own production Web Server!
http://tinyurl.com/4u5awh <-- View recorded webinar. This presentation covers how Alfresco can help users easily find information managed within and across Alfresco repositories, 3rd party applications and the Web.
Ian Robinson, Engineer at Neo4j, talks about how you productionize your Neo4j-based application. In this talk from GraphConnect San Francisco 2015, he looks at some of the most important considerations around designing, building and operating a Neo4j app.
Topics include:
* Where Neo4j fits in your application architecture, in both its embedded and server modes
* How to configure its clustering for high availability and high read throughput
* Backup strategies
* The new monitoring capabilities in Neo4j 2.3
Deep Dive: Alfresco Core Repository (... embedded in a micro-services style a...J V
Alfresco Summit 2014 (London)
Though best practice is to leverage Alfresco through the well defined API's, it can be useful to understand the internals of the repository so that your development efforts are the most effective. A deep understanding of the repository will help you to evaluate performance bottlenecks, look for bugs, or make contributions. This session provides an overview of the repository internals, including the major components, the key services, subsystems, and database. We then provide an example where we leverage the repository in a micro-service architecture while building Alfresco's future cloud products and show how the different parts of the repository interact to fulfill requests.
http://summit.alfresco.com/london/sessions/diving-deep-alfresco-repository
https://www.youtube.com/watch?v=TAE9UjC0xxc
Cross Datacenter Replication aka CDCR has been a long requested feature in Apache Solr. In this talk, we will discuss CDCR as released in Apache Solr 6.0 and beyond to understand its use-cases, limitations, setup and performance. We will also take a quick look at the future enhancements that can further simplify and scale this feature.
OSGi for real in the enterprise: Apache Karaf - NLJUG J-FALL 2010Adrian Trenaman
Want to know how to design, implement and deploy modular enterprise integration solutions using OSGi? The Apache Karaf OSGi shell, used by Apache Felix and Apache ServiceMix, enhances core OSGi implementations like Felix or Equinox with an easy to use, extendible command shell, providing logging, hot deployment, configuration, container administration, clustering, high availability and easy 'feature-based' dependency management In this session, you'll learn how Karaf works, and how you can leverage Karaf either on its own or embedded within ServiceMix to deploy business logic, RESTful services, EIP-based integration flows and web services. You'll learn how to extend the command shell with your own commands, and, use Spring-DM *or* OSGi BluePrint Services to make using OSGi a walk in the park.
Kerberos is the system which underpins the vast majority of strong authentication across the Apache HBase/Hadoop application stack. Kerberos errors have brought many to their knees and it is often referred to as “black magic” or “the dark arts”; a long-standing joke that there are so few who understand how it works. This talk will cover the types of problems that Kerberos solves and doesn’t solve for HBase, decrypt some jargon on related libraries and technology that enable Kerberos authentication in HBase and Hadoop, and distill some basic takeaways designed to ease users in developing an application that can securely communicate with a “kerberized” HBase installation.
Apache is the most popular web server software. The course is geared to make you a Apache pro. Once you digest all basics, the course will help you create your very own production Web Server!
I used these slides for an introductory lecture (90min) to a seminar on SPARQL. This slideset introduces the RDF query language SPARQL from a user's perspective.
"SPARQL Cheat Sheet" is a short collection of slides intended to act as a guide to SPARQL developers. It includes the syntax and structure of SPARQL queries, common SPARQL prefixes and functions, and help with RDF datasets.
The "SPARQL Cheat Sheet" is intended to accompany the SPARQL By Example slides available at http://www.cambridgesemantics.com/2008/09/sparql-by-example/ .
The Semantic Web is about to grow up. By efforts such as the Linked Open Data initiative, we finally find ourselves at the edge of a Web of Data becoming reality. Standards such as OWL 2, RIF and SPARQL 1.1 shall allow us to reason with and ask complex structured queries on this data, but still they do not play together smoothly and robustly enough to cope with huge amounts of noisy Web data. In this talk, we discuss open challenges relating to querying and reasoning with Web data and raise the question: can the emerging Web of Data ever catch up with the now ubiquitous HTML Web?
An introduction deck for the Web of Data to my team, including basic semantic web, Linked Open Data, primer, and then DBpedia, Linked Data Integration Framework (LDIF), Common Crawl Database, Web Data Commons.
This presentation shall address the web2py web framework, my favorite way to develop web apps.
web2py is a free, open-source web framework for agile development of secure database-driven web applications; it is written in Python and programmable in Python. web2py is a full-stack framework, meaning that it contains all the components you need to build fully functional web applications.
Ease of use is the primary goal for web2py. For us, this means reducing the learning and deployment time. This is why web2py is a full-stack framework without dependencies. It requires no installation and has no configuration files. Everything works out of the box, including a web server, database and a web-based IDE that gives access to all the main features.
I will show you why web2py can make you more productive by bringing the result of a reflection over the best ideas of the most popular MVC based web frameworks enforcing the best practices for a fast, scalable and secure web application with minimal effort. There will be a live demo where you can get a faster grasp on how does it work and how fun it can be.
For more: www.web2py.com
Building API's for a web 2.0 / web 3.0 aspiring service is very different than providing a tight integrated RPC service for some corporate client. It requires completely different ways of thinking and embracing new standards. I've composed a quick slideshow of all the architectural choices and considerations I've come across.
Open Device Programmability: Hands-on Intro to RESTCONF (and a bit of NETCONF)Cisco DevNet
In this small group, hands-on workshop session you'll learn how to write your first Python application that uses YANG, NETCONF and , RESTCONF to access operational and configuration data on a device.
Watch the DevNet 2044 replay from the Cisco Live On-Demand Library at: https://www.ciscolive.com/online/connect/sessionDetail.ww?SESSION_ID=92725&backBtn=true
Check out more and register for Cisco DevNet: http://ow.ly/jCNV3030OfS
More and more web projects require interfacing with the backend using a REST-ful interface. In this presentation we'll discuss Django-REST-Framework's features and walk through how to integrate it into your next project!
My presentation on RDFauthor at EKAW2010, Lisbon. For more information on RDFauthor visit http://aksw.org/Projects/RDFauthor; for the code visit http://code.google.com/p/rdfauthor/.
Similar to WWW09 - Triplify Light-Weight Linked Data Publication from Relational Databases (20)
Knowledge Graph Research and Innovation ChallengesSören Auer
Gives an overview on some challenges regarding the combination of machine-learning and knowledge graph technologies and the vision of devising a concept of Cognitive Knowledge Graphs consisting of graphlets instead of mere entity descriptions.
Towards Knowledge Graph based Representation, Augmentation and Exploration of...Sören Auer
Despite an improved digital access to scientific publications in the last decades, the fundamental principles of scholarly communication remain unchanged and continue to be largely document-based. The document-oriented workflows in science have reached the limits of adequacy as highlighted by recent discussions on the increasing proliferation of scientific literature, the deficiency of peer-review and the reproducibility crisis. We need to represent, analyse, augment and exploit scholarly communication in a knowledge-based way by expressing and linking scientific contributions and related artefacts through semantically rich, interlinked knowledge graphs. This should be based
on deep semantic representation of scientific contributions, their manual, crowd-sourced and automatic augmentation and finally the intuitive exploration and interaction employing question answering on the resulting scientific knowledge base. We need to synergistically combine automated extraction and augmentation techniques, with large-scale collaboration to reach an unprecedented level of knowledge graph breadth and depth. As a result, knowledge-based information flows can facilitate completely new ways of search and exploration. The efficiency and effectiveness of scholarly communication will significant increase, since ambiguities are reduced, reproducibility is facilitated, redundancy is avoided, provenance and contributions can be better traced and the interconnections of research contributions are made more explicit and transparent. In this talk we will present first steps in this direction in the context of our Open Research Knowledge Graph initiative and the ScienceGRAPH project.
Slides of my talk at OSLCfest in Stockholm Nov 6, 2019
Video recording of the talk is available here:
https://www.facebook.com/oslcfest/videos/2261640397437958/
Towards an Open Research Knowledge GraphSören Auer
The document-oriented workflows in science have reached (or already exceeded) the limits of adequacy as highlighted for example by recent discussions on the increasing proliferation of scientific literature and the reproducibility crisis. Now it is possible to rethink this dominant paradigm of document-centered knowledge exchange and transform it into knowledge-based information flows by representing and expressing knowledge through semantically rich, interlinked knowledge graphs. The core of the establishment of knowledge-based information flows is the creation and evolution of information models for the establishment of a common understanding of data and information between the various stakeholders as well as the integration of these technologies into the infrastructure and processes of search and knowledge exchange in the research library of the future. By integrating these information models into existing and new research infrastructure services, the information structures that are currently still implicit and deeply hidden in documents can be made explicit and directly usable. This has the potential to revolutionize scientific work because information and research results can be seamlessly interlinked with each other and better mapped to complex information needs. Also research results become directly comparable and easier to reuse.
Towards digitizing scholarly communicationSören Auer
Slides of the VIVO 2016 Conference keynote: Despite the availability of ubiquitous connectivity and information technology, scholarly communication has not changed much in the last hundred years: research findings are still encoded in and decoded from linear, static articles and the possibilities of digitization are rarely used. In this talk, we will discuss strategies for digitizing scholarly communication. This comprises in particular: the use of machine-readable, dynamic content; the description and interlinking of research artifacts using Linked Data; the crowd-sourcing of multilingual
educational and learning content. We discuss the relation of these developments to research information systems and how they could become part of an open ecosystem for scholarly communication.
Linked data for Enterprise Data IntegrationSören Auer
The Web evolves into a Web of Data. In parallel Intranets of large companies will evolve into Data Intranets based on the Linked Data principles. Linked Data has the potential to complement the SOA paradigm with a light-weight, adaptive data integration approach.
Introduction to the Data Web, DBpedia and the Life-cycle of Linked DataSören Auer
Over the past 4 years, the Semantic Web activity has gained momentum with the widespread publishing of structured data as RDF. The Linked Data paradigm has therefore evolved from a practical research idea into
a very promising candidate for addressing one of the biggest challenges
of computer science: the exploitation of the Web as a platform for data
and information integration. To translate this initial success into a
world-scale reality, a number of research challenges need to be
addressed: the performance gap between relational and RDF data
management has to be closed, coherence and quality of data published on
the Web have to be improved, provenance and trust on the Linked Data Web
must be established and generally the entrance barrier for data
publishers and users has to be lowered. This tutorial will discuss
approaches for tackling these challenges. As an example of a successful
Linked Data project we will present DBpedia, which leverages Wikipedia
by extracting structured information and by making this information
freely accessible on the Web. The tutorial will also outline some recent advances in DBpedia, such as the mappings Wiki, DBpedia Live as well as
the recently launched DBpedia benchmark.
GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...Neo4j
Leonard Jayamohan, Partner & Generative AI Lead, Deloitte
This keynote will reveal how Deloitte leverages Neo4j’s graph power for groundbreaking digital twin solutions, achieving a staggering 100x performance boost. Discover the essential role knowledge graphs play in successful generative AI implementations. Plus, get an exclusive look at an innovative Neo4j + Generative AI solution Deloitte is developing in-house.
Removing Uninteresting Bytes in Software FuzzingAftab Hussain
Imagine a world where software fuzzing, the process of mutating bytes in test seeds to uncover hidden and erroneous program behaviors, becomes faster and more effective. A lot depends on the initial seeds, which can significantly dictate the trajectory of a fuzzing campaign, particularly in terms of how long it takes to uncover interesting behaviour in your code. We introduce DIAR, a technique designed to speedup fuzzing campaigns by pinpointing and eliminating those uninteresting bytes in the seeds. Picture this: instead of wasting valuable resources on meaningless mutations in large, bloated seeds, DIAR removes the unnecessary bytes, streamlining the entire process.
In this work, we equipped AFL, a popular fuzzer, with DIAR and examined two critical Linux libraries -- Libxml's xmllint, a tool for parsing xml documents, and Binutil's readelf, an essential debugging and security analysis command-line tool used to display detailed information about ELF (Executable and Linkable Format). Our preliminary results show that AFL+DIAR does not only discover new paths more quickly but also achieves higher coverage overall. This work thus showcases how starting with lean and optimized seeds can lead to faster, more comprehensive fuzzing campaigns -- and DIAR helps you find such seeds.
- These are slides of the talk given at IEEE International Conference on Software Testing Verification and Validation Workshop, ICSTW 2022.
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
In the rapidly evolving landscape of technologies, XML continues to play a vital role in structuring, storing, and transporting data across diverse systems. The recent advancements in artificial intelligence (AI) present new methodologies for enhancing XML development workflows, introducing efficiency, automation, and intelligent capabilities. This presentation will outline the scope and perspective of utilizing AI in XML development. The potential benefits and the possible pitfalls will be highlighted, providing a balanced view of the subject.
We will explore the capabilities of AI in understanding XML markup languages and autonomously creating structured XML content. Additionally, we will examine the capacity of AI to enrich plain text with appropriate XML markup. Practical examples and methodological guidelines will be provided to elucidate how AI can be effectively prompted to interpret and generate accurate XML markup.
Further emphasis will be placed on the role of AI in developing XSLT, or schemas such as XSD and Schematron. We will address the techniques and strategies adopted to create prompts for generating code, explaining code, or refactoring the code, and the results achieved.
The discussion will extend to how AI can be used to transform XML content. In particular, the focus will be on the use of AI XPath extension functions in XSLT, Schematron, Schematron Quick Fixes, or for XML content refactoring.
The presentation aims to deliver a comprehensive overview of AI usage in XML development, providing attendees with the necessary knowledge to make informed decisions. Whether you’re at the early stages of adopting AI or considering integrating it in advanced XML development, this presentation will cover all levels of expertise.
By highlighting the potential advantages and challenges of integrating AI with XML development tools and languages, the presentation seeks to inspire thoughtful conversation around the future of XML development. We’ll not only delve into the technical aspects of AI-powered XML development but also discuss practical implications and possible future directions.
Sudheer Mechineni, Head of Application Frameworks, Standard Chartered Bank
Discover how Standard Chartered Bank harnessed the power of Neo4j to transform complex data access challenges into a dynamic, scalable graph database solution. This keynote will cover their journey from initial adoption to deploying a fully automated, enterprise-grade causal cluster, highlighting key strategies for modelling organisational changes and ensuring robust disaster recovery. Learn how these innovations have not only enhanced Standard Chartered Bank’s data infrastructure but also positioned them as pioneers in the banking sector’s adoption of graph technology.
Communications Mining Series - Zero to Hero - Session 1DianaGray10
This session provides introduction to UiPath Communication Mining, importance and platform overview. You will acquire a good understand of the phases in Communication Mining as we go over the platform with you. Topics covered:
• Communication Mining Overview
• Why is it important?
• How can it help today’s business and the benefits
• Phases in Communication Mining
• Demo on Platform overview
• Q/A
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...SOFTTECHHUB
The choice of an operating system plays a pivotal role in shaping our computing experience. For decades, Microsoft's Windows has dominated the market, offering a familiar and widely adopted platform for personal and professional use. However, as technological advancements continue to push the boundaries of innovation, alternative operating systems have emerged, challenging the status quo and offering users a fresh perspective on computing.
One such alternative that has garnered significant attention and acclaim is Nitrux Linux 3.5.0, a sleek, powerful, and user-friendly Linux distribution that promises to redefine the way we interact with our devices. With its focus on performance, security, and customization, Nitrux Linux presents a compelling case for those seeking to break free from the constraints of proprietary software and embrace the freedom and flexibility of open-source computing.
Essentials of Automations: The Art of Triggers and Actions in FMESafe Software
In this second installment of our Essentials of Automations webinar series, we’ll explore the landscape of triggers and actions, guiding you through the nuances of authoring and adapting workspaces for seamless automations. Gain an understanding of the full spectrum of triggers and actions available in FME, empowering you to enhance your workspaces for efficient automation.
We’ll kick things off by showcasing the most commonly used event-based triggers, introducing you to various automation workflows like manual triggers, schedules, directory watchers, and more. Plus, see how these elements play out in real scenarios.
Whether you’re tweaking your current setup or building from the ground up, this session will arm you with the tools and insights needed to transform your FME usage into a powerhouse of productivity. Join us to discover effective strategies that simplify complex processes, enhancing your productivity and transforming your data management practices with FME. Let’s turn complexity into clarity and make your workspaces work wonders!
A tale of scale & speed: How the US Navy is enabling software delivery from l...sonjaschweigert1
Rapid and secure feature delivery is a goal across every application team and every branch of the DoD. The Navy’s DevSecOps platform, Party Barge, has achieved:
- Reduction in onboarding time from 5 weeks to 1 day
- Improved developer experience and productivity through actionable findings and reduction of false positives
- Maintenance of superior security standards and inherent policy enforcement with Authorization to Operate (ATO)
Development teams can ship efficiently and ensure applications are cyber ready for Navy Authorizing Officials (AOs). In this webinar, Sigma Defense and Anchore will give attendees a look behind the scenes and demo secure pipeline automation and security artifacts that speed up application ATO and time to production.
We will cover:
- How to remove silos in DevSecOps
- How to build efficient development pipeline roles and component templates
- How to deliver security artifacts that matter for ATO’s (SBOMs, vulnerability reports, and policy evidence)
- How to streamline operations with automated policy checks on container images
Climate Impact of Software Testing at Nordic Testing DaysKari Kakkonen
My slides at Nordic Testing Days 6.6.2024
Climate impact / sustainability of software testing discussed on the talk. ICT and testing must carry their part of global responsibility to help with the climat warming. We can minimize the carbon footprint but we can also have a carbon handprint, a positive impact on the climate. Quality characteristics can be added with sustainability, and then measured continuously. Test environments can be used less, and in smaller scale and on demand. Test techniques can be used in optimizing or minimizing number of tests. Test automation can be used to speed up testing.
Dr. Sean Tan, Head of Data Science, Changi Airport Group
Discover how Changi Airport Group (CAG) leverages graph technologies and generative AI to revolutionize their search capabilities. This session delves into the unique search needs of CAG’s diverse passengers and customers, showcasing how graph data structures enhance the accuracy and relevance of AI-generated search results, mitigating the risk of “hallucinations” and improving the overall customer journey.
How to Get CNIC Information System with Paksim Ga.pptxdanishmna97
Pakdata Cf is a groundbreaking system designed to streamline and facilitate access to CNIC information. This innovative platform leverages advanced technology to provide users with efficient and secure access to their CNIC details.
How to Get CNIC Information System with Paksim Ga.pptx
WWW09 - Triplify Light-Weight Linked Data Publication from Relational Databases
1. Triplify – Linked Data Publication
from Relational Databases
Sören Auer, Sebastian Dietzold, Jens Lehmann,
Sebastian Hellmann, David Aumueller
AKSW, Institut für Informatik
2. Growth of the Semantic Data Web
Still outpaced by the traditional Web
4/24/2009 Triplify – Linked Data Publication from Relational DBs 2
4. Triplify Motivation
Project Area Downloads
• overcome the chicken-and-egg phpBB discussion forum 235480
dilemma of missing semantic Gallery photo gallery 166005
representations and search XOOPS CMS 115807
Coppermine photo gallery 113854
facilities on the Web Typo3 CMS 63641
• Triplify leverages relational Liferay Portal Portal 39615
representations behind existing eGroupWare group ware 33865
Alfresco CMS 31914
Web applications: e107 CMS 19996
– often open-source, deployed Lifetype Blogging 16730
hundred thousand times Plone CMS 13993
– structure and semantics encoded Compiere ERP + CRM 13718
in relational database schemes WebCalendar Calendar 12832
Nucleus Blogging 12739
(behind Web apps) is not
Tikiwiki Wiki 6368
accessible to Web search engines,
Monthly Web application downloads at
mashups etc. Sourceforge
4/24/2009 Triplify – Linked Data Publication from Relational DBs 4
6. Solution overview
• SQL is the industry standard language for
relational transformations
• Extend SQL with a few syntactic constructs
opaque to the SQL query processor
• Map URL patterns to sets of SQL query patterns
• For a concrete URL request, replace placeholders
in the query patterns, execute the query
• Transform resulting relations into various RDF
serializations (multiple view to class approach)
4/24/2009 Triplify – Linked Data Publication from Relational DBs 6
7. Triplify Solution: SQL-SELECT queries
map relational data to RDF
Triplify Configuration:
• number of SQL queries selecting information, which should be made publicly
available.
Special SQL query result structure required (in order to convert results into RDF:
• first column must contain identifiers for generating instance URIs (i.e. the primary
key of DB table)
• column names are used to generate property URIs, renaming columns allows to
reuse properties from existing vocabularies such as Dublin Core, FOAF, SIOC
– e.g. SELECT id, name AS 'foaf:name' FROM users
• individual cells contain data values or references to other instances
(eventually constitute the objects of resulting triples)
4/24/2009 Triplify – Linked Data Publication from Relational DBs 7
8. Example: Wordpress Blog Posts
Associate the URL path fragment 'post‘ with a number of
SQL patterns:
http://blog.aksw.org/triplify/post/(xxx)
Object property
1 SELECT id, post_author AS 'sioc:has_creator->user',
post_title AS 'dc:title',
post_content AS 'sioc:content', post_date AS
'dcterms:modified^^xsd:dateTime‘,
post_modified AS 'dcterms:created^^xsd:dateTime'
FROM posts
Datatype property
WHERE post_status='publish‘ (AND id=xxx)
2 SELECT post_id id, tag_label AS 'tag:taggedWithTag‘
FROM post2tag INNER JOIN tag ON(post2tag.tag_id=tag.tag_id)
(WHERE id=xxx)
3 SELECT post_id id, category_id AS 'belongsToCategory->category‘
FROM post2cat
(WHERE id=xxx)
4/24/2009 Triplify – Linked Data Publication from Relational DBs 8
9. RDF Conversion
http://blog.aksw.org/triplify/post/1
1 id post_author post_title post_content post_date post_modified
1 5 New DBpedia release Today we released … 200810201635 200810201635
id tag:taggedWithTag id belogsToCategory
3
2
1 DBpedia 1 34
1 Release …
..
http://blog.aksw.org/triplify/post/1 sioc:has_creator http://blog.aksw.org/triplify/user/5
http://blog.aksw.org/triplify/post/1 dc:title “New DBpedia release”
http://blog.aksw.org/triplify/post/1 sioc:content “Today we released …”
http://blog.aksw.org/triplify/post/1 dcterms:modified “20081020T1635”^^xsd:dateTime
http://blog.aksw.org/triplify/post/1 dcterms:created “20081020T1635”^^xsd:dateTime
http://blog.aksw.org/triplify/post/1 tag:taggedWithTag “DBpedia”
http://blog.aksw.org/triplify/post/1 tag:taggedWithTag “Release”
http://blog.aksw.org/triplify/post/1 belongsToCategory http://blog.aksw.org/triplify/category/34
4/24/2009 Triplify – Linked Data Publication from Relational DBs 9
10. Triplify Implementation: Simplicity
• Expose semantics as simple as possible
– No (new) mapping languages – easy to learn
– Few lines of code – easy to plug-in
– Simple, reusable configurations
• Available for most popular Web app languages
– PHP (ready), Ruby/Python under development
• Works with most popular Web app DBs
– MySQL (extensively tested), PHP-PDO DBs (SQLite, Oracle,
DB2, MS SQL, PostgreSQL etc.) should work, not needed
for Virtuoso
• Triplify exposes RDF/Ntriples, LinkedData and
RDF/JSON
4/24/2009 Triplify – Linked Data Publication from Relational DBs 10
11. <?php
include('../wp-config.php');
Example Config
$triplify['namespaces']=array(
'vocabulary'=>'http://triplify.org/vocabulary/Wordpress/',
'foaf'=>'http://xmlns.com/foaf/0.1/',
… );
$triplify['queries']=array(
'post'=>array(
quot;SELECT id,post_author 'sioc:has_creator->user',post_date 'dcterms:created',post_title 'dc:title', post_content 'sioc:content',
post_modified 'dcterms:modified‘ FROM {$table_prefix}posts WHERE post_status='publish'quot;,
quot;SELECT post_id id,tag_id 'tag:taggedWithTag' FROM {$table_prefix}post2tagquot;,
quot;SELECT post_id id,category_id 'belongsToCategory' FROM {$table_prefix}post2catquot;,
),
'tag'=>quot;SELECT tag_ID id,tag 'tag:tagName' FROM {$table_prefix}tagsquot;,
'category'=>quot;SELECT cat_ID id,cat_name 'skos:prefLabel',category_parent 'skos:narrower' FROM {$table_prefix}categoriesquot;,
'user'=>array(
quot;SELECT id,user_login 'foaf:accountName',SHA(CONCAT('mailto:',user_email)) 'foaf:mbox_sha1sum',
user_url 'foaf:homepage',display_name 'foaf:name' FROM {$table_prefix}usersquot;,
quot;SELECT user_id id,meta_value 'foaf:firstName' FROM {$table_prefix}usermeta WHERE meta_key='first_name'quot;,
quot;SELECT user_id id,meta_value 'foaf:family_name' FROM {$table_prefix}usermeta WHERE meta_key='last_name'quot;,
),
'comment'=>quot;SELECT comment_ID id,comment_post_id 'sioc:reply_of',comment_author AS 'foaf:name',
SHA(CONCAT('mailto:',comment_author_email)) 'foaf:mbox_sha1sum', comment_author_url 'foaf:homepage',
comment_date AS 'dcterms:created', comment_content 'sioc:content',comment_karma,comment_type
FROM {$table_prefix}comments WHERE comment_approved='1'quot;,
);
$triplify['objectProperties']=array(
'sioc:has_creator'=>'user', 'tag:taggedWithTag'=>'tag', 'belongsToCategory'=>'category‘,'skos:narrower'=>'category','sioc:reply_of'=>'post');
$triplify['classMap']=array('user'=>'foaf:person', 'post'=>'sioc:Post', 'tag'=>'tag:Tag', 'category'=>'skos:Concept');
$triplify['TTL']=0; // Caching
$triplify['db']=new PDO('mysql:host='.DB_HOST.';dbname='.DB_NAME,DB_USER,DB_PASSWORD);
?>
4/24/2009 Triplify – Linked Data Publication from Relational DBs 11
12. Configuration repository
• Triplify configurations are shared at:
http://Triplify.org
• Existing configurations for
OpenConf, Wordpress, WackoWiki, Drupal,
OJS, Joomla, osCommerce, Gallery, phpBB,
OMDB …
4/24/2009 Triplify – Linked Data Publication from Relational DBs 12
13. Triplify Endpoint Registry
• Simple REST endpoint registry:
http://triplify.org/Registry/?url=%rdf_source_URL%
• Itself available as Linked Data endpoint
• Enables building of mashups, vertical search
and other applications using information from
many sources – product search, blog search
etc.
4/24/2009 Triplify – Linked Data Publication from Relational DBs 13
14. Triplify Temporal Extension
Problem: How do next generation search engines know
something changed on the Data Web?
Different solutions:
• Try to crawl always everything: currently deployed on
the Web
• Ping a central update notification service:
PingTheSemanticWeb.com – will probably not scale if
the Data Web gets really deployed
• Each linked data endpoint publishes an update log:
Triplify Update Logs
4/24/2009 Triplify – Linked Data Publication from Relational DBs 14
15. Triplify Temporal Extension
special update path and vocabulary
http://example.com/Triplify/update
http://example.com/Triplify/update/2007 rdf:type update:UpdateCollection .
http://example.com/Triplify/update/2008 rdf:type update:UpdateCollection .
http://example.com/Triplify/update/2008
http://example.com/Triplify/update/2008/Jan rdf:type update:UpdateCollection .
http://example.com/Triplify/update/2008/Feb rdf:type update:UpdateCollection .
Nesting continues until we finally reach an URL, which exposes all updates performed in a certain
second in time…
http://example.com/Triplify/update/2008/Jan/01/17/58/06
http://example.com/Triplify/update/2008/Jan/01/17/58/06/user123
update:updatedResource http://example.com/Triplify/users/JohnDoe ;
update:updatedAt quot;20080101T17:58:06quot;^<xsd:dateTime> ;
update:updatedBy http://example.com/Triplify/users/JohnDoe .
4/24/2009 Triplify – Linked Data Publication from Relational DBs 15
16. LOD Update log generation
Updates have to be logged in the DB
Update log queries have to expose a date as first
column:
$triplify['queries']=array(
…
'update'=>quot;SELECT p.changed AS id,
p.id AS 'update:updatedResource->project‘
FROM project pquot;,
);
4/24/2009 Triplify – Linked Data Publication from Relational DBs 16
17. Triplify Spatial Extension:
Linked Open Geo Data
Spatial data is crucial for the Data Web in order to interlink geographically linked resources.
Open Street Map project (OSM) collects, organizes and publishes geo data the wiki way:
• 80.000 OSM users collected data about 22M km ways (roads, highways etc.) on earth, 25T
km are added daily
• OSM contains a vast amount points-of-interest descriptions e.g. shops, amenities, sports
venues, businesses, touristic and historic sights.
Goal: publish OSM geo data, interlink it with other data sources and provide efficient means
for browsing and authoring:
• Open Street Map data extraction works on the basis of OSM database dumps, a bi-
directional live integration of OSM and our Linked Geo Data browser and editor is currently
in the works.
• Triplify spatial data publishing, the Triplify script for publishing linked data from relational
databases is extended for publishing geo data, in particular with regard to the retrieval
of information about geographical areas.
• LinkedGeo Data browser and editor is a facet-based browser for geo content, which uses
an OLAP inspired hypercube for quickly retrieving aggregated information about any user
selected area on earth.
Linked Data Tutorial
18. Triplify Spatial Extension
How to publish geo-data using Triplify?
Radius Attribute Value
Lon Lat
http://linkedgeodata.org/near/48.213056,16.359722/1000/amenity=Hotel
http://linkedgeodata.org/node/212331
http://linkedgeodata.org/node/944523
http://linkedgeodata.org/node/234091
http://linkedgeodata.org/way/56719
node/150760824 amenity quot;pubquot;;
created_by quot;JOSMquot;;
distance quot;5995quot;;
name quot;La frioleraquot;;
geo#lat quot;40.4474quot;;
geo#long quot;-3.7173quot;.
4/24/2009 Triplify – Linked Data Publication from Relational DBs 18
20. RDB2RDF tool comparison
Tool Virtuoso RDF
Triplify D2RQ
Views
Technology
Scripting languages Whole middleware
Java
(PHP) solution
SPARQL endpoint
- X X
Mapping language
SQL RDF based RDF based
Mapping
Manual Semi-automatic Manual
generation
Scalability Medium-high
Medium High
(but no SPARQL)
More at: http://esw.w3.org/topic/Rdb2RdfXG/StateOfTheArt
4/24/2009 Triplify – Linked Data Publication from Relational DBs 20
21. Conclusion
• Triplify supports the “long tail” of deployed Web
applications
• Publishing RDF and Linked Data is simple
• Support for temporal and spatial data
dimensions
– LOD Update Logs enable differential crawling
– Linkedgeodata.org provides spatial identifiers for most
parts of the world
• More comprehensive solutions are (still) required
for SPARQL support
4/24/2009 Triplify – Linked Data Publication from Relational DBs 21
22. Should it be a Cloud or the Sky?
4/24/2009 Triplify – Linked Data Publication from Relational DBs 22
23. Thanks!
Sören Auer
auer@informatik.uni-leipzig.de
Research group Agile Knowledge Engineering & Semantic Web
(AKSW): http://aksw.org
• http://Triplify.org
• http://DBpedia.org
• http://OntoWiki.net
• http://OpenResearch.org
• http://aksw.org/projects/xOperator
• DL-Learner.org
• Cofundos.org
Linked Data Tutorial