Knowledge graphs have been conceived to collect heterogeneous data and knowledge about large domains, e.g. medical or engineering domains, and to allow versatile access to such collections by means of querying and logical reasoning. A surge of methods has responded to additional requirements in recent years. (i) Knowledge graph embeddings use similarity and analogy of structures to speculatively add to the collected data and knowledge. (ii) Queries with shapes and schema information can be typed to provide certainty about results. We survey both developments and find that the development of techniques happens in disjoint communities that mostly do not understand each other, thus limiting the proper and most versatile use of knowledge graphs.
I claim that none of the commonly used embedding methods capture any semantics.
It's fine if you want to move from a symbolic to a numeric or geometric representation, but when you do, don't throw the semantic baby out with the symbolic bathwater.
I argue that a useful definition of semantics is "predictable inference". This makes it possible to have semantics outside a logical framework.
A methodological warning from 1976: don't fool yourself that wishful mnemonics in your knowledge graph are "semantics". Therefore, knowledge graphs without a schema/ontology is just a data graph, without much semantics.
Finally, a discussion of some embedding methods that do manage to take semantics into account (TransOWL, ball embeddings like ELEm and EmEL++, and box embeddings like BoxEL and Box^2EL.
So: even if you do move to a non-symbolic representation (numerical, geometric), make sure you keep the semantics: don't throw the semantic baby out with the symbolic bathwater.
Vector space model or term vector model is an algebraic model for representing text documents as vectors of identifiers, such as, for example, index terms. It is used in information filtering, information retrieval, indexing and relevancy rankings. Its first use was in the SMART Information Retrieval System
See an example on writing a computer science dissertation literature review and get more information at https://www.literaturereviewwritingservice.com/
Making Decisions - From Software Architecture Theory to PracticeParis Avgeriou
Keynote at the 20th IEEE International Conference on Software Architecture (ICSA 2023).
Abstract: Around the mid 2000’s we had a ‘lightbulb moment’: architecting is more about making design decisions than drawing boxes and lines with UML or Architecture Description Languages. The enthusiasm of shifting the architecting paradigm was followed by a frenzy of research work in ontologies, methods and tools to harness Architecture Knowledge and empower architects in their decision making. Are we there yet? In this talk, I will revisit the state of the art and give my personal account on what worked, what failed, and how to move forward, especially in order to actually impact industry practice.
This chapter gives information about Social media analytics, Social network analysis, Text analytics, stopwords, tokenization, n-grams, Trend analytics, TF-IDF, Stemming and lemmatization
Keynote presentation from ECBS conference. The talk is about how to use machine learning and AI in improving software engineering. Experiences from our project in Software Center (www.software-center.se).
I claim that none of the commonly used embedding methods capture any semantics.
It's fine if you want to move from a symbolic to a numeric or geometric representation, but when you do, don't throw the semantic baby out with the symbolic bathwater.
I argue that a useful definition of semantics is "predictable inference". This makes it possible to have semantics outside a logical framework.
A methodological warning from 1976: don't fool yourself that wishful mnemonics in your knowledge graph are "semantics". Therefore, knowledge graphs without a schema/ontology is just a data graph, without much semantics.
Finally, a discussion of some embedding methods that do manage to take semantics into account (TransOWL, ball embeddings like ELEm and EmEL++, and box embeddings like BoxEL and Box^2EL.
So: even if you do move to a non-symbolic representation (numerical, geometric), make sure you keep the semantics: don't throw the semantic baby out with the symbolic bathwater.
Vector space model or term vector model is an algebraic model for representing text documents as vectors of identifiers, such as, for example, index terms. It is used in information filtering, information retrieval, indexing and relevancy rankings. Its first use was in the SMART Information Retrieval System
See an example on writing a computer science dissertation literature review and get more information at https://www.literaturereviewwritingservice.com/
Making Decisions - From Software Architecture Theory to PracticeParis Avgeriou
Keynote at the 20th IEEE International Conference on Software Architecture (ICSA 2023).
Abstract: Around the mid 2000’s we had a ‘lightbulb moment’: architecting is more about making design decisions than drawing boxes and lines with UML or Architecture Description Languages. The enthusiasm of shifting the architecting paradigm was followed by a frenzy of research work in ontologies, methods and tools to harness Architecture Knowledge and empower architects in their decision making. Are we there yet? In this talk, I will revisit the state of the art and give my personal account on what worked, what failed, and how to move forward, especially in order to actually impact industry practice.
This chapter gives information about Social media analytics, Social network analysis, Text analytics, stopwords, tokenization, n-grams, Trend analytics, TF-IDF, Stemming and lemmatization
Keynote presentation from ECBS conference. The talk is about how to use machine learning and AI in improving software engineering. Experiences from our project in Software Center (www.software-center.se).
Artificial intelligence and machine learning capabilities are growing at an unprecedented rate. These technologies have many widely beneficial applications, ranging from machine translation to medical image analysis. Countless more such applications are being developed and can be expected over the long term. Less attention has historically been paid to the ways in which artificial intelligence can be used maliciously. This report surveys the landscape of potential security threats from malicious uses of artificial intelligence technologies, and proposes ways to better forecast, prevent, and mitigate these threats. We analyze, but do not conclusively resolve, the question of what the long-term equilibrium between attackers and defenders will be. We focus instead on what sorts of attacks we are likely to see soon if adequate defenses are not developed.
Research about artificial intelligence (A.I)Alị Ŕỉźvị
These slides contains no extra ordinary textual information but they are very good as being creative. relevant to the topic animations were added which guarantee not to make the audience bore.
History of AI, Current Trends, Prospective TrajectoriesGiovanni Sileno
Talk given at the 2nd Winter Academy on Artificial Intelligence and International Law of the Asser Institute. The birth of AI: Dartmouth workshop. The biggest AI waves: classic symbolic AI (reasoning, knowledge systems, problem-solving), machine learning (induction). Current problems: explainability, trustworthyness, impact and transformation on society and people, the rise of artificially dumber systems.
The training content covers:
- Basics of Artificial Intelligence
- Penetration of AI in our daily lives
- Few examples and Use cases
- A brief on how future with AI looks like
Slides from HR Talks on Future of work: AI vs. Human.
Organized by HR Hub in Bucharest, on 23 Jan 2017.
Topics discussed:
* Automation
* AI
* Impact on HR
Past, Present and Future of Generative AIabhishek36461
Generative AI creates new content (images, text, music) based on learned patterns.
It learns from vast examples and can produce original, unseen works.
Capable of blending learned elements to generate unique outputs.
Can produce customized creations based on specific prompts.
Improves and refines its output over time with more data and feedback.
Technology for everyone - AI ethics and BiasMarion Mulder
Slides from my talk at #ToonTechTalks on 27 september 2018
We all see the great potential AI is bringing us. But is it really bringing it to everyone? How are we ensuring under-represented groups are included and vulnerable people are protected? What to do when our technology is unintended biased and discriminating against certain groups. And what if the data and AI is correct, but the by-effect of it is that some groups are put at risk? All questions we need to think about when we are advancing technology for the benefit of humanity.
Sharing what I've learned from my work in diversity, digital and from following great minds in this field such as Joanna Bryson, Virginia Dignum, Rumman Chowdhury, Juriaan van Diggelen, Valerie Frissen, Catelijne Muller, and many more.
Introduction to artifcial intelligence
Artificial intelligence (AI) is intelligence demonstrated by machines, unlike the natural intelligence displayed by humans and animals, which involves consciousness and emotionality. The distinction between the former and the latter categories is often revealed by the acronym chosen. 'Strong' AI is usually labelled as AGI (Artificial General Intelligence) while attempts to emulate 'natural' intelligence have been called ABI (Artificial Biological Intelligence). Leading AI textbooks define the field as the study of "intelligent agents": any device that perceives its environment and takes actions that maximize its chance of successfully achieving its goals.[3] Colloquially, the term "artificial intelligence" is often used to describe machines (or computers) that mimic "cognitive" functions that humans associate with the human mind, such as "learning" and "problem solving"
Vulnerability in AI
1- Introduction to AI
2- Vulnerability
3- The impact of AI on vulnerability management
4- Use of AI in cybersecurity
5- Vulnerability Management
6- Conclusion
AI Workshops at Computers In Libraries 2024Brian Pichman
While AI holds tremendous potential for libraries, it also comes with significant concerns and the potential for harm. We find ourselves sailing uncertain waters; there are few guardrails governing AI's use. Even as we acknowledge this truth, we must also note that library staff are already experimenting with the use of AI chatbots (most commonly ChatGPT), generative AI design tools (like Midjourney), and other variations of AI technology. In short, we have great potential, pitfalls, and a total lack of clarity. It is only through the thoughtful development of policy, procedure, and professionals that we can hope to articulate a vision for the ethical use of AI in our libraries. Join this conversation about new disruptive technology, take a deep breath, and get to work laying a foundation of policy guidelines and staff development to navigate the uncertain road ahead.
This interactive and hands-on workshop allows you to play and experiment with new tools which will spark ideas for the future of your library and community activities. It focuses on OpenAI’s API and how to get started building personalities in AI. It explores various tools to create AI images, videos, and more. Filled with tips, it will definitely be fun!
Workshop on "Building Successful Pipelines for Predictive Analytics in Healthcare" delivered by Danielle Belgrave, PhD, Researcher at Microsoft Research, Cambridge, UK.
Applications of Artificial Intelligence-Past, Present & FutureJamie Gannon
This presentation in Ignite format gives a brief look into the applications of Artificial Intelligence. Starting from the humble beginnings and working its way through present day and finally the future possibilities of Artificial intelligence.
Shapes for Sharing between Graph Data Spaces - and Epistemic Querying of RDF-...Steffen Staab
Data spaces in distributed environments should be allowed to evolve in agile ways providing data space owners with large flexibility about which data they store. Agility and heterogeneity, however, jeopardize data exchanges because representations may build on varying ontologies and data consumers may not rely on the semantic correctness of their queries in the context of semantically heterogeneous, evolving data spaces. Graph data spaces are one example of a powerful model for representing and querying data whose semantics may change over time. To assert and enforce conditions on individual graph data spaces, shape languages (e.g SHACL) have been developed. We investigate the question of how querying and programming can be guarded by reasoning over SHACL constraints in a distributed setting and we sketch a picture of how a future landscape based on semantically heterogeneous data spaces might look like.
Machine Learning Interpretability - Mateusz Dymczyk - H2O AI World London 2018Sri Ambati
This talk was recorded in London on Oct 30, 2018 and can be viewed here: https://youtu.be/p4iAnxwC_Eg
The good news is building fair, accountable, and transparent machine learning systems is possible. The bad news is it’s harder than many blogs and software package docs would have you believe. The truth is nearly all interpretable machine learning techniques generate approximate explanations, that the fields of eXplainable AI (XAI) and Fairness, Accountability, and Transparency in Machine Learning (FAT/ML) are very new, and that few best practices have been widely agreed upon. This combination can lead to some ugly outcomes!
This talk aims to make your interpretable machine learning project a success by describing fundamental technical challenges you will face in building an interpretable machine learning system, defining the real-world value proposition of approximate explanations for exact models, and then outlining the following viable techniques for debugging, explaining, and testing machine learning models
Mateusz is a software developer who loves all things distributed, machine learning and hates buzzwords. His favourite hobby data juggling.
He obtained his M.Sc. in Computer Science from AGH UST in Krakow, Poland, during which he did an exchange at L’ECE Paris in France and worked on distributed flight booking systems. After graduation he move to Tokyo to work as a researcher at Fujitsu Laboratories on machine learning and NLP projects, where he is still currently based.
Artificial intelligence and machine learning capabilities are growing at an unprecedented rate. These technologies have many widely beneficial applications, ranging from machine translation to medical image analysis. Countless more such applications are being developed and can be expected over the long term. Less attention has historically been paid to the ways in which artificial intelligence can be used maliciously. This report surveys the landscape of potential security threats from malicious uses of artificial intelligence technologies, and proposes ways to better forecast, prevent, and mitigate these threats. We analyze, but do not conclusively resolve, the question of what the long-term equilibrium between attackers and defenders will be. We focus instead on what sorts of attacks we are likely to see soon if adequate defenses are not developed.
Research about artificial intelligence (A.I)Alị Ŕỉźvị
These slides contains no extra ordinary textual information but they are very good as being creative. relevant to the topic animations were added which guarantee not to make the audience bore.
History of AI, Current Trends, Prospective TrajectoriesGiovanni Sileno
Talk given at the 2nd Winter Academy on Artificial Intelligence and International Law of the Asser Institute. The birth of AI: Dartmouth workshop. The biggest AI waves: classic symbolic AI (reasoning, knowledge systems, problem-solving), machine learning (induction). Current problems: explainability, trustworthyness, impact and transformation on society and people, the rise of artificially dumber systems.
The training content covers:
- Basics of Artificial Intelligence
- Penetration of AI in our daily lives
- Few examples and Use cases
- A brief on how future with AI looks like
Slides from HR Talks on Future of work: AI vs. Human.
Organized by HR Hub in Bucharest, on 23 Jan 2017.
Topics discussed:
* Automation
* AI
* Impact on HR
Past, Present and Future of Generative AIabhishek36461
Generative AI creates new content (images, text, music) based on learned patterns.
It learns from vast examples and can produce original, unseen works.
Capable of blending learned elements to generate unique outputs.
Can produce customized creations based on specific prompts.
Improves and refines its output over time with more data and feedback.
Technology for everyone - AI ethics and BiasMarion Mulder
Slides from my talk at #ToonTechTalks on 27 september 2018
We all see the great potential AI is bringing us. But is it really bringing it to everyone? How are we ensuring under-represented groups are included and vulnerable people are protected? What to do when our technology is unintended biased and discriminating against certain groups. And what if the data and AI is correct, but the by-effect of it is that some groups are put at risk? All questions we need to think about when we are advancing technology for the benefit of humanity.
Sharing what I've learned from my work in diversity, digital and from following great minds in this field such as Joanna Bryson, Virginia Dignum, Rumman Chowdhury, Juriaan van Diggelen, Valerie Frissen, Catelijne Muller, and many more.
Introduction to artifcial intelligence
Artificial intelligence (AI) is intelligence demonstrated by machines, unlike the natural intelligence displayed by humans and animals, which involves consciousness and emotionality. The distinction between the former and the latter categories is often revealed by the acronym chosen. 'Strong' AI is usually labelled as AGI (Artificial General Intelligence) while attempts to emulate 'natural' intelligence have been called ABI (Artificial Biological Intelligence). Leading AI textbooks define the field as the study of "intelligent agents": any device that perceives its environment and takes actions that maximize its chance of successfully achieving its goals.[3] Colloquially, the term "artificial intelligence" is often used to describe machines (or computers) that mimic "cognitive" functions that humans associate with the human mind, such as "learning" and "problem solving"
Vulnerability in AI
1- Introduction to AI
2- Vulnerability
3- The impact of AI on vulnerability management
4- Use of AI in cybersecurity
5- Vulnerability Management
6- Conclusion
AI Workshops at Computers In Libraries 2024Brian Pichman
While AI holds tremendous potential for libraries, it also comes with significant concerns and the potential for harm. We find ourselves sailing uncertain waters; there are few guardrails governing AI's use. Even as we acknowledge this truth, we must also note that library staff are already experimenting with the use of AI chatbots (most commonly ChatGPT), generative AI design tools (like Midjourney), and other variations of AI technology. In short, we have great potential, pitfalls, and a total lack of clarity. It is only through the thoughtful development of policy, procedure, and professionals that we can hope to articulate a vision for the ethical use of AI in our libraries. Join this conversation about new disruptive technology, take a deep breath, and get to work laying a foundation of policy guidelines and staff development to navigate the uncertain road ahead.
This interactive and hands-on workshop allows you to play and experiment with new tools which will spark ideas for the future of your library and community activities. It focuses on OpenAI’s API and how to get started building personalities in AI. It explores various tools to create AI images, videos, and more. Filled with tips, it will definitely be fun!
Workshop on "Building Successful Pipelines for Predictive Analytics in Healthcare" delivered by Danielle Belgrave, PhD, Researcher at Microsoft Research, Cambridge, UK.
Applications of Artificial Intelligence-Past, Present & FutureJamie Gannon
This presentation in Ignite format gives a brief look into the applications of Artificial Intelligence. Starting from the humble beginnings and working its way through present day and finally the future possibilities of Artificial intelligence.
Shapes for Sharing between Graph Data Spaces - and Epistemic Querying of RDF-...Steffen Staab
Data spaces in distributed environments should be allowed to evolve in agile ways providing data space owners with large flexibility about which data they store. Agility and heterogeneity, however, jeopardize data exchanges because representations may build on varying ontologies and data consumers may not rely on the semantic correctness of their queries in the context of semantically heterogeneous, evolving data spaces. Graph data spaces are one example of a powerful model for representing and querying data whose semantics may change over time. To assert and enforce conditions on individual graph data spaces, shape languages (e.g SHACL) have been developed. We investigate the question of how querying and programming can be guarded by reasoning over SHACL constraints in a distributed setting and we sketch a picture of how a future landscape based on semantically heterogeneous data spaces might look like.
Machine Learning Interpretability - Mateusz Dymczyk - H2O AI World London 2018Sri Ambati
This talk was recorded in London on Oct 30, 2018 and can be viewed here: https://youtu.be/p4iAnxwC_Eg
The good news is building fair, accountable, and transparent machine learning systems is possible. The bad news is it’s harder than many blogs and software package docs would have you believe. The truth is nearly all interpretable machine learning techniques generate approximate explanations, that the fields of eXplainable AI (XAI) and Fairness, Accountability, and Transparency in Machine Learning (FAT/ML) are very new, and that few best practices have been widely agreed upon. This combination can lead to some ugly outcomes!
This talk aims to make your interpretable machine learning project a success by describing fundamental technical challenges you will face in building an interpretable machine learning system, defining the real-world value proposition of approximate explanations for exact models, and then outlining the following viable techniques for debugging, explaining, and testing machine learning models
Mateusz is a software developer who loves all things distributed, machine learning and hates buzzwords. His favourite hobby data juggling.
He obtained his M.Sc. in Computer Science from AGH UST in Krakow, Poland, during which he did an exchange at L’ECE Paris in France and worked on distributed flight booking systems. After graduation he move to Tokyo to work as a researcher at Fujitsu Laboratories on machine learning and NLP projects, where he is still currently based.
Data-centric AI and the convergence of data and model engineering:opportunit...Paolo Missier
A keynote talk given to the IDEAL 2023 conference (Evora, Portugal Nov 23, 2023).
Abstract.
The past few years have seen the emergence of what the AI community calls "Data-centric AI", namely the recognition that some of the limiting factors in AI performance are in fact in the data used for training the models, as much as in the expressiveness and complexity of the models themselves. One analogy is that of a powerful engine that will only run as fast as the quality of the fuel allows. A plethora of recent literature has started the connection between data and models in depth, along with startups that offer "data engineering for AI" services. Some concepts are well-known to the data engineering community, including incremental data cleaning, multi-source integration, or data bias control; others are more specific to AI applications, for instance the realisation that some samples in the training space are "easier to learn from" than others. In this "position talk" I will suggest that, from an infrastructure perspective, there is an opportunity to efficiently support patterns of complex pipelines where data and model improvements are entangled in a series of iterations. I will focus in particular on end-to-end tracking of data and model versions, as a way to support MLDev and MLOps engineers as they navigate through a complex decision space.
INF 103(ASH) Possible Is Everything/newtonhelp.comlechenau71
For more course tutorials visit
www.newtonhelp.com
INF 103 Week 1 DQ 1 How Do You Currently Use Information Technology
INF 103 Week 1 DQ 2 Innovations in Hardware and Software
INF 103 Week 2 DQ 1 Copyright and Fair Use
INF 103 Week 2 DQ 2 Searching for Information
INF 103 Week 2 Assignment Using Microsoft Word What Does the Library Have to Offer
For more course tutorials visit
www.newtonhelp.com
INF 103 Week 1 DQ 1 How Do You Currently Use Information Technology
INF 103 Week 1 DQ 2 Innovations in Hardware and Software
INF 103 Week 2 DQ 1 Copyright and Fair Use
INF 103 Week 2 DQ 2 Searching for Information
INF 103 Week 2 Assignment Using Microsoft Word What Does the Library Have to Offer
Presentation of the Semantic Knowledge Graph research paper at the 2016 IEEE 3rd International Conference on Data Science and Advanced Analytics (Montreal, Canada - October 18th, 2016)
Abstract—This paper describes a new kind of knowledge representation and mining system which we are calling the Semantic Knowledge Graph. At its heart, the Semantic Knowledge Graph leverages an inverted index, along with a complementary uninverted index, to represent nodes (terms) and edges (the documents within intersecting postings lists for multiple terms/nodes). This provides a layer of indirection between each pair of nodes and their corresponding edge, enabling edges to materialize dynamically from underlying corpus statistics. As a result, any combination of nodes can have edges to any other nodes materialize and be scored to reveal latent relationships between the nodes. This provides numerous benefits: the knowledge graph can be built automatically from a real-world corpus of data, new nodes - along with their combined edges - can be instantly materialized from any arbitrary combination of preexisting nodes (using set operations), and a full model of the semantic relationships between all entities within a domain can be represented and dynamically traversed using a highly compact representation of the graph. Such a system has widespread applications in areas as diverse as knowledge modeling and reasoning, natural language processing, anomaly detection, data cleansing, semantic search, analytics, data classification, root cause analysis, and recommendations systems. The main contribution of this paper is the introduction of a novel system - the Semantic Knowledge Graph - which is able to dynamically discover and score interesting relationships between any arbitrary combination of entities (words, phrases, or extracted concepts) through dynamically materializing nodes and edges from a compact graphical representation built automatically from a corpus of data representative of a knowledge domain.
Predictive Model and Record Description with Segmented Sensitivity Analysis (...Greg Makowski
Describing a predictive data mining model can provide a competitive advantage for solving business problems with a model. The SSA approach can also provide reasons for the forecast for each record. This can help drive investigations into fields and interactions during a data mining project, as well as identifying "data drift" between the original training data, and the current scoring data. I am working on open source version of SSA, first in R.
In the last decade, several Scientific Knowledge Graphs (SKG) were released, representing scientific knowledge in a structured, interlinked, and semantically rich manner. But, what kind of information they describe? How they have been built? What can we do with them? In this lecture, I will first provide an overview of well-known SKGs, like Microsoft Academic Graph, Dimensions, and others. Then, I will present the Academia/Industry DynAmics (AIDA) Knowledge Graph, which describes 21M publications and 8M patents according to i) the research topics drawn from the Computer Science Ontology, ii) the type of the author's affiliations (e.g, academia, industry), and iii) 66 industrial sectors (e.g., automotive, financial, energy, electronics) from the Industrial Sectors Ontology (INDUSO). Finally, I will showcase a number of tools and approaches using such SKGs, supporting researchers, companies, and policymakers in making sense of research dynamics.
Natural language processing (NLP) is an area of artificial intelligence that helps computers understand and interpret human language. Innovations in Artificial intelligence, deep learning and compuational hardware is helping make major strides in NLP research. While the applications are many, it is important to understand the kinds of problems NLP techniques can help solve.
In this master class, we will introduce ten key NLP techniques that are predominantly used in the industry.
- Question Answering
- Neural Machine Translation
- Topic Summarization
- Natural Language Inference
- Semantic Role Labeling
- Text Classification
- Sentiment Analysis
- Relation extraction
- Goal-Oriented Dialogue
- Semantic Parsing
We will also illustrate a case study on NLP in Python using the QuSandbox.
Driverless AI Hands-on Focused on Machine Learning Interpretability - H2O.aiSri Ambati
Presented at #H2OWorld 2017 in Mountain View, CA.
Enjoy the video: https://youtu.be/axIqeaUhow0.
Learn more about H2O.ai: https://www.h2o.ai/.
Follow @h2oai: https://twitter.com/h2oai.
- - -
Abstract:
Usage of AI and machine learning models is likely to become more commonplace as larger swaths of the economy embrace automation and data-driven decision-making. While these predictive systems can be quite accurate, they have been treated as inscrutable black boxes in the past, that produce only numeric predictions with no accompanying explanations. Unfortunately, recent studies and recent events have drawn attention to mathematical and sociological flaws in prominent weak AI and ML systems, but practitioners usually don’t have the right tools to pry open machine learning black-boxes and debug them. This presentation introduces several new approaches to that increase transparency, accountability, and trustworthiness in machine learning models. If you are a data scientist or analyst and you want to explain a machine learning model to your customers or managers (or if you have concerns about documentation, validation, or regulatory requirements), then this presentation is for you!
Multi-Model Data Query Languages and Processing ParadigmsJiaheng Lu
Specifying users' interests with a formal query language is a typically challenging task, which becomes even harder in the context of multi-model data management because we have to deal with data variety. It usually lacks a unified schema to help the users issuing their queries, or has an incomplete schema as data come from disparate sources. Multi-Model DataBases (MMDBs) have emerged as a promising approach for dealing with this task as they are capable of accommodating and querying the multi-model data in a single system. This tutorial aims to offer a comprehensive presentation of a wide range of query languages for MMDBs and to make comparisons of their properties from multiple perspectives. We will discuss the essence of cross-model query processing and provide insights on the research challenges and directions for future work. The tutorial will also offer the participants hands-on experience in applying MMDBs to issue multi-model data queries.
Symbolic Background Knowledge for Machine LearningSteffen Staab
Machine learning aims at learning complex functions from data. Very often, this challenge remains ill-defined given the available amount of data, however, background knowledge that is available as knowledge graphs, ontologies or symbolic (physical) equations allows for an improved specification of the targeted solution. In this talk, we want to discuss several use cases that include symbolic background knowledge as regularizing priors, as constraints or as other inductive biases into machine learning tasks.
Soziale Netzwerke und Medien: Multi-disziplinäre Ansätze für ein multi-dimens...Steffen Staab
Präsentation von Oul Han und Steffen Staab
Workshop "Soziale Netzwerke und Medien" auf dem Treffen des Fakultätentags Informatik, 14. November 2019, Hamburg
Web Futures: Inclusive, Intelligent, SustainableSteffen Staab
Almost from its very beginning, the Web has been ambivalent.
It has facilitated freedom for information, but this also included the freedom to spread misinformation. It has faciliated intelligent personalization, but at the cost of intrusion into our private lifes. It has included more people than any other system before, but at the risk of exploiting them.
The Web is full of such ambivalences and the usage of artificial intelligences threatens to further amplify these ambivalences. To further the good and to contain the negative consequences, we need a research agenda studying and engineering the Web, as well as numerous activities by societies at large. In this talk, I will present and discuss a joint effort by an interdisciplinary team of Web Scientists to prepare and pursue such an agenda.
Concepts in Application Context ( How we may think conceptually )Steffen Staab
Formal concept analysis (FCA) derives a hierarchy of concepts
in a formal context that relates objects with attributes. This approach is very well aligned with the traditions of Frege, Saussure and Peirce, which relate a signifier (e.g. a word/an attribute) to a mental concept evoked by this word and meant to refer to a specific object in the real world. However, in the practice of natural languages as well as artificial languages (e.g. programming languages), the application context
often constitutes a latent variable that influences the interpretation of a signifier. We present some of our current work that analyzes the usage of words in natural language in varying application contexts as well as the usage of variables in programming languages in varying application contexts in order to provide conceptual constraints on these signifiers.
Storing and Querying Semantic Data in the CloudSteffen Staab
Daniel Janke and Steffen Staab. Tutorial at Reasoning Web
With proliferation of semantic data, there is a need to cope with trillions of triples by horizontally scaling data management in the cloud. To this end one needs to advance (i) strategies for data placement over compute and storage nodes, (ii) strategies for distributed query processing, and (iii) strategies for handling failure of compute and storage nodes. In this tutorial, we want to review challenges and how they have been addressed by research and development in the last 15 years.
Talk at Leopoldina Symposium on Digitization and its Effects on Man and Society
(Die Digitalisierung und ihre Auswirkungen auf Mensch und Gesellschaft)
leopoldina.org/de/veranstaltungen/veranstaltung/event/2464/
The evolution of the Web should move forward in an upward spiral that cylces between guiding values, engineering and science. Guiding values should comprise social values as well as system principles that further stabilization and growth of the Web. Principles I will talk about will include social inclusion, connectedness and fairness. Example efforts improve Web access for disabled, critically access Web structures and Web growth, and try to transfer knowledge about previously found patterns of Web growth to analogous cases.
(Semi-)Automatic analysis of online contentsSteffen Staab
How can media and discourse analyses combine approaches from humanities and statistical methods to deeply analyse large amounts of online contents.
Invited talk at Fachgruppen-Workshop der Deutschen Gesellschaft für Publizistik und Kommunikationswissenschaft
Soziale Medien – Echo-Kammer oder öffentlicher Raum?
Ansätze zur computergestützten Analyse von Internet-Korpora
6. Oktober 2016, Karlsruher Institut für Technologie (KIT)
Joint Keynote at Int. Conference on Knowledge Engineering and Semantic Web and Prague Computer Science Seminar, Prague, September 22, 2016
The challenges of Big Data are frequently explained by dealing with Volume, Velocity, Variety and Veracity. The large variety of data in organizations results from accessing different information systems with heterogeneous schemata or ontologies. In this talk I will present the research efforts that target the management of such broad data.
They include: (i) an integrated development environment for programming with broad data, (ii) a query language that allows for typing of query results, (iii) a typed lambda-calculus based on description logics, and (iv) efficient access to data repositories via schema indices.
We use metadata of various kind to improve and enrich text document clustering using an extension of Latent Dirichlet Allocation (LDA). The methods are fully implemented, evaluated and software is available on github.
These are the slides of an invited talk I gave September 8 at the Alexandria Workshop of TPDL-2016: http://alexandria-project.eu/events/3rd-workshop/
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
UiPath Test Automation using UiPath Test Suite series, part 4DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 4. In this session, we will cover Test Manager overview along with SAP heatmap.
The UiPath Test Manager overview with SAP heatmap webinar offers a concise yet comprehensive exploration of the role of a Test Manager within SAP environments, coupled with the utilization of heatmaps for effective testing strategies.
Participants will gain insights into the responsibilities, challenges, and best practices associated with test management in SAP projects. Additionally, the webinar delves into the significance of heatmaps as a visual aid for identifying testing priorities, areas of risk, and resource allocation within SAP landscapes. Through this session, attendees can expect to enhance their understanding of test management principles while learning practical approaches to optimize testing processes in SAP environments using heatmap visualization techniques
What will you get from this session?
1. Insights into SAP testing best practices
2. Heatmap utilization for testing
3. Optimization of testing processes
4. Demo
Topics covered:
Execution from the test manager
Orchestrator execution result
Defect reporting
SAP heatmap example with demo
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
JMeter webinar - integration with InfluxDB and GrafanaRTTS
Watch this recorded webinar about real-time monitoring of application performance. See how to integrate Apache JMeter, the open-source leader in performance testing, with InfluxDB, the open-source time-series database, and Grafana, the open-source analytics and visualization application.
In this webinar, we will review the benefits of leveraging InfluxDB and Grafana when executing load tests and demonstrate how these tools are used to visualize performance metrics.
Length: 30 minutes
Session Overview
-------------------------------------------
During this webinar, we will cover the following topics while demonstrating the integrations of JMeter, InfluxDB and Grafana:
- What out-of-the-box solutions are available for real-time monitoring JMeter tests?
- What are the benefits of integrating InfluxDB and Grafana into the load testing stack?
- Which features are provided by Grafana?
- Demonstration of InfluxDB and Grafana using a practice web application
To view the webinar recording, go to:
https://www.rttsweb.com/jmeter-integration-webinar
Accelerate your Kubernetes clusters with Varnish CachingThijs Feryn
A presentation about the usage and availability of Varnish on Kubernetes. This talk explores the capabilities of Varnish caching and shows how to use the Varnish Helm chart to deploy it to Kubernetes.
This presentation was delivered at K8SUG Singapore. See https://feryn.eu/presentations/accelerate-your-kubernetes-clusters-with-varnish-caching-k8sug-singapore-28-2024 for more details.
UiPath Test Automation using UiPath Test Suite series, part 3DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 3. In this session, we will cover desktop automation along with UI automation.
Topics covered:
UI automation Introduction,
UI automation Sample
Desktop automation flow
Pradeep Chinnala, Senior Consultant Automation Developer @WonderBotz and UiPath MVP
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Slack (or Teams) Automation for Bonterra Impact Management (fka Social Soluti...Jeffrey Haguewood
Sidekick Solutions uses Bonterra Impact Management (fka Social Solutions Apricot) and automation solutions to integrate data for business workflows.
We believe integration and automation are essential to user experience and the promise of efficient work through technology. Automation is the critical ingredient to realizing that full vision. We develop integration products and services for Bonterra Case Management software to support the deployment of automations for a variety of use cases.
This video focuses on the notifications, alerts, and approval requests using Slack for Bonterra Impact Management. The solutions covered in this webinar can also be deployed for Microsoft Teams.
Interested in deploying notification automations for Bonterra Impact Management? Contact us at sales@sidekicksolutionsllc.com to discuss next steps.
Essentials of Automations: Optimizing FME Workflows with ParametersSafe Software
Are you looking to streamline your workflows and boost your projects’ efficiency? Do you find yourself searching for ways to add flexibility and control over your FME workflows? If so, you’re in the right place.
Join us for an insightful dive into the world of FME parameters, a critical element in optimizing workflow efficiency. This webinar marks the beginning of our three-part “Essentials of Automation” series. This first webinar is designed to equip you with the knowledge and skills to utilize parameters effectively: enhancing the flexibility, maintainability, and user control of your FME projects.
Here’s what you’ll gain:
- Essentials of FME Parameters: Understand the pivotal role of parameters, including Reader/Writer, Transformer, User, and FME Flow categories. Discover how they are the key to unlocking automation and optimization within your workflows.
- Practical Applications in FME Form: Delve into key user parameter types including choice, connections, and file URLs. Allow users to control how a workflow runs, making your workflows more reusable. Learn to import values and deliver the best user experience for your workflows while enhancing accuracy.
- Optimization Strategies in FME Flow: Explore the creation and strategic deployment of parameters in FME Flow, including the use of deployment and geometry parameters, to maximize workflow efficiency.
- Pro Tips for Success: Gain insights on parameterizing connections and leveraging new features like Conditional Visibility for clarity and simplicity.
We’ll wrap up with a glimpse into future webinars, followed by a Q&A session to address your specific questions surrounding this topic.
Don’t miss this opportunity to elevate your FME expertise and drive your projects to new heights of efficiency.
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
Software Delivery At the Speed of AI: Inflectra Invests In AI-Powered QualityInflectra
In this insightful webinar, Inflectra explores how artificial intelligence (AI) is transforming software development and testing. Discover how AI-powered tools are revolutionizing every stage of the software development lifecycle (SDLC), from design and prototyping to testing, deployment, and monitoring.
Learn about:
• The Future of Testing: How AI is shifting testing towards verification, analysis, and higher-level skills, while reducing repetitive tasks.
• Test Automation: How AI-powered test case generation, optimization, and self-healing tests are making testing more efficient and effective.
• Visual Testing: Explore the emerging capabilities of AI in visual testing and how it's set to revolutionize UI verification.
• Inflectra's AI Solutions: See demonstrations of Inflectra's cutting-edge AI tools like the ChatGPT plugin and Azure Open AI platform, designed to streamline your testing process.
Whether you're a developer, tester, or QA professional, this webinar will give you valuable insights into how AI is shaping the future of software delivery.
Builder.ai Founder Sachin Dev Duggal's Strategic Approach to Create an Innova...Ramesh Iyer
In today's fast-changing business world, Companies that adapt and embrace new ideas often need help to keep up with the competition. However, fostering a culture of innovation takes much work. It takes vision, leadership and willingness to take risks in the right proportion. Sachin Dev Duggal, co-founder of Builder.ai, has perfected the art of this balance, creating a company culture where creativity and growth are nurtured at each stage.
Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...UiPathCommunity
💥 Speed, accuracy, and scaling – discover the superpowers of GenAI in action with UiPath Document Understanding and Communications Mining™:
See how to accelerate model training and optimize model performance with active learning
Learn about the latest enhancements to out-of-the-box document processing – with little to no training required
Get an exclusive demo of the new family of UiPath LLMs – GenAI models specialized for processing different types of documents and messages
This is a hands-on session specifically designed for automation developers and AI enthusiasts seeking to enhance their knowledge in leveraging the latest intelligent document processing capabilities offered by UiPath.
Speakers:
👨🏫 Andras Palfi, Senior Product Manager, UiPath
👩🏫 Lenka Dulovicova, Product Program Manager, UiPath
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf91mobiles
91mobiles recently conducted a Smart TV Buyer Insights Survey in which we asked over 3,000 respondents about the TV they own, aspects they look at on a new TV, and their TV buying preferences.
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf
Knowledge graphs for knowing more and knowing for sure
1. KI – Institute for Artificial Intelligence
Knowledge graphs for
knowing more and knowing for sure
Steffen Staab
@ststaab
https://www.ki.uni-stuttgart.de
https://semanux.com
https://southampton.ac.uk/research/institutes-centres/web-internet-science
2. provides an international forum for presentation and
discussion of research on information and knowledge
management, as well as recent advances on data and
knowledge bases.
2
Conference on Information and Knowledge Management
Rather: Conference on Large Language Models?
Let’s explore the role of knowledge bases/graphs!
3. 1. What is a Knowledge Graph?
2. Some Applications of Knowledge Graphs
3. Knowledge Graphs for Knowing for Sure
4. Knowledge Graphs for Knowing More
5. Large Language Models as Knowledge Bases
6. Large Language Models as AI Assistants
3
Plan for my talk
5. What is a Knowledge Graph?
A model for knowledge structures with
5
C22.0
Patient2342
treatedBy
„liver tumor“ / „PhValue 7.5“
Concepts
Entities
Relations
Labels / Values
6. Queries
• Scalability to
billions of facts
• Answering with
• facts
• predictions
• recommendations
6
What does a knowledge graph do for us?
What are the difficulties?
Example from medical project:
• Foundational Model of Anatomy:
75.000 concepts, 120.000 labels,
> 2 Mio facts
• Not even patient data yet!
7. Queries
Ontologies & Facts
• How to develop and integrate
ontologies?
• How to provide facts?
• Reasoning?
• Learning?
• Guarantees?
7
What does a knowledge graph do for us?
What are the difficulties?
Example from medical project:
• Foundational Model of Anatomy
• RadLex
• ICD-10
8. Queries
Ontologies & Facts
What can be represented?
• Provenance
• Uncertainty
• Time
• …
8
What does a knowledge graph do for us?
What are the difficulties?
Example from medical project:
• Patient history
• patient measurements
11. 02.11.2020
Steffen Staab, Universität Stuttgart, @ststaab, https://www.ipvs.uni-stuttgart.de/departments/ac/ 11
Wonderful ressource
– but not representative
13. Application 2: KG for Circular Factory
Product
Production
Co-
Design
Knowledge Graph contains knowledge about design, production and product
including plans, sensor measurements and intra-logistics
17. • Updates and deletions with dependencies [EKAW18],
also at the ontological level [KR2020]
• Federation [WWW08]
• Lacking views with deletions and updates
• Transaction locking [ESWC2013]
• Lacking recent standards (SHACL) and optimistic schemes
• Uncertainties
• Managing identities
(„does re-designed column preserve its identity?“)
• ...
17
Applications Imply Wealth of Requirements
rudimentary
available
research
18. Encyclopedic KGs
• Facts are reported often
• Who is Douglas Adams?
• What is the capital of France?
• Head of distribution of world
knowledge on the Web
• Answers with high precision
retrieval desired
Engineering KGs
• Point facts exist once
• w3476 instOf AngleGrinder
• faceGear4223
maxDeviation 0.3mm
• Processes are important
• Answers must be correct
18
Sliding scale of knowledge graph requirements
Currently fashionable
research
“we build a system”
under-researched
19. A lot of research in Knowledge Graphs builds on the
assumption that we want to query encyclopedias
but we have many other requirements in industry.
19
Observation 1
21. KG 1 KG 2 KG 3
App A App B App C
Scenario in Architecture, Engineering, Construction
(AEC)
22. 22
SOLID Project
https://solidproject.org/
• people store their data
securely in decentralized
data stores - Pods
• people control access to
the data in their Pod
• standard, open, and
interoperable data
formats and protocols
Focus:
authentication &
authorization
23. KG 1 KG 2 KG 3
App A App B App C
Can my app B work on my KG2?
24. Example: How old are the students?
Query for all students, access age
Query fails during evaluation
let students = query { SELECT ?x WHERE {?x a Student. } }
for student in students do
printfn „%A“ (student.age)
bob
alice 𝑏1
Student
University
subClass
type
studiesAt
type
211... "Bob"
matrNr name
25 "Alice"
age name
Person
[ESOP17,ISWC19]
25. Example: How old are the students?
let students = query { SELECT ?x WHERE {?x a Student. } }
for student in students do
printfn „%A“ (student.age)
Should we use this relation on this signifier?
Depends on:
1. Conceptualization of data source
2. Querying of data source
3. Software code
bob
alice 𝑏1
Student
University
subClass
type
studiesAt
type
211... "Bob"
matrNr name
25 "Alice"
age name
Person
26. Closed-world conceptualization of classes and relations
SHACL – SHApes Constraint Language
• SHACL shapes are integrity constraints
• Namespaces omitted for brevity
:StudentShape a :NodeShape;
:targetClass :Student;
:class :Person;
:property [
:path :studiesAt;
:minCount 1;
:class :University;
].
:PersonShape a :NodeShape;
:targetClass :Person;
:property [
:path :name;
:minCount 1;
:datatype xsd:string;
].
27. Closed-world conceptualization of code (1)
Type checking discovers (potential) run-time errors
let students = query { SELECT ?x WHERE {?x a Student. } }
for student in students do
printfn „%A“ (student.age)
Set of all students (StudentShape)
One value of
StudentShape
set
Not allowed since
StudentShape ⊈ ≥𝟏age.⊤
when considering
all conceptually possible RDF graphs
28. Closed-world conceptualization of code (2)
• Access: matrNr
• No error during evaluation
• Unsafe: Rejected by type checking,
conceptualization not guaranteed
let students = query { SELECT ?x WHERE {?x a Student. } }
for student in students do
printfn „%A“ (student.matrNr)
bob
alice 𝑏1
Student
University
subClass
type
studiesAt
type
211... "Bob"
matrNr name
25 "Alice"
age name
Person
29. Closed-world conceptualization of code (3)
• Query for: matrNr
• Type safe access:
matrNr inferred to be given for all values of student
let students = query { SELECT ?x WHERE {?x matrNr ?y. } }
for student in students do
printfn „%A“ (student.matrNr)
bob
alice 𝑏1
Student
University
subClass
type
studiesAt
type
211... "Bob"
matrNr name
25 "Alice"
age name
Person
[ESOP17,ISWC19]
30. 1. Use available SHACL constraints
2. Infer additional SHACL constraints from queries
3. Type check using inference
Determine type safety
let students = query { SELECT ?x WHERE {?x a Student. } }
for student in students do
printfn „%A“ (student.name)
Query shape(2) including StudentShape (1)
One value of
StudentShape
set
StudentShape ⊆ PersonShape and
PersonShape ⊆ ≥1name. ⊤ in all possible graphs
Inference (3)
[ESOP17,ISWC19]
31. KG 1 KG 2 KG 3
App A App B App C
Scenario: Can my app B work on my view of KG1?
32. 32
Shapes to Shapes
KG 1 KG 2
App B
Input Shape
Sin = { :Person ⊑ :Agent
}
Input query defining view
q = CONSTRUCT {
?x a :Person .
?y a :Agent
} WHERE {
?x a :Person .
?y a :Agent
“Every Person
is an Agent” Output Shapes
“Which data can App B expect?”
s2s(Sin, q) → Sout
view
[Seifer2023]
33. Tracing Query Concepts (and Relations)
Sin = { :Person1 ⊑
:Agent }
q = CONSTRUCT {
?x a :Person3 .
?y a :Agent
} WHERE {
?x a :Person2 .
?y a :Agent
}
Sout = { :Person3 ⊑
:Agent }
Are concepts
:Person1
:Person2
:Person3
the same?
33
Yes!
[Seifer2023]
34. Tracing Query Concepts (and Relations)
Sin = { :Person1 ⊑ :Agent
}
q = CONSTRUCT {
?x a :Person3 .
?y a :Agent
} WHERE {
?x a :Person2 .
?x a :Teacher .
?y a :Agent
}
Sout = { :Person3 ⊑ :Agent
}
Are concepts
:Person1
:Person2
:Person3
still the same?
34
NO!
[Seifer2023]
Hard problem even for restricted
query and constraint languages
35. KG problems occur at ontological and at fact level.
Knowledge Graph technologies lack crucial capabilities
for guaranteeing results.
35
Observation 2
39. 39
Finding and Exploiting Patterns of Similarity & Analogy
Stuttgart
Area
worksFor
locatedIn
Koblenz
Area
Wolv.
Area
Steffen
Frank
Ingo
birthdate
livesIn
prediction impossible prediction possible
40. Correct [2013]:
“TransE significantly outperforms state-of-the-art methods in link
prediction on two knowledge bases.”
Misleading:
“Our work focuses on modeling multi-relational data from KBs
(Wordnet [9] and Freebase [1] in this paper), with the goal of
providing an efficient tool to complete them by automatically
adding new facts, without requiring extra knowledge.”
A. Bordes et al. [TransE 2013]
Knowing More than What is Stated in a Knowledge Graph
41. Geometric Reasoning with EL Ontology A-Box
Concept assertion 𝐶(𝑎)
𝑎
𝐶
Geometric membership
[ISWC2022]
42. 42
Geometric Reasoning with EL Ontology T-Box
Box affine
transformation
Box entailment Box intersection
Box disjointedness
[ISWC2022]
43. Geometric Reasoning with EL Ontology A-Box
4
3
Concept assertion 𝐶(𝑎)
𝑎
𝐶
r(𝑎, 𝑏)
𝑇𝑟
𝑏
𝑎
Role assertion
Geometric membership
Affine transformation
between two points
[ISWC2022]
44. 44
Geometric Reasoning with Fact Attributions in ShrinkE
• Modeling primal triple as a spatial spanning (from a point to a box)
• Modeling qualifiers as a spatial (monotonically) shrinking of the box
• Qualifier implication and exclusion are geometrically modeled as
box containment and disjointedness
[ACL2023]
Check out https://kg-beyond-triple.github.io/
45. 45
Geometric Reasoning with Fact Attributions in ShrinkE
[ACL2023]
• Box embedding
• Box shrinking is a box-to-box transform that monotonically shrinks the size
46. • WD50k: excerpt from Wikidata
• JF17K: excerpt from Freebase
• WikiPeople: excerpt from Wikidata
• FB15k-237: excerpt from Freebase
• …
Datasets for evaluating knowledge graph embeddings
Many datasets, but all biased in the same direction
47. 02.11.2020
PhD thesis in preparation by Fabian Sasse, KIT 47
Selecting manufacturing measurement technology
in immature production processes
49. Knowledge Graph embedding techniques
do not complete knowledge graphs,
they perform similarity and analogical reasoning.
Evaluations of Knowledge Graph embedding methods
remain biased towards encyclopedic knowledge.
49
Observation 3
54. Statistically frequent knowledge
• Commonsense knowledge:
• “cows eat grass”
• “apples fall towards earth if
unsupported”
• Commonsense expert
knowledge
• “halting problem is undecidable”
• “3SAT is NP-complete”
“Point knowledge”
• Steffen Staab is a professor
at University of Stuttgart
54
Knowledge in text
55. • Smoothing a data set: create an approximating function that
preserves patterns in the data, while leaving out noise or fine-
scale structures. [Shortened from Wikipedia]
• Laplacian smoothing for Naïve Bayes:
argmax𝑐 𝑃 𝑐 𝑥1, … , 𝑥𝑛 ≈ argmax𝑐𝑃 𝑐 𝑃 𝑥1 𝑐 ⋯ 𝑃 𝑥𝑛 𝑐
• Smoothing for language models [ACL14]
𝑃 𝑤𝑛 𝑤𝑛−𝑘 ⋯ 𝑤𝑛−1
must not be 0 for unobserved 𝑤𝑛−𝑘 ⋯ 𝑤𝑛−1 𝑤𝑛
55
Language models smoothen probability distributions
must not be 0
56. • What other terms could appear in a masked position?
• High “temperature” → diversity of answers
• Varying answers for “Write a poem about <your name>”
56
Smoothing is the core task of Large Language Models
63. 63
Few shot in context learning on KB question answering
Tianle Li, Xueguang Ma, Alex Zhuang, Yu Gu, Yu Su,
and Wenhu Chen. 2023. Few-shot In-context Learning
on Knowledge Base Question Answering. In ACL-2023
67. Knowing for Sure
• Research required for
dealing with federated,
overlapping KGs with
multiple authorities
Knowing More
• Know what you get and
evaluate not only with
encyclopedic KGs
LLMs as knowledge bases
• Commonsense knowledge
• Frequently observed
knowledge
LLMs as AI assistants
• entering and retrieving
“point knowledge”
Do not (always) go with the flow
68. Thank you!
E-Mail
www.
Universität Stuttgart
KI – Institute for Artificial Intelligence
Universitätsstraße 32, 70569 Stuttgart
Steffen Staab
ki.uni-stuttgart.de
Analytic Computing, KI
Steffen.staab@ki.uni-stuttgart.de
Many thanks go to my
PhD students, PostDocs and
collaborators who made the work
possible portrayed in this talk
check out references!
I hire
PostDoc & PhD student
for circular factory project!
69. 1. [Potyka23] Nico Potyka, Yuqicheng Zhu, Evgeny Kharlamov and Steffen Staab.
Uncertainty-aware Knowledge Extraction from Large Language Models using
Social Choice Theory. TechReport.
2. [ISWC2022] B. Xiong, N. Potyka, T.-K. Tran, M. Nayyeri, S. Staab. For “Faithful
Embeddings for EL++ Knowledge Bases”. In: 21st International Semantic Web
Conference (ISWC2022)
3. [SIGIR23] J. Lu, J. Shen, B. Xiong, W. Ma, S. Staab, C. Yang. HiPrompt: Few-
Shot Biomedical Knowledge Fusion via Hierarchy-Oriented Prompting. In:
Proceedings of ACM SIGIR-2023, Taipei, Taiwan, July 23-27, 2023.
4. [ISWC2023] M. Nayyeri, Z. Wang, M. M. Akter, M. Mohtashim, Md R. Al Hasan
Rony, J. Lehmann, S. Staab. Integrating Knowledge Graph Embeddings and Pre-
trained Language Models in Hypercomplex Spaces. In: 22nd Int. Semantic Web
Conference (ISWC2023), Athens, GR, November 6-10, 2023.
5. [TransE 2013] Bordes, Antoine, et al. "Translating embeddings for modeling
multi-relational data." Advances in neural information processing systems 26
(2013).
References related to Knowing More
70. 1. [ISWC19] M. Leinberger, P. Seifer, C. Schon, R. Lämmel, S. Staab. Type Checking Program Code using SHACL. In: Proc.
of Int. Semantic Web Conference (ISWC-2019). Auckland, New Zealand, October 2019.
2. [Seifer2023] Philipp Seifer, Daniel Hernández, Ralf Lämmel, Steffen Staab. From Shapes to Shapes: Inferring SHACL
Shapes for Results of SPARQL CONSTRUCT Queries. TechReport.
3. [ESOP17] M. Leinberger, R. Lämmel, S. Staab. The essence of functional programming on semantic data. In 26th
European Symposium on Programming (ESOP 2017), Uppsala, SE, 22 - 29 Apr 2017, pp. 750-776.
4. [CAAD Futures 2023] D. Elshani, D. Hernandez, A. Lombardi, L. Siriwardena, T. Schwinn, A. Fisher, S. Staab, A. Menges,
T. Wortmann. Building Information Validation and Reasoning Using Semantic Web Technologies. In: Computer-Aided
Architectural Design. CAAD Futures 2023. Springer, Cham, 2023.
5. [KR2020] T. Rienstra, C. Schon, S. Staab. Concept Contraction in the Description Logic EL. In: Principles of Knowledge
Representation and Reasoning: Proceedings of the Seventeenth International Conference, KR 2020, pp. 723-732.
6. [EKAW18] C. Schon, S. Staab, P. Kügler, P. Kestel, B. Schleich, S. Wartzack. Metaproperty-guided Deletion from the
Instance-Level of a Knowledge Base. In: Proc. of EKAW 2018, 21st International Conference on Knowledge Engineering
and Knowledge Management, November 12-16, 2018, Nancy, France, Springer 2018.
7. [ESWC2013] S. Scheglmann, S. Staab, M. Thimm, G. Gröner. Locking for Concurrent Transactions on Ontologies. In: 10th
Extended Semantic Web Conference (ESWC2013), Montpellier, France, May 26-30, 2013.
8. [WWW08] S. Schenk, S. Staab. Networked Graphs: A Declarative Mechanism for SPARQL Rules, SPARQL Views and RDF
Data Integration on the Web. In: Proc. of WWW-2008, 17th Int. World Wide Web Conference, Bejing, China, April 21-25,
2008, pp. 585-594.
References related to Knowing for Sure
71. [ACL14] R. Pickhardt, T. Gottron, M. Körner, P. G. Wagner, T. Speicher, S. Staab. A Generalized Language Model as the
Combination of Skipped n-grams and Modified Kneser Ney Smoothing. In: Proc. of ACL-2014 - The 52nd Annual Meeting of the
Association for Computational Linguistics. Baltimore, June 22-27, 2014.
02.11.2020
Steffen Staab, Universität Stuttgart, @ststaab, https://www.ipvs.uni-stuttgart.de/departments/ac/ 71
Others
Editor's Notes
If it looks like a duck, walks like a duck and quacks like a duck, then it just may be a duck.
Huey, Dewey, and Louie live in Duckburg
If it looks like a duck, walks like a duck and quacks like a duck, then it just may be a duck.
Huey, Dewey, and Louie live in Duckburg
If it looks like a duck, walks like a duck and quacks like a duck, then it just may be a duck.
Huey, Dewey, and Louie live in Duckburg
750 million triples, fast growing, not easy to manage
status: proposal for funding by 20 PIs, mostly engineering, mostly from KIT
7 year excellence cluster at Uni Stuttgart
medical knowledge graphs and applications may be found on either side
Now I am gonna to present those geometric interpretations and the corresponding loss term for each axiom.
In Abox, we have two types of axioms: Concept assertion and Role assertion r(a, b).
For concept assertion, the geometric interpretation is that the point of instance a should be inside the box of the class C. That means that our loss should enforce every dimension of the point a to be between the low-left corner of box C and upper-right corner of box C.
We also have role assertion r(a, b) saying that a has a relation r with b, the geometric interpretation is that the point a, after a affine transformation of r, should be near the point of b.
The corresponding loss term can be defined by minimizing the L2 distance between the transformed point of a and the point b.
We proved that our terms satisfy the soundness guarantees that means our loss terms are zero if and only if the corresponding geometric interpretations are satisfied.
Now I am gonna to present those geometric interpretations and the corresponding loss term for each axiom.
In Abox, we have two types of axioms: Concept assertion and Role assertion r(a, b).
For concept assertion, the geometric interpretation is that the point of instance a should be inside the box of the class C. That means that our loss should enforce every dimension of the point a to be between the low-left corner of box C and upper-right corner of box C.
We also have role assertion r(a, b) saying that a has a relation r with b, the geometric interpretation is that the point a, after a affine transformation of r, should be near the point of b.
The corresponding loss term can be defined by minimizing the L2 distance between the transformed point of a and the point b.
We proved that our terms satisfy the soundness guarantees that means our loss terms are zero if and only if the corresponding geometric interpretations are satisfied.
[Hypertext2008]
Yulan talked in her keynote about voting in order to improve confidence – though I also have observed non-i.i.d. behaviour and then voting may be bad
usefulness may be an issue
In the following I will give you webpage content about a soccer club. Represent the facts that you find in this text in RDF turtle notation. Effizienter VfB siegt bei Union Berlin Die Siegesserie des VfB geht weiter. Beim 1. FC Union Berlin setzt sich die Mannschaft mit dem Brustring mit 3:0 durch. Es ist der sechste Erfolg in Serie und der erste gegen Union in der Bundesliga. Der Spielverlauf: Der VfB ging mit einer auf zwei Positionen veränderten Startformation in das Duell beim 1. FC Union Berlin. Maxi Mittelstädt und Dan-Axel Zagadou begannen für Pascal Stenzel sowie Hiroki Ito (beide Bank). Die Mannschaft mit dem an diesem Tag schwarzen Brustring startete selbstbewusst in die Partie und hatte in der Anfangsviertelstunde deutlich mehr Ballbesitz. Die höheren Spielanteile münzte der VfB schnell in die verdiente Führung um. Wer sonst als Serhou Guirassy hätte der Torschütze zum 1:0 sein können (siehe „Die Tore“). Die Jungs aus Cannstatt kontrollierten die Partie auch in der Folge, musste nach knapp einer halben Stunde aber schon wechseln. Serhou Guirassy verließ den Platz angeschlagen mit muskulären Problemen im hinteren linken Oberschenkel, Deniz Undav kam für ihn in die Partie. Der VfB war dennoch bis zum Pausenpfiff das tonangebende Team. Silas und Deniz Undav sorgen für die Entscheidung Nach dem Wiederanpfiff entwickelte sich eine umkämpfte Partie mit vielen Situationen zwischen den Strafräumen. Klare Torchancen konnte sich zunächst keines der Teams erspielen. In der 60. Minute hatte Jamie Leweling jedoch die große Chance, auf 2:0 zu erhöhen. Der 22-Jährige scheiterte in aussichtsreicher Position frei vor dem Tor an Unions Torhüter Frederik Rönnow. In der 77. Minute war Alexander Nübel auf der Gegenseite hellwach und klärte die Situation gegen den heranstürmenden Kevin Behrens. Kurz darauf sorgte der VfB mit einem Konter für das beruhigende 2:0. Der eingewechselte Silas war mit seinem dritten Saisontor erfolgreich. Den Endstand zum 3:0 stellte Deniz Undav mit einem Kopfball her. Der VfB siegte am Ende verdient, weil er seine Chancen konsequent nutzte und über die gesamte Spielzeit hinweg kaum Chancen des Gegners zuließ. Den gesamten Spielverlauf im VfB-Liveticker nachlesen. Die Tore: 16. Minute: Serhou Guirassy köpft nach einer Flanke von Anthony Rouault zum 1:0 ein. Es ist das 14. Saisontor des VfB-Stürmers. 81. Minute: Silas kommt über Karazor und Millot an den Ball, setzt sich gegen die aufgerückten Union-Verteidiger durch und schließt überlegt zum 2:0 ab. 88. Minute: Der VfB erobert in Höhe des gegnerischen Strafraums den Ball, Wooyeong Jeong flankt von rechts auf Deniz Undav, der zentral zum 3:0 einköpft. Die Stimmen: VfB-Cheftrainer Sebastian Hoeneß: „Es war eine reife Leistung von uns. Wir haben sehr erwachsen gespielt. Die Druckphasen des Gegners waren nie so sehr ausgeprägt. Dass wir das Spiel am Ende so klar auf unsere Seite ziehen, macht mich stolz. Wir haben aktuell einen Lauf und den wollen wir so lange wie möglich mitnehmen.“ Chris Führich: „Es war eine Riesenteamleistung von uns. Wir wussten, wie schwierig es ist, hier zu gewinnen. Wir haben von der ersten bis zur letzten Minute unsere Taktik durchgezogen. Es ist auch sehr wichtig gegen die lange Bälle und wuchtigen Spieler von Union gut zu stehen. Das ist uns gut gelungen und wir haben letztlich auch verdient gewonnen.“ Maxi Mittelstädt: „Es ist ein schönes Gefühl, dass wir gewonnen haben. Wir haben eine reife Leistung gezeigt und wenig anbrennen lassen. Natürlich gab es auch Phasen, die wir überstehen mussten. Wir haben einen breiten und starken Kader. Auch heute war wichtig, welche Impulse die Einwechselspieler in die Partie gebracht haben. Das Kollektiv macht uns aktuell stark. Manchmal muss man sich angesichts der Siegesserie kneifen, aber wir haben uns das auch über die vergangenen Monate und Wochen erarbeitet. Ich freue mich auf die kommenden Herausforderungen.“ Die Besonderheiten: Der gebürtige Berliner Maximilian Mittelstädt gab sein Startelfdebüt für den VfB. Neu-Nationalspieler Chris Führich machte an diesem Samstag sein 75. Pflichtspiel im Trikot mit dem roten Brustring. Der ehemalige VfBler Rani Khedira gab auf Seiten der Berliner gegen den Club aus Cannstatt sein Comeback nach einer längeren Wadenverletzung. Der Schiedsrichter der Partie Bastian Dankert und seine Assistenz René Rohde leiteten ihr jeweils 150. Bundesligaspiel. 21 Punkte nach acht Spieltagen hat der VfB in seiner Vereinshistorie noch nie auf dem Konto gehabt. Die nächsten Spiele: Am kommenden Samstag empfängt der VfB die TSG Hoffenheim in der MHPArena. Dieses Spiel ist bereits ausverkauft, ebenso wie die Heimpartie gegen Borussia Dortmund am 11. November. Der Mitgliedervorverkauf für das Pokal-Heimspiel gegen den 1. FC Union Berlin am Dienstagabend, 31. Oktober, 18 Uhr läuft, genauso wie für die Heimbegegnung gegen den SV Werder Bremen am Samstag, 2. Dezember, 18:30 Uhr. Zum VfB-Onlineshop.