After the amazing breakthroughs of machine learning (deep learning or otherwise) in the past decade, the shortcomings of machine learning are also becoming increasingly clear: unexplainable results, data hunger and limited generalisability are all becoming bottlenecks.
In this talk we will look at how the combination with symbolic AI (in the form of very large knowledge graphs) can give us a way forward, towards machine learning systems that can explain their results, that need less data, and that generalise better outside their training set.
--
Frank van Harmelen leads the Knowledge Representation & Reasoning group in the CS Department of the VU University Amsterdam. He is also Principal investigator of the Hybrid Intelligence Centre, a 20Μ€, 10 year collaboration between researchers at 6 Dutch universities into AI that collaborates with people instead of replacing them.
--
Unleash Your Potential - Namagunga Girls Coding Club
Systems that learn and reason | Frank Van Harmelen
1. Systems that learn and reason
Frank van Harmelen
Knowledge Representation & Reasoning Group
Vrije Universiteit Amsterdam
Creative Commons License
CC BY 3.0:
Allowed to copy, redistribute
remix & transform
But must attribute
1
2. Direction of travel in modern AI
“making artificial intelligence more human-
centred”
AI systems should “support people and be
competent partners”
But:
“current AI systems are often incompetent
because
they lack background and contextual
knowledge” 2
3. So what’s holding AI back?
For a long time,
AI researchers have locked themselves
in one of two towers 3
“current AI systems are often incompetent because
they lack background and contextual knowledge”
and “they cannot explain their actions”
4. AI: a tale of
two towers?
Statistical AI Symbolic AI
10. Strengths & Weaknesses
Symbolic Connectionist
Construction Human effort Data hunger
Scaleable +/- +/-
Explainable + -
Generalisable Performance cliff Performance cliff
10
11. Strengths & Weaknesses
Symbolic Connectionist
Construction Human effort Data hunger
Scaleable +/- +/-
Explainable + -
Generalisable Performance cliff Performance cliff
Very large knowledge graphs:
Billions of facts and rules,
But expensive to build & maintain:
- Crowd sourcing
11
12. Strengths & Weaknesses
10M training samples
Symbolic Connectionist
Construction Human effort Data hunger
Scalable +/- +/-
Explainable + -
Generalisable Performance cliff Performance cliff
4.8M training games
12
13. Strengths & Weaknesses
Symbolic Connectionist
Construction Human effort Data hunger
Scaleable +/- +/-
Explainable + -
Generalisable Performance cliff Performance cliff
worse with
more data
worse with
less data
13
14. Strengths & Weaknesses
Symbolic Connectionist
Construction Human effort Data hunger
Scaleable +/- +/-
Explainable + -
Generalisable Performance cliff Performance cliff
15. Strengths & Weaknesses
Symbolic Statistical
Construction Human effort Data hunger
Scaleable +/- +/-
Explainable + -
Generalisable Performance cliff Performance cliff
15
“black box problem”
16. Strengths & Weaknesses
Symbolic Connectionist
Construction Human effort Data hunger
Scaleable +/- +/-
Explainable + -
Generalisable Performance cliff Performance cliff
quality
generality 16
17. Strengths & Weaknesses
Symbolic Connectionist
Construction Human effort Data hunger
Scaleable +/- +/-
Explainable + -
Generalisable Performance cliff Performance cliff
17
Class: 793
Label: n04209133 (shower cap)
Certainty: 99.7%
Problem:
“out of sample generalizability”
18. Strengths & Weaknesses
Symbolic Connectionist
Construction Human effort Data hunger
Scaleable +/- +/-
Explainable + -
Generalisable Performance cliff Performance cliff
18
27. Context aware ML using KR
(informed ML)
See survey of 100+ systems in Von Rueden et al, Learning, 2019
flower?
cushion?
“Parts of a chair are:
cushion and armrest”
“Given the context of chair,
a cushion is much more likely
than a flower”
P(cushion|chair) >> P(flower|chair)
37. 1. Machine Learning can benefit from injecting
symbolic knowledge (knowledge graphs)
• Explanations
• Ranking hypotheses
• Transfer learning
• Sample efficiency
• Pipeline engineering
2. The required symbolic knowledge is available
at very large scale + corresponding tools
“Why learn what you already know?”
37