Machine Learning (ML) and Artificial Intelligence (AI) have made great strides in this decade. We have a plethora of ML algorithms that can be used to perform a given task, be it face recognition, image classification or natural language processing. However, explainability of ML/AI algorithms remains a big problem. Explainable AI (XAI) is a branch of ML that is devoted to unravelling the black-box nature of AI so that we understand the reasons behind the decisions/output. However, there are concerns that XAI sometimes produce “tools for computer scientists to explain things to other computer scientists”, which defeats its purpose. To this end, a growing number of researchers have called for integration with social sciences to make truly explainable and trustworthy AI, because philosophy and social sciences have debated the meaning and function of an explanation for millennia and have deeper insights1. In this talk, we present such an integration2.
Our problem domain is algorithm evaluation, which considers a portfolio of algorithms and its performance on a set of problems. For example, it can be a portfolio of regression algorithms. The goal is to understand meaningful, explainable insights about the algorithms from the performance results. As the social science linkage, we use Item Response Theory (IRT), a methodology from educational psychometrics. IRT is traditionally used to evaluate the difficulty and discrimination of test questions and the ability of students and has causal interpretations. Using IRT we obtain explainable insights about algorithms relating to their stable/consistent nature, the difficulty level of problems they can handle and their behaviour. In addition, we visualise the problem spectrum and find regions on the spectrum where algorithms exhibit strengths. The causal interpretations of IRT transfer to the algorithm evaluation domain as we gain a deeper understanding of algorithms.
References
1. Miller, T. Explanation in artificial intelligence: Insights from the social sciences. Artif Intell 267, 1–38 (2019).
2. Kandanaarachchi, S. & Smith-Miles, K. Comprehensive Algorithm Portfolio Evaluation using Item Response Theory. Journal of Machine Learning Research 24, 1–52 (2023).
2. To explain or to predict? – Galit Shmueli
● Paper by Shmueli in 2010
● Talks about these two topics
● Argues that explanation and prediction are different
● Two modelling paths – for predicting and explaining
● Social sciences have done explanatory models for a long time
3. What is an explanation?
• To explain an event is to provide some information about its causal history. –
Lewis, 1986 (Causal Explanation)
• A statement or an account that makes something clear – Google
• It is important to note that the solution to explainable AI is not just ‘more AI’ -
Miller, 2019
• Miller (2019) argues for Social Science + Computer Science in XAI
In the fields of philosophy, cognitive psychology/science, and social
psychology, there is a vast and mature body of work that studies these exact
topics.
3
4. Message: Bring in the
social scientists to the
party!
Integrate their methods!
4
5. What is this talk about?
● Using a method in social sciences to do two ML type tasks
● Evaluate algorithms
● It gives us more meaningful metrics about algorithms
● Has some causal interpretations
● Visually inspect where algorithms perform well (and poorly)
● Ensembles
● Anomaly detection ensembles
6. What is algorithm evaluation?
• Performance of many algorithms to many problems
• How do you explain the algorithm performance?
• Standard statistical analysis misses many things
Algo 1 Algo 2 Algo 3 Algo 4
Problem 1
Problem 2
Problem 3
Problem 4
Problem 5
6
7. We want to evaluate algorithms
performance . . .
. . . in a way that we understand algorithms
and problems better!
7
9. Item Response Theory
(IRT)
• Modelsusedinsocialsciences/psychometrics
• Unobservablecharacteristicsandobserved
outcomes
• Verbalormathematicalability
• Racialprejudiceor stressproneness
• Politicalinclinations
• Intrinsic“quality”thatcannotbemeasured
directly
9
This Photo by Unknown Author is licensed under CC BY-SA
10. IRT in education
• Finds the discrimination and difficulty of test
questions
• And the ability of the test participants
• By fitting an IRT model
• In education – questions that can discriminate
between students with different ability is preferred
to “very difficult” questions.
10
11. How it works
11
Questions
Students Q 1 Q 2 Q 3 Q 4
Stu 1 0.95 0.87 0.67 0.84
Stu 2 0.57 0.49 0.78 0.77
Stu n 0.75 0.86 0.57 0.45
IRT Model
Discrimination of
questions
Difficulty of questions
Ability of students (latent trait)
Matrix
βj
αj
θi
12. What does IRT give us?
• Q1 - discrimination , difficulty
• Q2 - discrimination , difficulty
• Q3 - discrimination , difficulty
• Q4 - discrimination , difficulty
• Student 1 ability
︙
• Student n ability
α1 β1
α2 β2
α3 β3
α4 β4
θ1
θn
Q 1 Q 2 Q 3 Q 4
Stu 1 0.95 0.87 0.67 0.84
Stu 2 0.57 0.49 0.78 0.77
Stu n 0.75 0.86 0.57 0.45
12
18. Fitting the IRT model
• Maximising the expectation
• - discrimination parameter of algorithm
• - scaling parameter for the algorithm
• - difficulty parameter for the algorithm
• - score of the algorithm on the dataset/problem
• - prior probabilities
Eθ|Λ(t),Z [ln p (Λ|θ, Z)] = N
n
∑
j=1
(ln αj + ln γj) −
1
2
N
∑
i=1
n
∑
j=1
α2
j ((βj + γjzij − μ(t)
i )
2
+ σ(t)2
)
+ ln p (Λ) + const
αj
γj
βj
zij
Λ
19. New meaning of IRT parameters?
● Discrimination -> anomalousness, algorithm consistency
● Difficulty -> algorithm difficulty limit
● Student ability -> Dataset difficulty spectrum
20. What can we say . . .
• Consistent algorithms give similar performance for easy or hard datasets
• Algorithms with higher difficulty limits can handle harder problems
• Anomalous algorithms give bad performance for easy problems and good
performance for difficult problems
20
21. Dataset difficulty
●
● Is a function of discrimination, difficulty and scores
dataset i di
ffi
culty = −
∑j
̂
α2
j (
̂
βj + ̂
γjzij)
∑j
̂
α2
j
22. Example – fitting IRT model
• Anomaly detection algorithms
• 8 anomaly detection algorithms
• 3142 datasets
• Our performance metric is AUROC (looking at the performance vs actual)
22
25. Focusing on algorithm performance
• We have algorithm performance (y axis)
and problem difficulty (x axis)
• We can fit a model and find how each
algorithm performs
• We use smoothing splines
• Can visualize them
• No parameters to specify
25
29. AIRT performs well, when . . .
• The set of algorithms is diverse.
• Ties back to IRT basics
• IRT in education – If all the questions are equally discriminative and difficult,
IRT doesn’t add much
• IRT useful when we have a diverse set of questions and we want to know
• Whichquestionsaremorediscriminative
• Whichquestionsaredifficult
29
32. IRT to build an ensemble
● Previous work was using performance - marks
● What about using original responses?
● Like survey questions
● Rosenberg's Self-Esteem Scale - I feel I am a person of worth
(Strongly agree/Agree/Neutral/Disagree/Strongly disagree)
● No right or wrong answer
● Latent trait gives the person’s self esteem
● Latent trait uncovers the “hidden quality”
33. Unsupervised algorithms
● Instead of performance values what if you have original responses?
Q 1 Q 2 Q 3 Q 4
Stu 1 0.95 0.87 0.67 0.84
Stu 2 0.57 0.49 0.78 0.77
Stu n 0.75 0.86 0.57 0.45
latent value (i) = −
∑j
̂
α2
j (
̂
βj + ̂
γjzij)
∑j
̂
α2
j
34. What is an anomaly detection ensemble?
Dataset
Unsupervised
AD methods
• The AD methods are heterogenous methods
• Ensembles – use existing methods to come up with better anomaly detection/
scores
AD ensemble
Ensemble
Score
35. Anomaly detection ensembles
● Latent trait = anomalousness of the observations = the ensemble score
●
, a weighted score of original responses
●
ensemble score of observa
ti
on (i) = −
∑j
̂
α2
j (
̂
βj + ̂
γjzij)
∑j
̂
α2
j
40. New AIRT parameters
●
●
●
● Where M, V and C denote mean, variance, and covariance terms
γ(t+1)
j
=
V (μ(t)
i ) + σ(t)2
Cj (zij, μ(t)
i )
β(t+1)
j
= M (μ(t)
i ) − γ(t+1)
j
Mj (zij)
α(t+1)
j
=
(
γ(t+1)2
j
Vj(zij) − V (μ(t)
i ) − σ(t)2
)
−1/2
41. Explaining notation
●
●
●
●
●
Mj (zij) =
∑i
zij
N
M (μ(t)
i ) =
∑i
μ(t)
i
N
V (zij) =
∑i
z2
ij
N
− Mj (zij)
2
V (μ(t)
i ) =
∑i
μ(t)2
i
N
− M (μ(t)
i )
2
Cj (zij, μ(t)
i ) =
∑i
zijμ(t)
i
N
− Mj (zij) M (μ(t)
i )
42. More notation
●
● is the th iteration
●
●
p (θi |Λ(t)
, zi) =
𝒩
(θi |μ(t)
i
, σ(t)2
)
Λ(t)
= (λ1
(t)
, …, λn
(t)
) , λj
(t)
= (α(t)
j
, β(t)
j
, γ(t)
j )
T
t
σ(t)2
=
∑
j
α(t)2
j
+ σ−2
−1
μ(t)
i
= σ(t)2
∑
j
α(t)2
j (β(t)
j
+ γ(t)
j
zij) + μ
43. Algorithm portfolio selection
• Can use algorithm strengths to select a good portfolio of algorithms
• We call this portfolio airt portfolio
• airt – Algorithm IRT (old Scottish word – to guide)
43
45. What happens to the IRT parameters?
• IRT - ability of student
• As increases probability of a higher
score increases
• What is in terms of a dataset?
• easiness of the dataset
• Dataset difficulty score
θi
θi
θi
−θi
45
46. Discrimination parameter
• Discrimination of item
• increases slope of curve increases
• What is in terms of an algorithm?
• - lack of stability/robustness of
algo
•
Consistency of algo
αj
αj
αj
αj
1
|αj |
46
47. Consistent algorithms
• Education – such a question doesn’t
give any information
• Algorithms – these algorithms are
really stable or consistent
•
Consistency =
1
|αj |
47
48. Anomalous algorithms
• Algorithms that perform poorly on
easy datasets and well on difficult
datasets
• Negative discrimination
• In education – such items are
discarded or revised
• If an algorithm anomalous, it is
interesting
• Anomalousness = = sign(αj)
This Photo by Unknown Author is licensed under CC BY-NC-ND
48