The document discusses prototype-based machine learning and its applications in bio-medical domains. It provides an overview of unsupervised and supervised prototype-based learning techniques, including competitive learning, Kohonen's self-organizing map (SOM), and learning vector quantization (LVQ). Examples of applying these methods to cluster proteomics data and identify biomarkers for rheumatoid arthritis are also mentioned.
Generalizing Scientific Machine Learning and Differentiable Simulation Beyond...Chris Rackauckas
The combination of scientific models into deep learning structures, commonly referred to as scientific machine learning (SciML), has made great strides in the last few years in incorporating models such as ODEs and PDEs into deep learning through differentiable simulation. However, the vast space of scientific simulation also includes models like jump diffusions, agent-based models, and more. Is SciML constrained to the simple continuous cases or is there a way to generalize to more advanced model forms? This talk will dive into the mathematical aspects of generalizing differentiable simulation to discuss cases like chaotic simulations, differentiating stochastic simulations like particle filters and agent-based models, and solving inverse problems of Bayesian inverse problems (i.e. differentiation of Markov Chain Monte Carlo methods). We will then discuss the evolving numerical stability issues, implementation issues, and other interesting mathematical tidbits that are coming to light as these differentiable programming capabilities are being adopted.
Bio: Dr. Chris Rackauckas is the VP of Modeling and Simulation at JuliaHub, the Director of Scientific Research at Pumas-AI, Co-PI of the Julia Lab at MIT, and the lead developer of the SciML Open Source Software Organization. For his work in mechanistic machine learning, his work is credited for the 15,000x acceleration of NASA Launch Services simulations and recently demonstrated a 60x-570x acceleration over Modelica tools in HVAC simulation, earning Chris the US Air Force Artificial Intelligence Accelerator Scientific Excellence Award. See more at https://chrisrackauckas.com/. He is the lead developer of the Pumas project and has received a top presentation award at every ACoP in the last 3 years for improving methods for uncertainty quantification, automated GPU acceleration of nonlinear mixed effects modeling (NLME), and machine learning assisted construction of NLME models with DeepNLME. For these achievements, Chris received the Emerging Scientist award from ISoP.
Generalizing Scientific Machine Learning and Differentiable Simulation Beyond...Chris Rackauckas
The combination of scientific models into deep learning structures, commonly referred to as scientific machine learning (SciML), has made great strides in the last few years in incorporating models such as ODEs and PDEs into deep learning through differentiable simulation. However, the vast space of scientific simulation also includes models like jump diffusions, agent-based models, and more. Is SciML constrained to the simple continuous cases or is there a way to generalize to more advanced model forms? This talk will dive into the mathematical aspects of generalizing differentiable simulation to discuss cases like chaotic simulations, differentiating stochastic simulations like particle filters and agent-based models, and solving inverse problems of Bayesian inverse problems (i.e. differentiation of Markov Chain Monte Carlo methods). We will then discuss the evolving numerical stability issues, implementation issues, and other interesting mathematical tidbits that are coming to light as these differentiable programming capabilities are being adopted.
Bio: Dr. Chris Rackauckas is the VP of Modeling and Simulation at JuliaHub, the Director of Scientific Research at Pumas-AI, Co-PI of the Julia Lab at MIT, and the lead developer of the SciML Open Source Software Organization. For his work in mechanistic machine learning, his work is credited for the 15,000x acceleration of NASA Launch Services simulations and recently demonstrated a 60x-570x acceleration over Modelica tools in HVAC simulation, earning Chris the US Air Force Artificial Intelligence Accelerator Scientific Excellence Award. See more at https://chrisrackauckas.com/. He is the lead developer of the Pumas project and has received a top presentation award at every ACoP in the last 3 years for improving methods for uncertainty quantification, automated GPU acceleration of nonlinear mixed effects modeling (NLME), and machine learning assisted construction of NLME models with DeepNLME. For these achievements, Chris received the Emerging Scientist award from ISoP.
“Statistical Physics Studies of Machine Learning Problems" by Lenka Zdeborova, Researcher @CNRS
Abstract : We will talk about some insight of the following questions: What makes problems studied in machine and statistical physics related? How can this relation be used to understand better the performance and limitations of machine learning systems? What happens when a phase transition is found in a computational problem? How do phase transitions influence algorithmic hardness?
Tutorial at the Winter School on Machine Learning, Gran Canaria, January 2020 (ppsx format, 52 slides)
Michael Biehl, University of Groningen, The Netherlands
Camp IT: Making the World More Efficient Using AI & Machine LearningKrzysztof Kowalczyk
Slides from the introductory lecture I gave for students at Camp IT 2019. I tried to cover artificial inteligence, machine learning, most popular algorithms and their applications to business as broadly as possible - for in-depth materials on the given topics, see links and references in the presentation.
The Positive Effects of Fuzzy C-Means Clustering on Supervised Learning Class...Waqas Tariq
Selection of inputs is one of the most substantial components of classification algorithms for data mining and pattern recognition problems since even the best classifier will perform badly if the inputs are not selected very well. Big data and computational complexity are main cause of bad performance and low accuracy for classical classifiers. In other words, the complexity of classifier method is inversely proportional with its classification efficiency. For this purpose, two hybrid classifiers have been developed by using both type-1 and type-2 fuzzy c-means clustering with cascaded a classifier. In this proposed classifier, a large number of data points are reduced by using fuzzy c-means clustering before applied to a classifier algorithm as inputs. The aim of this study is to investigate the effect of fuzzy clustering on well-known and useful classifiers such as artificial neural networks (ANN) and support vector machines (SVM). Then the role of positive effects of these proposed algorithms were investigated on applied different data sets.
The Positive Effects of Fuzzy C-Means Clustering on Supervised Learning Class...CSCJournals
Selection of inputs is one of the most substantial components of classification algorithms for data mining and pattern recognition problems since even the best classifier will perform badly if the inputs are not selected very well. Big data and computational complexity are main cause of bad performance and low accuracy for classical classifiers. In other words, the complexity of classifier method is inversely proportional with its classification efficiency. For this purpose, two hybrid classifiers have been developed by using both type-1 and type-2 fuzzy c-means clustering with cascaded a classifier. In this proposed classifier, a large number of data points are reduced by using fuzzy c-means clustering before applied to a classifier algorithm as inputs. The aim of this study is to investigate the effect of fuzzy clustering on well-known and useful classifiers such as artificial neural networks (ANN) and support vector machines (SVM). Then the role of positive effects of these proposed algorithms were investigated on applied different data sets.
Special Plenary Lecture at the International Conference on VIBRATION ENGINEERING AND TECHNOLOGY OF MACHINERY (VETOMAC), Lisbon, Portugal, September 10 - 13, 2018
http://www.conf.pt/index.php/v-speakers
Propagation of uncertainties in complex engineering dynamical systems is receiving increasing attention. When uncertainties are taken into account, the equations of motion of discretised dynamical systems can be expressed by coupled ordinary differential equations with stochastic coefficients. The computational cost for the solution of such a system mainly depends on the number of degrees of freedom and number of random variables. Among various numerical methods developed for such systems, the polynomial chaos based Galerkin projection approach shows significant promise because it is more accurate compared to the classical perturbation based methods and computationally more efficient compared to the Monte Carlo simulation based methods. However, the computational cost increases significantly with the number of random variables and the results tend to become less accurate for a longer length of time. In this talk novel approaches will be discussed to address these issues. Reduced-order Galerkin projection schemes in the frequency domain will be discussed to address the problem of a large number of random variables. Practical examples will be given to illustrate the application of the proposed Galerkin projection techniques.
Interpretable machine learning in endocrinology, M. Biehl, APPIS 2024University of Groningen
An introduction to interpretable machine learning in endocrinology.
In particular, the application of Generalized Matrix Relevance LVQ to the classification of andrenocortical tumors and the differential diagnosis of primary aldosteronism is given.
“Statistical Physics Studies of Machine Learning Problems" by Lenka Zdeborova, Researcher @CNRS
Abstract : We will talk about some insight of the following questions: What makes problems studied in machine and statistical physics related? How can this relation be used to understand better the performance and limitations of machine learning systems? What happens when a phase transition is found in a computational problem? How do phase transitions influence algorithmic hardness?
Tutorial at the Winter School on Machine Learning, Gran Canaria, January 2020 (ppsx format, 52 slides)
Michael Biehl, University of Groningen, The Netherlands
Camp IT: Making the World More Efficient Using AI & Machine LearningKrzysztof Kowalczyk
Slides from the introductory lecture I gave for students at Camp IT 2019. I tried to cover artificial inteligence, machine learning, most popular algorithms and their applications to business as broadly as possible - for in-depth materials on the given topics, see links and references in the presentation.
The Positive Effects of Fuzzy C-Means Clustering on Supervised Learning Class...Waqas Tariq
Selection of inputs is one of the most substantial components of classification algorithms for data mining and pattern recognition problems since even the best classifier will perform badly if the inputs are not selected very well. Big data and computational complexity are main cause of bad performance and low accuracy for classical classifiers. In other words, the complexity of classifier method is inversely proportional with its classification efficiency. For this purpose, two hybrid classifiers have been developed by using both type-1 and type-2 fuzzy c-means clustering with cascaded a classifier. In this proposed classifier, a large number of data points are reduced by using fuzzy c-means clustering before applied to a classifier algorithm as inputs. The aim of this study is to investigate the effect of fuzzy clustering on well-known and useful classifiers such as artificial neural networks (ANN) and support vector machines (SVM). Then the role of positive effects of these proposed algorithms were investigated on applied different data sets.
The Positive Effects of Fuzzy C-Means Clustering on Supervised Learning Class...CSCJournals
Selection of inputs is one of the most substantial components of classification algorithms for data mining and pattern recognition problems since even the best classifier will perform badly if the inputs are not selected very well. Big data and computational complexity are main cause of bad performance and low accuracy for classical classifiers. In other words, the complexity of classifier method is inversely proportional with its classification efficiency. For this purpose, two hybrid classifiers have been developed by using both type-1 and type-2 fuzzy c-means clustering with cascaded a classifier. In this proposed classifier, a large number of data points are reduced by using fuzzy c-means clustering before applied to a classifier algorithm as inputs. The aim of this study is to investigate the effect of fuzzy clustering on well-known and useful classifiers such as artificial neural networks (ANN) and support vector machines (SVM). Then the role of positive effects of these proposed algorithms were investigated on applied different data sets.
Special Plenary Lecture at the International Conference on VIBRATION ENGINEERING AND TECHNOLOGY OF MACHINERY (VETOMAC), Lisbon, Portugal, September 10 - 13, 2018
http://www.conf.pt/index.php/v-speakers
Propagation of uncertainties in complex engineering dynamical systems is receiving increasing attention. When uncertainties are taken into account, the equations of motion of discretised dynamical systems can be expressed by coupled ordinary differential equations with stochastic coefficients. The computational cost for the solution of such a system mainly depends on the number of degrees of freedom and number of random variables. Among various numerical methods developed for such systems, the polynomial chaos based Galerkin projection approach shows significant promise because it is more accurate compared to the classical perturbation based methods and computationally more efficient compared to the Monte Carlo simulation based methods. However, the computational cost increases significantly with the number of random variables and the results tend to become less accurate for a longer length of time. In this talk novel approaches will be discussed to address these issues. Reduced-order Galerkin projection schemes in the frequency domain will be discussed to address the problem of a large number of random variables. Practical examples will be given to illustrate the application of the proposed Galerkin projection techniques.
Interpretable machine learning in endocrinology, M. Biehl, APPIS 2024University of Groningen
An introduction to interpretable machine learning in endocrinology.
In particular, the application of Generalized Matrix Relevance LVQ to the classification of andrenocortical tumors and the differential diagnosis of primary aldosteronism is given.
A tutorial given at the AMALEA workshop 2022.
This talk presents the statistical physics based theory of machine learning in terms of simple example systems. As a recent application, the occurrence of phase transitions in layered networks is discussed.
The statistical physics of learning revisted: Phase transitions in layered ne...University of Groningen
"The statistical physics of learning revisted: Phase transitions in layered neural networks"
Physics Colloquium at the University of Leipzig/Germany, June 29, 2021
24 slides, ca 45 minutes
Invited lecture on Machine Learning in Medicine at the joint "Integrated Omics" course of Hanze University and University Hospital UMCG, Groningen, The Netherlands
Short presentation (15 minutes) focussing on the application of unsupervised and supervised machine learning in the paper "Tissue- and development-stage specific mRNA and heterogeneous CNV signatures of human ribosomal proteins in normal and cancer samples
Talk presented at WSOM 2016 in Houston/Texas.
Machine learning based classification of FDG-PET scan data for the diagnosis of neurodegenerative disorders
Introduction:
RNA interference (RNAi) or Post-Transcriptional Gene Silencing (PTGS) is an important biological process for modulating eukaryotic gene expression.
It is highly conserved process of posttranscriptional gene silencing by which double stranded RNA (dsRNA) causes sequence-specific degradation of mRNA sequences.
dsRNA-induced gene silencing (RNAi) is reported in a wide range of eukaryotes ranging from worms, insects, mammals and plants.
This process mediates resistance to both endogenous parasitic and exogenous pathogenic nucleic acids, and regulates the expression of protein-coding genes.
What are small ncRNAs?
micro RNA (miRNA)
short interfering RNA (siRNA)
Properties of small non-coding RNA:
Involved in silencing mRNA transcripts.
Called “small” because they are usually only about 21-24 nucleotides long.
Synthesized by first cutting up longer precursor sequences (like the 61nt one that Lee discovered).
Silence an mRNA by base pairing with some sequence on the mRNA.
Discovery of siRNA?
The first small RNA:
In 1993 Rosalind Lee (Victor Ambros lab) was studying a non- coding gene in C. elegans, lin-4, that was involved in silencing of another gene, lin-14, at the appropriate time in the
development of the worm C. elegans.
Two small transcripts of lin-4 (22nt and 61nt) were found to be complementary to a sequence in the 3' UTR of lin-14.
Because lin-4 encoded no protein, she deduced that it must be these transcripts that are causing the silencing by RNA-RNA interactions.
Types of RNAi ( non coding RNA)
MiRNA
Length (23-25 nt)
Trans acting
Binds with target MRNA in mismatch
Translation inhibition
Si RNA
Length 21 nt.
Cis acting
Bind with target Mrna in perfect complementary sequence
Piwi-RNA
Length ; 25 to 36 nt.
Expressed in Germ Cells
Regulates trnasposomes activity
MECHANISM OF RNAI:
First the double-stranded RNA teams up with a protein complex named Dicer, which cuts the long RNA into short pieces.
Then another protein complex called RISC (RNA-induced silencing complex) discards one of the two RNA strands.
The RISC-docked, single-stranded RNA then pairs with the homologous mRNA and destroys it.
THE RISC COMPLEX:
RISC is large(>500kD) RNA multi- protein Binding complex which triggers MRNA degradation in response to MRNA
Unwinding of double stranded Si RNA by ATP independent Helicase
Active component of RISC is Ago proteins( ENDONUCLEASE) which cleave target MRNA.
DICER: endonuclease (RNase Family III)
Argonaute: Central Component of the RNA-Induced Silencing Complex (RISC)
One strand of the dsRNA produced by Dicer is retained in the RISC complex in association with Argonaute
ARGONAUTE PROTEIN :
1.PAZ(PIWI/Argonaute/ Zwille)- Recognition of target MRNA
2.PIWI (p-element induced wimpy Testis)- breaks Phosphodiester bond of mRNA.)RNAse H activity.
MiRNA:
The Double-stranded RNAs are naturally produced in eukaryotic cells during development, and they have a key role in regulating gene expression .
Richard's entangled aventures in wonderlandRichard Gill
Since the loophole-free Bell experiments of 2020 and the Nobel prizes in physics of 2022, critics of Bell's work have retreated to the fortress of super-determinism. Now, super-determinism is a derogatory word - it just means "determinism". Palmer, Hance and Hossenfelder argue that quantum mechanics and determinism are not incompatible, using a sophisticated mathematical construction based on a subtle thinning of allowed states and measurements in quantum mechanics, such that what is left appears to make Bell's argument fail, without altering the empirical predictions of quantum mechanics. I think however that it is a smoke screen, and the slogan "lost in math" comes to my mind. I will discuss some other recent disproofs of Bell's theorem using the language of causality based on causal graphs. Causal thinking is also central to law and justice. I will mention surprising connections to my work on serial killer nurse cases, in particular the Dutch case of Lucia de Berk and the current UK case of Lucy Letby.
This pdf is about the Schizophrenia.
For more details visit on YouTube; @SELF-EXPLANATORY;
https://www.youtube.com/channel/UCAiarMZDNhe1A3Rnpr_WkzA/videos
Thanks...!
Multi-source connectivity as the driver of solar wind variability in the heli...Sérgio Sacani
The ambient solar wind that flls the heliosphere originates from multiple
sources in the solar corona and is highly structured. It is often described
as high-speed, relatively homogeneous, plasma streams from coronal
holes and slow-speed, highly variable, streams whose source regions are
under debate. A key goal of ESA/NASA’s Solar Orbiter mission is to identify
solar wind sources and understand what drives the complexity seen in the
heliosphere. By combining magnetic feld modelling and spectroscopic
techniques with high-resolution observations and measurements, we show
that the solar wind variability detected in situ by Solar Orbiter in March
2022 is driven by spatio-temporal changes in the magnetic connectivity to
multiple sources in the solar atmosphere. The magnetic feld footpoints
connected to the spacecraft moved from the boundaries of a coronal hole
to one active region (12961) and then across to another region (12957). This
is refected in the in situ measurements, which show the transition from fast
to highly Alfvénic then to slow solar wind that is disrupted by the arrival of
a coronal mass ejection. Our results describe solar wind variability at 0.5 au
but are applicable to near-Earth observatories.
THE IMPORTANCE OF MARTIAN ATMOSPHERE SAMPLE RETURN.Sérgio Sacani
The return of a sample of near-surface atmosphere from Mars would facilitate answers to several first-order science questions surrounding the formation and evolution of the planet. One of the important aspects of terrestrial planet formation in general is the role that primary atmospheres played in influencing the chemistry and structure of the planets and their antecedents. Studies of the martian atmosphere can be used to investigate the role of a primary atmosphere in its history. Atmosphere samples would also inform our understanding of the near-surface chemistry of the planet, and ultimately the prospects for life. High-precision isotopic analyses of constituent gases are needed to address these questions, requiring that the analyses are made on returned samples rather than in situ.
Slide 1: Title Slide
Extrachromosomal Inheritance
Slide 2: Introduction to Extrachromosomal Inheritance
Definition: Extrachromosomal inheritance refers to the transmission of genetic material that is not found within the nucleus.
Key Components: Involves genes located in mitochondria, chloroplasts, and plasmids.
Slide 3: Mitochondrial Inheritance
Mitochondria: Organelles responsible for energy production.
Mitochondrial DNA (mtDNA): Circular DNA molecule found in mitochondria.
Inheritance Pattern: Maternally inherited, meaning it is passed from mothers to all their offspring.
Diseases: Examples include Leber’s hereditary optic neuropathy (LHON) and mitochondrial myopathy.
Slide 4: Chloroplast Inheritance
Chloroplasts: Organelles responsible for photosynthesis in plants.
Chloroplast DNA (cpDNA): Circular DNA molecule found in chloroplasts.
Inheritance Pattern: Often maternally inherited in most plants, but can vary in some species.
Examples: Variegation in plants, where leaf color patterns are determined by chloroplast DNA.
Slide 5: Plasmid Inheritance
Plasmids: Small, circular DNA molecules found in bacteria and some eukaryotes.
Features: Can carry antibiotic resistance genes and can be transferred between cells through processes like conjugation.
Significance: Important in biotechnology for gene cloning and genetic engineering.
Slide 6: Mechanisms of Extrachromosomal Inheritance
Non-Mendelian Patterns: Do not follow Mendel’s laws of inheritance.
Cytoplasmic Segregation: During cell division, organelles like mitochondria and chloroplasts are randomly distributed to daughter cells.
Heteroplasmy: Presence of more than one type of organellar genome within a cell, leading to variation in expression.
Slide 7: Examples of Extrachromosomal Inheritance
Four O’clock Plant (Mirabilis jalapa): Shows variegated leaves due to different cpDNA in leaf cells.
Petite Mutants in Yeast: Result from mutations in mitochondrial DNA affecting respiration.
Slide 8: Importance of Extrachromosomal Inheritance
Evolution: Provides insight into the evolution of eukaryotic cells.
Medicine: Understanding mitochondrial inheritance helps in diagnosing and treating mitochondrial diseases.
Agriculture: Chloroplast inheritance can be used in plant breeding and genetic modification.
Slide 9: Recent Research and Advances
Gene Editing: Techniques like CRISPR-Cas9 are being used to edit mitochondrial and chloroplast DNA.
Therapies: Development of mitochondrial replacement therapy (MRT) for preventing mitochondrial diseases.
Slide 10: Conclusion
Summary: Extrachromosomal inheritance involves the transmission of genetic material outside the nucleus and plays a crucial role in genetics, medicine, and biotechnology.
Future Directions: Continued research and technological advancements hold promise for new treatments and applications.
Slide 11: Questions and Discussion
Invite Audience: Open the floor for any questions or further discussion on the topic.
Earliest Galaxies in the JADES Origins Field: Luminosity Function and Cosmic ...Sérgio Sacani
We characterize the earliest galaxy population in the JADES Origins Field (JOF), the deepest
imaging field observed with JWST. We make use of the ancillary Hubble optical images (5 filters
spanning 0.4−0.9µm) and novel JWST images with 14 filters spanning 0.8−5µm, including 7 mediumband filters, and reaching total exposure times of up to 46 hours per filter. We combine all our data
at > 2.3µm to construct an ultradeep image, reaching as deep as ≈ 31.4 AB mag in the stack and
30.3-31.0 AB mag (5σ, r = 0.1” circular aperture) in individual filters. We measure photometric
redshifts and use robust selection criteria to identify a sample of eight galaxy candidates at redshifts
z = 11.5 − 15. These objects show compact half-light radii of R1/2 ∼ 50 − 200pc, stellar masses of
M⋆ ∼ 107−108M⊙, and star-formation rates of SFR ∼ 0.1−1 M⊙ yr−1
. Our search finds no candidates
at 15 < z < 20, placing upper limits at these redshifts. We develop a forward modeling approach to
infer the properties of the evolving luminosity function without binning in redshift or luminosity that
marginalizes over the photometric redshift uncertainty of our candidate galaxies and incorporates the
impact of non-detections. We find a z = 12 luminosity function in good agreement with prior results,
and that the luminosity function normalization and UV luminosity density decline by a factor of ∼ 2.5
from z = 12 to z = 14. We discuss the possible implications of our results in the context of theoretical
models for evolution of the dark matter halo mass function.
Deep Behavioral Phenotyping in Systems Neuroscience for Functional Atlasing a...Ana Luísa Pinho
Functional Magnetic Resonance Imaging (fMRI) provides means to characterize brain activations in response to behavior. However, cognitive neuroscience has been limited to group-level effects referring to the performance of specific tasks. To obtain the functional profile of elementary cognitive mechanisms, the combination of brain responses to many tasks is required. Yet, to date, both structural atlases and parcellation-based activations do not fully account for cognitive function and still present several limitations. Further, they do not adapt overall to individual characteristics. In this talk, I will give an account of deep-behavioral phenotyping strategies, namely data-driven methods in large task-fMRI datasets, to optimize functional brain-data collection and improve inference of effects-of-interest related to mental processes. Key to this approach is the employment of fast multi-functional paradigms rich on features that can be well parametrized and, consequently, facilitate the creation of psycho-physiological constructs to be modelled with imaging data. Particular emphasis will be given to music stimuli when studying high-order cognitive mechanisms, due to their ecological nature and quality to enable complex behavior compounded by discrete entities. I will also discuss how deep-behavioral phenotyping and individualized models applied to neuroimaging data can better account for the subject-specific organization of domain-general cognitive systems in the human brain. Finally, the accumulation of functional brain signatures brings the possibility to clarify relationships among tasks and create a univocal link between brain systems and mental functions through: (1) the development of ontologies proposing an organization of cognitive processes; and (2) brain-network taxonomies describing functional specialization. To this end, tools to improve commensurability in cognitive science are necessary, such as public repositories, ontology-based platforms and automated meta-analysis tools. I will thus discuss some brain-atlasing resources currently under development, and their applicability in cognitive as well as clinical neuroscience.
Observation of Io’s Resurfacing via Plume Deposition Using Ground-based Adapt...Sérgio Sacani
Since volcanic activity was first discovered on Io from Voyager images in 1979, changes
on Io’s surface have been monitored from both spacecraft and ground-based telescopes.
Here, we present the highest spatial resolution images of Io ever obtained from a groundbased telescope. These images, acquired by the SHARK-VIS instrument on the Large
Binocular Telescope, show evidence of a major resurfacing event on Io’s trailing hemisphere. When compared to the most recent spacecraft images, the SHARK-VIS images
show that a plume deposit from a powerful eruption at Pillan Patera has covered part
of the long-lived Pele plume deposit. Although this type of resurfacing event may be common on Io, few have been detected due to the rarity of spacecraft visits and the previously low spatial resolution available from Earth-based telescopes. The SHARK-VIS instrument ushers in a new era of high resolution imaging of Io’s surface using adaptive
optics at visible wavelengths.
Comparing Evolved Extractive Text Summary Scores of Bidirectional Encoder Rep...University of Maribor
Slides from:
11th International Conference on Electrical, Electronics and Computer Engineering (IcETRAN), Niš, 3-6 June 2024
Track: Artificial Intelligence
https://www.etran.rs/2024/en/home-english/
1. Michael Biehl
Bernoulli Institute for Mathematics,
Computer Science and Artificial Intelligence
University of Groningen
The Netherlands
www.cs.rug.nl/~biehl
Prototype-based machine learning:
bio-medical applications
7. AMALEA 2022 4
overview
1. Introduction / Motivation
prototypes and exemplars, neural activation / learning
3. Supervised Learning
Learning Vector Quantization (LVQ)
Adaptive distances and Relevance Learning
2. Unsupervised Learning
Competitive Learning
Kohonen’s Self-Organizing Map (SOM)
Examples and illustrations : Bio-medical applications
- clustering of proteomics data
- biomarkers for rheumatoid arthritis
(- FDG-Pet brain scans)
8. AMALEA 2022 5
1. Introduction
prototypes, exemplars:
representation of information in terms of
typical representatives (e.g. of a class of objects),
much debated concept in cognitive psychology
9. AMALEA 2022 5
1. Introduction
prototypes, exemplars:
representation of information in terms of
typical representatives (e.g. of a class of objects),
much debated concept in cognitive psychology
10. AMALEA 2022 5
1. Introduction
prototypes, exemplars:
representation of information in terms of
typical representatives (e.g. of a class of objects),
much debated concept in cognitive psychology
11. AMALEA 2022 5
1. Introduction
prototypes, exemplars:
representation of information in terms of
typical representatives (e.g. of a class of objects),
much debated concept in cognitive psychology
neural activation:
external stimulus to a network of neurons
response acc. to weights (expected inputs)
12. AMALEA 2022 5
1. Introduction
prototypes, exemplars:
representation of information in terms of
typical representatives (e.g. of a class of objects),
much debated concept in cognitive psychology
neural activation:
external stimulus to a network of neurons
response acc. to weights (expected inputs)
best matching unit (and neighbors)
weights represent different expected stimuli (prototypes)
13. AMALEA 2022 5
1. Introduction
prototypes, exemplars:
representation of information in terms of
typical representatives (e.g. of a class of objects),
much debated concept in cognitive psychology
neural activation:
external stimulus to a network of neurons
response acc. to weights (expected inputs)
best matching unit (and neighbors)
weights represent different expected stimuli (prototypes)
learning: change of weights result in even stronger
response to similar stimuli in the future
14. AMALEA 2022 6
even independent from the above:
attractive framework for machine learning based data analysis
- trained system is parameterized in the feature space
- facilitates discussions with domain experts
- transparent (white box) and provides insights into the
applied criteria (classification, regression, clustering etc.)
- easy to implement, efficient computation
- versatile, successfully applied in many different application areas
15. AMALEA 2022 7
2. Unsupervised Learning
Potential aims:
dimension reduction: compression, visualization, ...
exploration of data structure: clustering, density estimation, ...
pre-processing: supvervised learning, classification, regression,...
16. AMALEA 2022 7
2. Unsupervised Learning
Potential aims:
dimension reduction: compression, visualization, ...
exploration of data structure: clustering, density estimation, ...
pre-processing: supvervised learning, classification, regression,...
Vector Quantization: identify (few) typical representatives
from a set of feature vectors
w1
, w2
, . . . , wK
wk
2 I
RN
x1
, x2
, . . . , xP
xµ
2 I
RN
17. AMALEA 2022 7
2. Unsupervised Learning
Potential aims:
dimension reduction: compression, visualization, ...
exploration of data structure: clustering, density estimation, ...
pre-processing: supvervised learning, classification, regression,...
Vector Quantization: identify (few) typical representatives
from a set of feature vectors
w1
, w2
, . . . , wK
wk
2 I
RN
x1
, x2
, . . . , xP
xµ
2 I
RN
assign xµ to winning prototype w⇤
= argminj d(wj
, xµ
)
d(w, x) =
N
X
n=1
(wn xn)
2
for instance w.r.t. :
squared Euclidean distance
18. AMALEA 2022 8
, random sequence of single data:
… the winner takes it all:
initially: randomized wk
competitive learning
competition for updates
learning rate / step size η <1
⌘ (xµ
w⇤
)
w⇤
! w⇤
+ ⌘ (xµ
w⇤
)
19. AMALEA 2022 8
, random sequence of single data:
… the winner takes it all:
initially: randomized wk
competitive learning
competition for updates
learning rate / step size η <1
⌘ (xµ
w⇤
)
w⇤
! w⇤
+ ⌘ (xµ
w⇤
)
20. AMALEA 2022 8
, random sequence of single data:
… the winner takes it all:
initially: randomized wk
competitive learning
competition for updates
learning rate / step size η <1
⌘ (xµ
w⇤
)
w⇤
! w⇤
+ ⌘ (xµ
w⇤
)
competitive VQ = stochastic gradient descent w.r.t. Quantization Error
- assign each data to closest prototype
- measure the corresponding distance (e.g. squared Euclidean)
- sum over all assigned data points
measures the quality of the representation
defines a (one possible) criterion to evaluate / compare
the quality of different prototype configurations
{
QE
24. AMALEA 2022 9
data
initial
prototypes
dead
units
WTA training
general problem: local minima of the quantization error,
initialization-dependent outcome of training
competitive learning
improvement: rank-based updates (winner, second, third,… )
25. AMALEA 2022 9
data
initial
prototypes
dead
units
WTA training
general problem: local minima of the quantization error,
initialization-dependent outcome of training
competitive learning
improvement: rank-based updates (winner, second, third,… )
introduce rank-based neighborhood cooperativeness
[Martinetz, Berkovich, Schulten, IEEE Trans. Neural Netw. 1993]
Neural Gas: many prototypes to represent the density of data
26. AMALEA 2022
Self-Organizing Map
T. Kohonen. Self-Organizing Maps. Springer (1995)
neighborhood cooperativeness on a predefined low-dim. lattice
27. AMALEA 2022
Self-Organizing Map
T. Kohonen. Self-Organizing Maps. Springer (1995)
neighborhood cooperativeness on a predefined low-dim. lattice
lattice A of neurons
i.e. prototypes
wr 2 I
RN
at r 2 I
Rd
A
28. AMALEA 2022
Self-Organizing Map
T. Kohonen. Self-Organizing Maps. Springer (1995)
neighborhood cooperativeness on a predefined low-dim. lattice
lattice A of neurons
i.e. prototypes
ws
wr 2 I
RN
at r 2 I
Rd
upon presentation of xµ :
- determine the winner (best matching unit)
in feature space: (at position s in A)
A
29. AMALEA 2022
Self-Organizing Map
T. Kohonen. Self-Organizing Maps. Springer (1995)
neighborhood cooperativeness on a predefined low-dim. lattice
lattice A of neurons
i.e. prototypes
- update winner and lattice neighborhood:
where
range ρ w.r.t. distances in lattice A
ws
wr 2 I
RN
at r 2 I
Rd
h⇢(r, s) = exp
✓
|| r s ||2
A
2⇢2
◆
upon presentation of xµ :
- determine the winner (best matching unit)
in feature space: (at position s in A)
<latexit sha1_base64="P/AKHBol8ZLEa4RgJDSz5KstNjs=">AAACY3icbVHLSgMxFM2MWrU+Oj52KgRFaLGWGVF0KbhxWcFaoVOHTJrphGYeJBm1DrP0i9z6Fe7EjRv/w0wfUqsXkns491zuzYkbMyqkab5r+szsXGF+YbG4tLyyWjLW1m9ElHBMGjhiEb91kSCMhqQhqWTkNuYEBS4jTbd3kdeb94QLGoXXsh+TdoC6IfUoRlJRjvFkB0j6rpc+ZA6HtozgiIAjpgoP8ssmEuXZd2zuR+VxF8+qYyiyihLkUpd2fwSP2Z0dJPAQTs7JFRXoGHtmzRwE/AusEdg73/aeP+5fXuuO8WZ3IpwEJJSYISFalhnLdoq4pJiRrGgngsQI91CXtBQMUUBEOx14lMF9xXSgF3F1QgkH7GRHigIh+oGrlPmmYrqWk//VWon0ztopDeNEkhAPB3kJg8rK3HDYoZxgyfoKIMyp2hViH3GEpfqWojLBmn7yX3BzVLNOauaVcuMYDGMBbIFdUAYWOAXn4BLUQQNg8KkVtJJmaF/6kr6ubw6lujbq2QC/Qt/5BtYMuc8=</latexit>
wr ! wr + ⌘ h⇢(r, s) xµ
wr
A
33. AMALEA 2022 12
- many extensions of the basic concept, e.g.
cost function based SOM [Heskes]
Generative Topographic Map (GTM), probabilistic
formulation of the mapping to low-dim. lattice
[Bishop, Svensen, Williams, 1998]
specific modifications of SOM or Neural Gas for
- time series / functional data
- “non-vectorial” relational data
- graphs and trees
- supervised learning
Remarks
39. AMALEA 2022
transcription:
DNA è (m)RNA
translation
mRNA è
proteins
the “central dogma” of molecular biology
Ribosome
protein,è
function
40. AMALEA 2022
• is an ancient molecular machine, ‘3D-printer’ for proteins
the ribosome…
41. AMALEA 2022
• is an ancient molecular machine, ‘3D-printer’ for proteins
• ~ 107 cytoplasmic ribosomes per cell
the ribosome…
42. AMALEA 2022
• is an ancient molecular machine, ‘3D-printer’ for proteins
• ~ 107 cytoplasmic ribosomes per cell
• is believed to have universal function and the same
the ribosome…
43. AMALEA 2022
• is an ancient molecular machine, ‘3D-printer’ for proteins
• ~ 107 cytoplasmic ribosomes per cell
• is believed to have universal function and the same
composition in different tissues and across species
the ribosome…
44. AMALEA 2022
• is an ancient molecular machine, ‘3D-printer’ for proteins
• ~ 107 cytoplasmic ribosomes per cell
• is believed to have universal function and the same
composition in different tissues and across species
• consists of RNA and
the ribosome…
45. AMALEA 2022
• is an ancient molecular machine, ‘3D-printer’ for proteins
• ~ 107 cytoplasmic ribosomes per cell
• is believed to have universal function and the same
composition in different tissues and across species
• consists of RNA and
ribosomal proteins (RP)
the ribosome…
46. AMALEA 2022
• is an ancient molecular machine, ‘3D-printer’ for proteins
• ~ 107 cytoplasmic ribosomes per cell
• is believed to have universal function and the same
composition in different tissues and across species
• consists of RNA and
ribosomal proteins (RP)
the ribosome…
also coded by DNA which
is transcribed to mRNA
47. AMALEA 2022
• is an ancient molecular machine, ‘3D-printer’ for proteins
• ~ 107 cytoplasmic ribosomes per cell
• is believed to have universal function and the same
composition in different tissues and across species
• consists of RNA and
ribosomal proteins (RP)
the ribosome…
also coded by DNA which
is transcribed to mRNA
here: analysis of
RP mRNA expression
48. AMALEA 2022
Klijn et al. Nature Biotechnol. 33(3): 306-312 (2015)
675 cell lines
public domain data sets
GTeX (v6p) www.gtexportal.org
8,555 normal samples from 53 different tissues (with >50 samples)
TCGA (NCI-GDC, v7) www.cancer.gov
10363 tumor samples, 730 tumor-adjacent normals
49. AMALEA 2022
Klijn et al. Nature Biotechnol. 33(3): 306-312 (2015)
675 cell lines
public domain data sets
GTeX (v6p) www.gtexportal.org
8,555 normal samples from 53 different tissues (with >50 samples)
TCGA (NCI-GDC, v7) www.cancer.gov
10363 tumor samples, 730 tumor-adjacent normals
mRNA
expression
50. AMALEA 2022
Klijn et al. Nature Biotechnol. 33(3): 306-312 (2015)
675 cell lines
public domain data sets
GTeX (v6p) www.gtexportal.org
8,555 normal samples from 53 different tissues (with >50 samples)
TCGA (NCI-GDC, v7) www.cancer.gov
10363 tumor samples, 730 tumor-adjacent normals
normalization:
constant sum of reads
for each of the 78 RP
depending on method:
log-transform, z-score
mRNA
expression
51. AMALEA 2022
normal samples (GTeX)
whole blood - brain tissues - rest
different tissues have different RP mRNA signatures: PCA
57. AMALEA 2022
RP mRNA signatures vary with
Ÿ tissue type Ÿ tumor type and sub-type Ÿ developmental stage
RP translation rates are proportional to RP mRNA levels
RP mRNA and profiling different in cell cultures vs. cells in-vivo
main findings: (only very few presented here, see NAR paper)
summary/conclusion
58. AMALEA 2022
RP mRNA signatures vary with
Ÿ tissue type Ÿ tumor type and sub-type Ÿ developmental stage
RP translation rates are proportional to RP mRNA levels
RP mRNA and profiling different in cell cultures vs. cells in-vivo
main findings: (only very few presented here, see NAR paper)
speculative (yet plausible) conclusions:
RP composition and function (?)
Ÿ is tissue-, tumor-, development-, environment-specific
Ÿ adds a novel layer to the regulatory network of the cell
Ÿ might play an important role in cancer
summary/conclusion
59. AMALEA 2022
RP mRNA signatures vary with
Ÿ tissue type Ÿ tumor type and sub-type Ÿ developmental stage
RP translation rates are proportional to RP mRNA levels
RP mRNA and profiling different in cell cultures vs. cells in-vivo
main findings: (only very few presented here, see NAR paper)
speculative (yet plausible) conclusions:
RP composition and function (?)
Ÿ is tissue-, tumor-, development-, environment-specific
Ÿ adds a novel layer to the regulatory network of the cell
Ÿ might play an important role in cancer
caveats: composition could be independent of RP abundance
possible extra-ribosomal functions of RP
direct inspection of ribosome is difficult
summary/conclusion
62. AMALEA 2022
supervised learning
classification / regression / prediction
based on labeled example data
generic workflow:
example data model apply to novel data
training working
validation
estimate working performance
set parameters of model / training
compare different models
63. AMALEA 2022
supervised learning
classification / regression / prediction
based on labeled example data
generic workflow:
example data model apply to novel data
training working
obvious performance measures: overall / class-wise accuracy
ROC, Precision Recall ...
validation
estimate working performance
set parameters of model / training
compare different models
64. AMALEA 2022
21
supervised learning
classification / regression / prediction
based on labeled example data
generic workflow:
example data model apply to novel data
training working
obvious performance measures: overall / class-wise accuracy
ROC, Precision Recall ...
validation
estimate working performance
set parameters of model / training
compare different models
accuracy is not enough - interpretable “white-box” systems
example: prototype-based models, distance-based classifiers
66. AMALEA 2022
distance-based classifiers
a simple distance-based system: NN classifier
store a set of labeled examples
classify a query according to the
label of the Nearest Neighbor
in the data set
?
N-dim. feature space
67. AMALEA 2022
distance-based classifiers
a simple distance-based system: NN classifier
store a set of labeled examples
classify a query according to the
label of the Nearest Neighbor
in the data set
piece-wise linear decision
boundaries according to e.g.
(squared) Euclidean distance:
?
N-dim. feature space
d(xµ
, x⌫
) =
N
X
j=1
xµ
j x⌫
j
2
68. AMALEA 2022
distance-based classifiers
a simple distance-based system: NN classifier
store a set of labeled examples
classify a query according to the
label of the Nearest Neighbor
in the data set
piece-wise linear decision
boundaries according to e.g.
(squared) Euclidean distance:
?
N-dim. feature space
d(xµ
, x⌫
) =
N
X
j=1
xµ
j x⌫
j
2
69. AMALEA 2022
distance-based classifiers
a simple distance-based system: NN classifier
store a set of labeled examples
classify a query according to the
label of the Nearest Neighbor
in the data set
piece-wise linear decision
boundaries according to e.g.
(squared) Euclidean distance:
?
N-dim. feature space
+ conceptually simple,
+ no training phase
- expensive (storage, computation)
- sensitive to mislabeled data
- overly complex decision boundaries
d(xµ
, x⌫
) =
N
X
j=1
xµ
j x⌫
j
2
70. AMALEA 2022
prototype based classification
a prototype based classifier [Kohonen 1990]
represent the data by one or
several prototypes per class
N-dim. feature space
71. AMALEA 2022
prototype based classification
a prototype based classifier [Kohonen 1990]
represent the data by one or
several prototypes per class
classify a query according to the
label of the nearest prototype
(or alternative schemes)
local decision boundaries
acc. to Euclidean distances
from the prototypes
piece-wise linear class borders
parameterized by prototypes
N-dim. feature space
72. AMALEA 2022
prototype based classification
a prototype based classifier [Kohonen 1990]
represent the data by one or
several prototypes per class
classify a query according to the
label of the nearest prototype
(or alternative schemes)
local decision boundaries
acc. to Euclidean distances
from the prototypes
piece-wise linear class borders
parameterized by prototypes
N-dim. feature space
+
less sensitive to outliers, lower storage needs,
little computational effort in the working phase
73. AMALEA 2022
prototype based classification
a prototype based classifier [Kohonen 1990]
represent the data by one or
several prototypes per class
classify a query according to the
label of the nearest prototype
(or alternative schemes)
local decision boundaries
acc. to Euclidean distances
from the prototypes
piece-wise linear class borders
parameterized by prototypes
N-dim. feature space
+
less sensitive to outliers, lower storage needs,
little computational effort in the working phase
-
training phase required in order to place prototypes,
model selection problem: number of prototypes per class etc.
74. AMALEA 2022
∙ identification of prototype vectors from labeled example data
∙ distance based classification (e.g. Euclidean)
Learning Vector Quantization
N-dimensional data, feature vectors
75. AMALEA 2022
∙ identification of prototype vectors from labeled example data
∙ distance based classification (e.g. Euclidean)
Learning Vector Quantization
N-dimensional data, feature vectors
competitive learning: heuristic LVQ1 [Kohonen, 1990]
76. AMALEA 2022
∙ identification of prototype vectors from labeled example data
∙ distance based classification (e.g. Euclidean)
Learning Vector Quantization
N-dimensional data, feature vectors
• initialize prototype vectors for each class
competitive learning: heuristic LVQ1 [Kohonen, 1990]
77. AMALEA 2022
∙ identification of prototype vectors from labeled example data
∙ distance based classification (e.g. Euclidean)
Learning Vector Quantization
N-dimensional data, feature vectors
• initialize prototype vectors for each class
competitive learning: heuristic LVQ1 [Kohonen, 1990]
• present a single example
78. AMALEA 2022
∙ identification of prototype vectors from labeled example data
∙ distance based classification (e.g. Euclidean)
Learning Vector Quantization
N-dimensional data, feature vectors
• initialize prototype vectors for each class
competitive learning: heuristic LVQ1 [Kohonen, 1990]
• identify the winner (closest prototype)
• present a single example
79. AMALEA 2022
∙ identification of prototype vectors from labeled example data
∙ distance based classification (e.g. Euclidean)
Learning Vector Quantization
N-dimensional data, feature vectors
• initialize prototype vectors for each class
competitive learning: heuristic LVQ1 [Kohonen, 1990]
• identify the winner (closest prototype)
• present a single example
• move the winner
- closer towards the data (same class)
80. AMALEA 2022
∙ identification of prototype vectors from labeled example data
∙ distance based classification (e.g. Euclidean)
Learning Vector Quantization
N-dimensional data, feature vectors
• initialize prototype vectors for each class
competitive learning: heuristic LVQ1 [Kohonen, 1990]
• identify the winner (closest prototype)
• present a single example
• move the winner
- closer towards the data (same class)
81. AMALEA 2022
∙ identification of prototype vectors from labeled example data
∙ distance based classification (e.g. Euclidean)
Learning Vector Quantization
N-dimensional data, feature vectors
• initialize prototype vectors for each class
competitive learning: heuristic LVQ1 [Kohonen, 1990]
• identify the winner (closest prototype)
• present a single example
• move the winner
- closer towards the data (same class)
82. AMALEA 2022
∙ identification of prototype vectors from labeled example data
∙ distance based classification (e.g. Euclidean)
Learning Vector Quantization
N-dimensional data, feature vectors
• initialize prototype vectors for each class
competitive learning: heuristic LVQ1 [Kohonen, 1990]
• identify the winner (closest prototype)
• present a single example
• move the winner
- closer towards the data (same class)
- away from the data (different class)
83. AMALEA 2022
∙ identification of prototype vectors from labeled example data
∙ distance based classification (e.g. Euclidean)
Learning Vector Quantization
N-dimensional data, feature vectors
• initialize prototype vectors for each class
competitive learning: heuristic LVQ1 [Kohonen, 1990]
• identify the winner (closest prototype)
• present a single example
• move the winner
- closer towards the data (same class)
- away from the data (different class)
• many variants, including
cost-function-based schemes, e.g.
Generalized LVQ (approximates # of misclassifications)
84. AMALEA 2022
∙ identification of prototype vectors from labeled example data
∙ distance based classification (e.g. squared Euclidean)
Learning Vector Quantization
N-dimensional data, feature vectors
∙ tesselation of feature space
[piece-wise linear]
∙ distance-based classification
[here: Euclidean distances]
∙ aim: generalization ability
correct classification of new data
85. AMALEA 2022
LVQ distance measures
? key question: appropriate distance / (dis-) similarity measure
fixed, pre-defined distance measures:
(G)LVQ can formulated for general (differentiable) distances
86. AMALEA 2022
LVQ distance measures
? key question: appropriate distance / (dis-) similarity measure
fixed, pre-defined distance measures:
(G)LVQ can formulated for general (differentiable) distances
examples: Minkowski distances (p≠2), correlation based,
statistical divergences, ... not necessarily metrics!
87. AMALEA 2022
LVQ distance measures
? key question: appropriate distance / (dis-) similarity measure
fixed, pre-defined distance measures:
(G)LVQ can formulated for general (differentiable) distances
examples: Minkowski distances (p≠2), correlation based,
statistical divergences, ... not necessarily metrics!
standard work-flow
- consider several distance measures
- compare performances in, e.g., cross-validation
88. AMALEA 2022
LVQ distance measures
? key question: appropriate distance / (dis-) similarity measure
fixed, pre-defined distance measures:
(G)LVQ can formulated for general (differentiable) distances
examples: Minkowski distances (p≠2), correlation based,
statistical divergences, ... not necessarily metrics!
standard work-flow
- consider several distance measures
- compare performances in, e.g., cross-validation
elegant approach:
Relevance Learning / adaptive distances
- employ parameterized distance measure
- optimize in the data-driven training process (cost function!)
89. AMALEA 2022
Generalized Matrix Relevance LVQ
generalized quadratic distance in LVQ: [Schneider, Biehl, Hammer, 2009]
d(w, x) = (w x)
>
⇤ (w x)
(GMLVQ)
90. AMALEA 2022
Generalized Matrix Relevance LVQ
generalized quadratic distance in LVQ: [Schneider, Biehl, Hammer, 2009]
d(w, x) = (w x)
>
⇤ (w x)
(GMLVQ)
91. AMALEA 2022
Generalized Matrix Relevance LVQ
generalized quadratic distance in LVQ: [Schneider, Biehl, Hammer, 2009]
d(w, x) = (w x)
>
⇤ (w x)
(GMLVQ)
= [ ⌦ (w x) ]
2
92. AMALEA 2022
GMLVQ
generalized quadratic distance in LVQ: [Schneider, Biehl, Hammer, 2009]
training: adaptation of prototypes
and distance measure guided by
GLVQ cost function
= [ ⌦ (w x) ]
2
d(w, x) = (w x)
>
⇤ (w x)
Generalized Matrix Relevance LVQ:
93. AMALEA 2022
GMLVQ
generalized quadratic distance in LVQ: [Schneider, Biehl, Hammer, 2009]
variants:
one global, several local, class-wise relevance matrices
rectangular low-dim. representation / visualization
[Bunte et al., 2012]
diagonal matrices: single feature weights [Hammer et al., 2002]
training: adaptation of prototypes
and distance measure guided by
GLVQ cost function
= [ ⌦ (w x) ]
2
d(w, x) = (w x)
>
⇤ (w x)
Generalized Matrix Relevance LVQ:
95. AMALEA 2022
But this is just Mahalonobis distance…
[Mahalonobis, 1936]
S covariance matrix of random vectors
(calculated once from the data, fixed definition, not adaptive)
x 2 RN
(‘two point version’)
No.
dM (x, y) =
q
(x y)> S 1 (x y)
96. AMALEA 2022
But this is just Mahalonobis distance…
[Mahalonobis, 1936]
S covariance matrix of random vectors
(calculated once from the data, fixed definition, not adaptive)
x 2 RN
if you insist…
(‘two point version’)
So it is a generalized Mahalonobis distance ?
No.
dM (x, y) =
q
(x y)> S 1 (x y)
97. AMALEA 2022
But this is just Mahalonobis distance…
[Mahalonobis, 1936]
S covariance matrix of random vectors
(calculated once from the data, fixed definition, not adaptive)
x 2 RN
if you insist…
(‘two point version’)
So it is a generalized Mahalonobis distance ?
No.
a generalized
broccoli
E = ~!
a generalization
of Ohm’s Law
dM (x, y) =
q
(x y)> S 1 (x y)
99. AMALEA 2022 99
interpretation
summarizes
• the contribution of a single dimension
• the relevance of original features in the classifier
⇤ij
quantifies the contribution of the pair
of features (i,j) to the distance
after training:
prototypes represent typical class properties or subtypes (hope)
Relevance Matrix
100. AMALEA 2022 100
interpretation
summarizes
• the contribution of a single dimension
• the relevance of original features in the classifier
Note: interpretation assumes implicitly that
features have equal order of magnitude
e.g. after z-score-transformation →
(averages over data set)
⇤ij
quantifies the contribution of the pair
of features (i,j) to the distance
after training:
prototypes represent typical class properties or subtypes (hope)
Relevance Matrix
101. Urine Steroid Metabolomics as a Biomarker Tool for
Detecting Malignancy in Patients with Adrenal Tumors
www.ensat.org
W. Arlt, M. Biehl, A. Taylor, S. Hahner, R. Libé, B. Hughes, P. Schneider,
D. Smith, H. Stiekema, N. Krone, E. Porfiri, G. Opocher, J. Bertherat,
F. Mantero, B. Allolio, M. Terzolo, P. Nightingale, C. Shackleton,
X. Bertagna, M.Fassnacht, P. Stewart
J Clinical Endocrinology & Metabolism 96: 3775-3784 (2011)
tumor classification
102. Urine Steroid Metabolomics as a Biomarker Tool for
Detecting Malignancy in Patients with Adrenal Tumors
www.ensat.org
W. Arlt, M. Biehl, A. Taylor, S. Hahner, R. Libé, B. Hughes, P. Schneider,
D. Smith, H. Stiekema, N. Krone, E. Porfiri, G. Opocher, J. Bertherat,
F. Mantero, B. Allolio, M. Terzolo, P. Nightingale, C. Shackleton,
X. Bertagna, M.Fassnacht, P. Stewart
J Clinical Endocrinology & Metabolism 96: 3775-3784 (2011)
tumor classification
insight: marker selection, patented diagnosis tool
follow-up: recurrence detection, other disorders, tumor sub-types...
104. AMALEA 2022 104
two recent application examples
I) cytokine expression data:
- insights into disease mechanisms of (early) rheumatoid arthritis
based on synovial tissue samples
~ 50 samples represented by 117 cytokine expressions
in synovial tissue, PCA+GMLVQ combined
105. AMALEA 2022 105
two recent application examples
I) cytokine expression data:
- insights into disease mechanisms of (early) rheumatoid arthritis
based on synovial tissue samples
~ 50 samples represented by 117 cytokine expressions
in synovial tissue, PCA+GMLVQ combined
II) FDG-PET brain scans:
- ultimate goal: diagnosis of neurodegenerative diseases
~ 100 samples, ~200000 voxels per scan
SSM/PCA+GMLVQ combined
106. Early diagnosis (?) of Rheumatoid Arthritis
Expression of chemokines CXCL4 and CXCL7 by synovial
macrophages defines an early stage of rheumatoid arthritis
Annals of the Rheumatic Diseases 75:763-771 (2016)
L. Yeo, N. Adlard, M. Biehl, M. Juarez, M. Snow
C.D. Buckley, A. Filer, K. Raza, D. Scheel-Toellner
107. AMALEA 2022 34
Rheumatoid Arthritis
Rheumatoid Arthritis (RA)
- chronicle inflammatory disease
- immune system affects joints
- RA leads to deformation and disability
110. AMALEA 2022
uninflamed control established RA early inflammation
resolving early RA
ultimate goals:
understand pathogenesis and
mechanism of progression
rheumatoid arthritis (RA)
111. AMALEA 2022
uninflamed control established RA early inflammation
resolving early RA
cytokine based diagnosis of RA
at earliest possible stage ?
ultimate goals:
understand pathogenesis and
mechanism of progression
rheumatoid arthritis (RA)
115. AMALEA 2022
GMLVQ analysis
pre-processing
• log-transformed expression values
• 21 leading principal components explain 95% of the variation
Two binary problems: (A) established RA vs. uninflamed controls (!)
(B) early RA vs. resolving inflammation (")
• 1 prototype per class, global relevance matrix, distance measure:
x 2 I
R117
, x = e
x 2 I
R21
d(e
x, e
w) = (e
x e
w)
> e
⇤ (e
x e
w) = (x w)
> > e
⇤
| {z }
⇤
(x w)
116. AMALEA 2022
GMLVQ analysis
pre-processing
• log-transformed expression values
• 21 leading principal components explain 95% of the variation
Two binary problems: (A) established RA vs. uninflamed controls (!)
(B) early RA vs. resolving inflammation (")
• 1 prototype per class, global relevance matrix, distance measure:
x 2 I
R117
, x = e
x 2 I
R21
d(e
x, e
w) = (e
x e
w)
> e
⇤ (e
x e
w) = (x w)
> > e
⇤
| {z }
⇤
(x w)
d(e
x, e
w) = (e
x e
w)
> e
⇤ (e
x e
w) = (x w)
> > e
⇤
| {z }
⇤
(x w)
in
<latexit sha1_base64="S6bznRCRGpJCx80D30CfTaQNebg=">AAAB83icbVDLSsNAFL2pr1pfUZduBovgqiSi6M6CG1dSxT6giWUynbRDJ5MwMxFK6G8I6kIRt36Av+HOv3HSdqGtBwYO59zLPXOChDOlHefbKiwsLi2vFFdLa+sbm1v29k5DxakktE5iHstWgBXlTNC6ZprTViIpjgJOm8HgIveb91QqFotbPUyoH+GeYCEjWBvJ8yKs+0GQ3Yzurjp22ak4Y6B54k5J+fzzMcdTrWN/ed2YpBEVmnCsVNt1Eu1nWGpGOB2VvFTRBJMB7tG2oQJHVPnZOPMIHRili8JYmic0Gqu/NzIcKTWMAjOZZ1SzXi7+57VTHZ75GRNJqqkgk0NhypGOUV4A6jJJieZDQzCRzGRFpI8lJtrUVDIluLNfnieNo4p7UnGunXL1GCYowh7swyG4cApVuIQa1IFAAg/wAq9Waj1bb9b7ZLRgTXd24Q+sjx8yWJYz</latexit>
RN
117. AMALEA 2022
GMLVQ analysis
pre-processing
• log-transformed expression values
• 21 leading principal components explain 95% of the variation
Two binary problems: (A) established RA vs. uninflamed controls (!)
(B) early RA vs. resolving inflammation (")
• 1 prototype per class, global relevance matrix, distance measure:
• leave-two-out validation (one from each class)
evaluation in terms of Receiver Operating Characteristics
x 2 I
R117
, x = e
x 2 I
R21
d(e
x, e
w) = (e
x e
w)
> e
⇤ (e
x e
w) = (x w)
> > e
⇤
| {z }
⇤
(x w)
d(e
x, e
w) = (e
x e
w)
> e
⇤ (e
x e
w) = (x w)
> > e
⇤
| {z }
⇤
(x w)
in
<latexit sha1_base64="S6bznRCRGpJCx80D30CfTaQNebg=">AAAB83icbVDLSsNAFL2pr1pfUZduBovgqiSi6M6CG1dSxT6giWUynbRDJ5MwMxFK6G8I6kIRt36Av+HOv3HSdqGtBwYO59zLPXOChDOlHefbKiwsLi2vFFdLa+sbm1v29k5DxakktE5iHstWgBXlTNC6ZprTViIpjgJOm8HgIveb91QqFotbPUyoH+GeYCEjWBvJ8yKs+0GQ3Yzurjp22ak4Y6B54k5J+fzzMcdTrWN/ed2YpBEVmnCsVNt1Eu1nWGpGOB2VvFTRBJMB7tG2oQJHVPnZOPMIHRili8JYmic0Gqu/NzIcKTWMAjOZZ1SzXi7+57VTHZ75GRNJqqkgk0NhypGOUV4A6jJJieZDQzCRzGRFpI8lJtrUVDIluLNfnieNo4p7UnGunXL1GCYowh7swyG4cApVuIQa1IFAAg/wAq9Waj1bb9b7ZLRgTXd24Q+sjx8yWJYz</latexit>
RN
118. AMALEA 2022
false positive rate
true
positive
rate
diagonal Λii vs. cytokine index i
(A) established RA vs.
uninflamed control
Relevances
diagonal relevances
leave-one-out
119. AMALEA 2022
false positive rate
true
positive
rate
t
rue
positive
rate
diagonal Λii vs. cytokine index i
(A) established RA vs.
uninflamed control
(B) early RA vs.
resolving inflammation
Relevances
diagonal relevances
leave-one-out
120. AMALEA 2022
false positive rate
true
positive
rate
t
rue
positive
rate
diagonal Λii vs. cytokine index i
(A) established RA vs.
uninflamed control
(B) early RA vs.
resolving inflammation
Relevances
diagonal relevances
leave-one-out
122. AMALEA 2022
CXCL4 chemokine (C-X-C motif) ligand 4
CXCL7 chemokine (C-X-C motif) ligand 7
direct study on protein level, staining / imaging of sinovial tissue:
macrophages : predominant source of CXCL4/7 expression
protein level studies
123. AMALEA 2022
CXCL4 chemokine (C-X-C motif) ligand 4
CXCL7 chemokine (C-X-C motif) ligand 7
direct study on protein level, staining / imaging of sinovial tissue:
macrophages : predominant source of CXCL4/7 expression
protein level studies
• high levels of CXCL4 and
CXLC7 in early RA
• expression on macrophages
outside of blood vessels
discriminates
early RA / resolving cases
124. AMALEA 2022
false positive rate
true
positive
rate
t
rue
positive
rate
diagonal Λii vs. cytokine index i
(A) established RA vs.
uninflamed control
(B) early RA vs.
resolving inflammation
relevant cytokines
macrophage
stimulating 1
diagonal relevances
leave-one-out
125. Machine learning analysis of FDG-PET
brain images for the diagnosis of
neurodegenerative disorders
K.L. Leenders, S. Meles, … UMCG Groningen, Neurology
R. van Veen, S. Lövdal Bernoulli Institute, Computer Science
…
128. AMALEA 2022 42
Subjects
Source HC PD AD
CUN 19 49 -
UGOSM 44 58 55
UMCG 19 20 21
FDG-PET brain scans from 3 centers
• Clínica Universidad de Navarra
• Univ. Genova/IRCCS San Martino
• Univ. Medical Center Groningen
Glucose
uptake
http://glimpsproject.com
subjects
A
B
C
FDG-PET 3D images
Fluorodeoxyglucose
positron emission tomography
Healthy Controls HC
Parkinson’s Disease PD
Alzheimer’s Disease AD
data
130. AMALEA 2022 43
work flow
subjects
~
200000
voxels
subject specific
anatomy
high intensity,
low noise voxels
log-transform
double centering
masking (*)
low-dimensional
projections (*) details of pre-processing:
D. Mudali et al.
Computational and Mathematical Methods in Medicine.
March 2015, Art.ID 136921 and references. therein
(*) Scaled Subprofile Model / PCA based
on a disjoint reference group of subjects
131. AMALEA 2022 43
work flow
subjects
~
200000
voxels
subject specific
anatomy
high intensity,
low noise voxels
log-transform
double centering
masking (*)
low-dimensional
projections (*)
subject
socres
subjects
details of pre-processing:
D. Mudali et al.
Computational and Mathematical Methods in Medicine.
March 2015, Art.ID 136921 and references. therein
(*) Scaled Subprofile Model / PCA based
on a disjoint reference group of subjects
133. AMALEA 2022 44
work flow
subjects
subject
socres
subjects
applied to
novel subject
test
labels
(condition)
classification:
GMLVQ, SVM
?
~
200000
voxels
134. AMALEA 2022 44
work flow
subjects
subject
socres
subjects
applied to
novel subject
test
labels
(condition)
classification:
GMLVQ, SVM
?
~
200000
voxels
135. AMALEA 2022 45
(A) Perceptron of optimal stability (aka “SVM with linear kernel”)
- linear threshold classifier
- large margin (with errors)
two classifiers
136. AMALEA 2022 45
(A) Perceptron of optimal stability (aka “SVM with linear kernel”)
- linear threshold classifier
- large margin (with errors)
two classifiers
(B) Learning Vector Quantization (Generalized Matrix LVQ)
- prototype- and distance-based classifier
- relevance learning
137. AMALEA 2022 45
(A) Perceptron of optimal stability (aka “SVM with linear kernel”)
- linear threshold classifier
- large margin (with errors)
performance evaluation:
averages over 10 randomized runs of 10-fold cross-validation
accuracies, sensitivity /specificity
Receiver Operating Characteristics for binary classification
both classifiers outperformed Decision Trees in previous projects
two classifiers
(B) Learning Vector Quantization (Generalized Matrix LVQ)
- prototype- and distance-based classifier
- relevance learning
139. AMALEA 2022 46
results
subjects from one center only
here: UGOSM unbiased classifiers ROC
relatively good within-center performance
also in three-class settings
±0.008
142. AMALEA 2022 48
results
subjects from one center only
here: UGOSM, PD vs. AD unbiased classifiers ROC
PD vs. AD
subjects from centers combined for training and testing
example: UMCG and UGOSM
143. AMALEA 2022 48
results
subjects from one center only
here: UGOSM, PD vs. AD unbiased classifiers ROC
PD vs. AD
subjects from centers combined for training and testing
example: UMCG and UGOSM
reasonable (yet lower) overall performance
also in the other classification problems
144. AMALEA 2022 49
here: PD vs. HC
unbiased classifiers ROC
within center
(example: UGMOS)
results
145. AMALEA 2022 49
here: PD vs. HC
unbiased classifiers ROC
within center
(example: UGMOS)
across centers: poor performance
results
146. AMALEA 2022
50
UMCG vs UGOSM
experiment - classify subjects according to medical center
here: AD patients only
results/conclusions
147. AMALEA 2022
50
UMCG vs UGOSM
experiment - classify subjects according to medical center
here: AD patients only
results/conclusions
possible explanations:
- center-specific (pre-)processing
despite identical equipment and work flows
- significantly different patient cohorts (not the case)
need for more consistent protocols, calibration / pre-processing
aim: unified classifiers with good inter-center performance
148. AMALEA 2022 51
Matlab:
K Bunte: Relevance and Matrix adaptation in Learning Vector
Quantization (GRLVQ, GMLVQ and LiRaM LVQ)
-> code
F Westerman, R Veen, M.B: A no-nonsense beginners’ tool for GMLVQ
http://www.cs.rug.nl/~biehl/gmlvq
sklvq: Scikit Learning Vector Quantization
R van Veen, GJ de Vries, M. Biehl, JMLR 22 (2021), 1-6
https://www.cs.rug.nl/~biehl/
CITEC Bielefeld: scikit-learn compatible LVQ implementations
from the machine learning group at CITEC Bielefeld:
Java:
plug-in for WEKA from the CI Group Mittweida
M. Kästner, T. Villmann
Python: