Michael Biehl presented prototype-based machine learning methods. Prototype models represent data as exemplars and allow for interpretable and flexible classification. Unsupervised methods like vector quantization and self-organizing maps perform clustering. Supervised learning examples include Learning Vector Quantization for classification and Generalized Matrix Relevance LVQ, which learns distance metrics from data. Prototype models provide insights into data structures while achieving high accuracy.
Tutorial at the Winter School on Machine Learning, Gran Canaria, January 2020 (ppsx format, 52 slides)
Michael Biehl, University of Groningen, The Netherlands
Novel Machine Learning Methods for Extraction of Features Characterizing Data...Velimir (monty) Vesselinov
Vesselinov, V.V., Novel Machine Learning Methods for Extraction of Features Characterizing Datasets and Models, AGU Fall meeting, Washington D.C., 2018.
The variational Gaussian process (VGP), a Bayesian nonparametric model which adapts its shape to match com- plex posterior distributions. The VGP generates approximate posterior samples by generating latent inputs and warping them through random non-linear mappings; the distribution over random mappings is learned during inference, enabling the transformed outputs to adapt to varying complexity.
Universal Approximation Property via Quantum Feature Maps
----
The quantum Hilbert space can be used as a quantum-enhanced feature space in machine learning (ML) via the quantum feature map to encode classical data into quantum states. We prove the ability to approximate any continuous function with optimal approximation rate via quantum ML models in typical quantum feature maps.
---
Contributed talk at Quantum Techniques in Machine Learning 2021, Tokyo, November 8-12 2021.
By Quoc Hoan Tran, Takahiro Goto and Kohei Nakajima
Tutorial at the Winter School on Machine Learning, Gran Canaria, January 2020 (ppsx format, 52 slides)
Michael Biehl, University of Groningen, The Netherlands
Novel Machine Learning Methods for Extraction of Features Characterizing Data...Velimir (monty) Vesselinov
Vesselinov, V.V., Novel Machine Learning Methods for Extraction of Features Characterizing Datasets and Models, AGU Fall meeting, Washington D.C., 2018.
The variational Gaussian process (VGP), a Bayesian nonparametric model which adapts its shape to match com- plex posterior distributions. The VGP generates approximate posterior samples by generating latent inputs and warping them through random non-linear mappings; the distribution over random mappings is learned during inference, enabling the transformed outputs to adapt to varying complexity.
Universal Approximation Property via Quantum Feature Maps
----
The quantum Hilbert space can be used as a quantum-enhanced feature space in machine learning (ML) via the quantum feature map to encode classical data into quantum states. We prove the ability to approximate any continuous function with optimal approximation rate via quantum ML models in typical quantum feature maps.
---
Contributed talk at Quantum Techniques in Machine Learning 2021, Tokyo, November 8-12 2021.
By Quoc Hoan Tran, Takahiro Goto and Kohei Nakajima
Kernel methods and variable selection for exploratory analysis and multi-omic...tuxette
Nathalie Vialaneix
4th course on Computational Systems Biology of Cancer: Multi-omics and Machine Learning Approaches
International course, Curie training
https://training.institut-curie.org/courses/sysbiocancer2021
(remote)
September 29th, 2021
An Importance Sampling Approach to Integrate Expert Knowledge When Learning B...NTNU
The introduction of expert knowledge when learning Bayesian Networks from data is known to be an excellent approach to boost the performance of automatic learning methods, specially when the data is scarce. Previous approaches for this problem based on Bayesian statistics introduce the expert knowledge modifying the prior probability distributions. In this study, we propose a new methodology based on Monte Carlo simulation which starts with non-informative priors and requires knowledge from the expert a posteriori, when the simulation ends. We also explore a new Importance Sampling method for Monte Carlo simulation and the definition of new non-informative priors for the structure of the network. All these approaches are experimentally validated with five standard Bayesian networks.
Read more:
http://link.springer.com/chapter/10.1007%2F978-3-642-14049-5_70
Conventional tools in array signal processing have traditionally relied on the availability of a large number of samples acquired at each sensor or array element (antenna, hydrophone, microphone, etc.). Large sample size assumptions typically guarantee the consistency of estimators, detectors, classifiers and multiple other widely used signal processing procedures. However, practical scenario and array mobility conditions, together with the need for low latency and reduced scanning times, impose strong limits on the total number of observations that can be effectively processed. When the number of collected samples per sensor is small, conventional large sample asymptotic approaches are not relevant anymore. Recently, large random matrix theory tools have been proposed in order to address the small sample support problem in array signal processing. In fact, it has been shown that the most important and longstanding problems in this field can be reformulated and studied according to this asymptotic paradigm. By exploiting the latest advances in large random matrix theory and high dimensional statistics, a novel and unconventional methodology can be established, which provides an unprecedented treatment of the finite sample-per-sensor regime. In this talk, we will see that random matrix theory establishes a unifying framework for the study of array signal processing techniques under the constraint of a small number of observations per sensor, which has radically changed the way in which array processing methodologies have been traditionally established. We will show how this unconventional way of revisiting classical array processing has lead to major advances in the design and analysis of signal processing techniques for multidimensional observations.
There is now a huge literature on Bayesian methods for variable selection that use spike-and-slab priors. Such methods, in particular, have been quite successful for applications in a variety of different fields. High-throughput genomics and neuroimaging are two of such examples. There, novel methodological questions are being generated, requiring the integration of different concepts, methods, tools and data types. These have in particular motivated the development of variable selection priors that go beyond the independence assumptions of a simple Bernoulli prior on the variable inclusion indicators. In this talk I will describe various prior constructions that incorporate information about structural dependencies among the variables. I will also address extensions of the models to the analysis of count data. I will motivate the development of the models using specific applications from neuroimaging and from studies that use microbiome data.
Introduction to search and optimisation for the design theoristAkin Osman Kazakci
An historically important design theory is the state-space search by Herbert Simon. Over the years, the importance of this model has been consistently downplayed for various reasons. Today, it is not being used or discussed very frequently - except to downplay its significance even more - usually without an in-depth analysis.
However, the young generation of (design) researchers do not know well-enough the underlying formalism and how it can be used to interpret design phenomena.
This short introduction intends to give the basics of search, optimisation and problem-solving formalisms in a very intuitive way - which also helps to understand more complicated formal models of design.
Unsupervised Learning and Image Classification in High Performance Computing ...HPCC Systems
Itauma Itauma, Wayne State University, presents at the 2016 HPCC Systems Engineering Summit Community Day.
, Detroit, MI
At DSAL (Data Sciences and Analytics Lab) at Wayne State University, the focus is on studying, researching and developing algorithms that are applicable to various big data analysis problems occurring in areas of Machine Learning, Data Mining, Bioinformatics, and Healthcare Informatics.
The objective of this presentation is to share with the community our published work on, “Unsupervised Learning and Image Classification in High Performance Computing Cluster”.
Identifying good features has various benefits for object classification with respect to reducing computational cost and increasing classification accuracy. In our study, we implement a new multimodal feature learning method and object identification framework using High Performance Computing Cluster (HPCC Systems) that leads to faster optimization/calculation of algorithms with low costs of hardware designs.
The framework first learns representative weights over unlabeled data for each model through the K-means unsupervised learning method. Then, the desired features are extracted from the labeled data using the correlation between the labeled data and representative bases. These labeled features are fused and fed to the classifiers to make the final recognition. Our algorithms are implemented in ECL and we made use of the HPCC Systems machine learning library. Our framework is evaluated using various databases such as the CALTECH-101, AR databases, and a subset of wild PubFig83 data in which multimedia content is added.
We show that HPCC Systems can be used by machine learning researchers to speed up the running time of any computationally intensive algorithm; lowers budget costs by using existing computers instead of designing an expensive system with GPUs; and is scalable with respect to code reuse irrespective of the size of the dataset and number of nodes configured. Our novel identity recognition algorithm can lead to further exploration of face recognition problems using the HPCC Systems environment.
Itauma Itauma
Itauma Itauma has an undergraduate degree in Electrical Engineering from the University of Ilorin and two Masters degrees, an MSc. in Computer Engineering from Istanbul Technical University, majoring in human-robot interaction and an MSc. in Computer Science from Wayne State University where his thesis was based on leveraging HPCC Systems for Big Data analytics. He is currently a PhD student in Instructional Design and Technology at Keiser University and a Part-Time Computer Science Lecturer at Wayne State University. He loves teaching. His interests lie in the areas of robotics, big data, and social and behavioral sciences. His current research involves using natural language processing tools in processing unstructured data, thereby automating and optimizing the process of qualitative data analysis.
Learning to discover monte carlo algorithm on spin ice manifoldKai-Wen Zhao
The global update Monte Carlo sampler can be discovered naturally by trained machine using policy gradient method on topologically constrained environment.
Invited lecture on Machine Learning in Medicine at the joint "Integrated Omics" course of Hanze University and University Hospital UMCG, Groningen, The Netherlands
Kernel methods and variable selection for exploratory analysis and multi-omic...tuxette
Nathalie Vialaneix
4th course on Computational Systems Biology of Cancer: Multi-omics and Machine Learning Approaches
International course, Curie training
https://training.institut-curie.org/courses/sysbiocancer2021
(remote)
September 29th, 2021
An Importance Sampling Approach to Integrate Expert Knowledge When Learning B...NTNU
The introduction of expert knowledge when learning Bayesian Networks from data is known to be an excellent approach to boost the performance of automatic learning methods, specially when the data is scarce. Previous approaches for this problem based on Bayesian statistics introduce the expert knowledge modifying the prior probability distributions. In this study, we propose a new methodology based on Monte Carlo simulation which starts with non-informative priors and requires knowledge from the expert a posteriori, when the simulation ends. We also explore a new Importance Sampling method for Monte Carlo simulation and the definition of new non-informative priors for the structure of the network. All these approaches are experimentally validated with five standard Bayesian networks.
Read more:
http://link.springer.com/chapter/10.1007%2F978-3-642-14049-5_70
Conventional tools in array signal processing have traditionally relied on the availability of a large number of samples acquired at each sensor or array element (antenna, hydrophone, microphone, etc.). Large sample size assumptions typically guarantee the consistency of estimators, detectors, classifiers and multiple other widely used signal processing procedures. However, practical scenario and array mobility conditions, together with the need for low latency and reduced scanning times, impose strong limits on the total number of observations that can be effectively processed. When the number of collected samples per sensor is small, conventional large sample asymptotic approaches are not relevant anymore. Recently, large random matrix theory tools have been proposed in order to address the small sample support problem in array signal processing. In fact, it has been shown that the most important and longstanding problems in this field can be reformulated and studied according to this asymptotic paradigm. By exploiting the latest advances in large random matrix theory and high dimensional statistics, a novel and unconventional methodology can be established, which provides an unprecedented treatment of the finite sample-per-sensor regime. In this talk, we will see that random matrix theory establishes a unifying framework for the study of array signal processing techniques under the constraint of a small number of observations per sensor, which has radically changed the way in which array processing methodologies have been traditionally established. We will show how this unconventional way of revisiting classical array processing has lead to major advances in the design and analysis of signal processing techniques for multidimensional observations.
There is now a huge literature on Bayesian methods for variable selection that use spike-and-slab priors. Such methods, in particular, have been quite successful for applications in a variety of different fields. High-throughput genomics and neuroimaging are two of such examples. There, novel methodological questions are being generated, requiring the integration of different concepts, methods, tools and data types. These have in particular motivated the development of variable selection priors that go beyond the independence assumptions of a simple Bernoulli prior on the variable inclusion indicators. In this talk I will describe various prior constructions that incorporate information about structural dependencies among the variables. I will also address extensions of the models to the analysis of count data. I will motivate the development of the models using specific applications from neuroimaging and from studies that use microbiome data.
Introduction to search and optimisation for the design theoristAkin Osman Kazakci
An historically important design theory is the state-space search by Herbert Simon. Over the years, the importance of this model has been consistently downplayed for various reasons. Today, it is not being used or discussed very frequently - except to downplay its significance even more - usually without an in-depth analysis.
However, the young generation of (design) researchers do not know well-enough the underlying formalism and how it can be used to interpret design phenomena.
This short introduction intends to give the basics of search, optimisation and problem-solving formalisms in a very intuitive way - which also helps to understand more complicated formal models of design.
Unsupervised Learning and Image Classification in High Performance Computing ...HPCC Systems
Itauma Itauma, Wayne State University, presents at the 2016 HPCC Systems Engineering Summit Community Day.
, Detroit, MI
At DSAL (Data Sciences and Analytics Lab) at Wayne State University, the focus is on studying, researching and developing algorithms that are applicable to various big data analysis problems occurring in areas of Machine Learning, Data Mining, Bioinformatics, and Healthcare Informatics.
The objective of this presentation is to share with the community our published work on, “Unsupervised Learning and Image Classification in High Performance Computing Cluster”.
Identifying good features has various benefits for object classification with respect to reducing computational cost and increasing classification accuracy. In our study, we implement a new multimodal feature learning method and object identification framework using High Performance Computing Cluster (HPCC Systems) that leads to faster optimization/calculation of algorithms with low costs of hardware designs.
The framework first learns representative weights over unlabeled data for each model through the K-means unsupervised learning method. Then, the desired features are extracted from the labeled data using the correlation between the labeled data and representative bases. These labeled features are fused and fed to the classifiers to make the final recognition. Our algorithms are implemented in ECL and we made use of the HPCC Systems machine learning library. Our framework is evaluated using various databases such as the CALTECH-101, AR databases, and a subset of wild PubFig83 data in which multimedia content is added.
We show that HPCC Systems can be used by machine learning researchers to speed up the running time of any computationally intensive algorithm; lowers budget costs by using existing computers instead of designing an expensive system with GPUs; and is scalable with respect to code reuse irrespective of the size of the dataset and number of nodes configured. Our novel identity recognition algorithm can lead to further exploration of face recognition problems using the HPCC Systems environment.
Itauma Itauma
Itauma Itauma has an undergraduate degree in Electrical Engineering from the University of Ilorin and two Masters degrees, an MSc. in Computer Engineering from Istanbul Technical University, majoring in human-robot interaction and an MSc. in Computer Science from Wayne State University where his thesis was based on leveraging HPCC Systems for Big Data analytics. He is currently a PhD student in Instructional Design and Technology at Keiser University and a Part-Time Computer Science Lecturer at Wayne State University. He loves teaching. His interests lie in the areas of robotics, big data, and social and behavioral sciences. His current research involves using natural language processing tools in processing unstructured data, thereby automating and optimizing the process of qualitative data analysis.
Learning to discover monte carlo algorithm on spin ice manifoldKai-Wen Zhao
The global update Monte Carlo sampler can be discovered naturally by trained machine using policy gradient method on topologically constrained environment.
Invited lecture on Machine Learning in Medicine at the joint "Integrated Omics" course of Hanze University and University Hospital UMCG, Groningen, The Netherlands
A tutorial given at the AMALEA workshop 2022:
Unsupervised and supervised prototype-based learning is illustrated in terms of bio-medical applications.
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology.
3D Scene Analysis via Sequenced Predictions over Points and RegionsFlavia Grosan
I gave this talk in Machine Vision seminar at Jacobs University. I presented the state of the art in 3D point cloud classification and I described X. Xiong et al approach in a paper published in 2010.
Classification of Iris Data using Kernel Radial Basis Probabilistic Neural N...Scientific Review SR
Radial Basis Probabilistic Neural Network (RBPNN) has a broader generalized capability that been
successfully applied to multiple fields. In this paper, the Euclidean distance of each data point in RBPNN is
extended by calculating its kernel-induced distance instead of the conventional sum-of squares distance. The
kernel function is a generalization of the distance metric that measures the distance between two data points as the
data points are mapped into a high dimensional space. During the comparing of the four constructed classification
models with Kernel RBPNN, Radial Basis Function networks, RBPNN and Back-Propagation networks as
proposed, results showed that, model classification on Iris Data with Kernel RBPNN display an outstanding
performance in this regard
Similar to 2017: Prototype-based models in unsupervised and supervised machine learning (20)
Interpretable machine learning in endocrinology, M. Biehl, APPIS 2024University of Groningen
An introduction to interpretable machine learning in endocrinology.
In particular, the application of Generalized Matrix Relevance LVQ to the classification of andrenocortical tumors and the differential diagnosis of primary aldosteronism is given.
A tutorial given at the AMALEA workshop 2022.
This talk presents the statistical physics based theory of machine learning in terms of simple example systems. As a recent application, the occurrence of phase transitions in layered networks is discussed.
The statistical physics of learning revisted: Phase transitions in layered ne...University of Groningen
"The statistical physics of learning revisted: Phase transitions in layered neural networks"
Physics Colloquium at the University of Leipzig/Germany, June 29, 2021
24 slides, ca 45 minutes
Short presentation (15 minutes) focussing on the application of unsupervised and supervised machine learning in the paper "Tissue- and development-stage specific mRNA and heterogeneous CNV signatures of human ribosomal proteins in normal and cancer samples
Talk presented at WSOM 2016 in Houston/Texas.
Machine learning based classification of FDG-PET scan data for the diagnosis of neurodegenerative disorders
June 2017: Biomedical applications of prototype-based classifiers and relevan...University of Groningen
A presentation of several biomedical applications of prototype-based machine learning and relevance learning. Invited talk at the AlCoB conference 2017 in Aveiro/Portugal.
THE IMPORTANCE OF MARTIAN ATMOSPHERE SAMPLE RETURN.Sérgio Sacani
The return of a sample of near-surface atmosphere from Mars would facilitate answers to several first-order science questions surrounding the formation and evolution of the planet. One of the important aspects of terrestrial planet formation in general is the role that primary atmospheres played in influencing the chemistry and structure of the planets and their antecedents. Studies of the martian atmosphere can be used to investigate the role of a primary atmosphere in its history. Atmosphere samples would also inform our understanding of the near-surface chemistry of the planet, and ultimately the prospects for life. High-precision isotopic analyses of constituent gases are needed to address these questions, requiring that the analyses are made on returned samples rather than in situ.
DERIVATION OF MODIFIED BERNOULLI EQUATION WITH VISCOUS EFFECTS AND TERMINAL V...Wasswaderrick3
In this book, we use conservation of energy techniques on a fluid element to derive the Modified Bernoulli equation of flow with viscous or friction effects. We derive the general equation of flow/ velocity and then from this we derive the Pouiselle flow equation, the transition flow equation and the turbulent flow equation. In the situations where there are no viscous effects , the equation reduces to the Bernoulli equation. From experimental results, we are able to include other terms in the Bernoulli equation. We also look at cases where pressure gradients exist. We use the Modified Bernoulli equation to derive equations of flow rate for pipes of different cross sectional areas connected together. We also extend our techniques of energy conservation to a sphere falling in a viscous medium under the effect of gravity. We demonstrate Stokes equation of terminal velocity and turbulent flow equation. We look at a way of calculating the time taken for a body to fall in a viscous medium. We also look at the general equation of terminal velocity.
Deep Behavioral Phenotyping in Systems Neuroscience for Functional Atlasing a...Ana Luísa Pinho
Functional Magnetic Resonance Imaging (fMRI) provides means to characterize brain activations in response to behavior. However, cognitive neuroscience has been limited to group-level effects referring to the performance of specific tasks. To obtain the functional profile of elementary cognitive mechanisms, the combination of brain responses to many tasks is required. Yet, to date, both structural atlases and parcellation-based activations do not fully account for cognitive function and still present several limitations. Further, they do not adapt overall to individual characteristics. In this talk, I will give an account of deep-behavioral phenotyping strategies, namely data-driven methods in large task-fMRI datasets, to optimize functional brain-data collection and improve inference of effects-of-interest related to mental processes. Key to this approach is the employment of fast multi-functional paradigms rich on features that can be well parametrized and, consequently, facilitate the creation of psycho-physiological constructs to be modelled with imaging data. Particular emphasis will be given to music stimuli when studying high-order cognitive mechanisms, due to their ecological nature and quality to enable complex behavior compounded by discrete entities. I will also discuss how deep-behavioral phenotyping and individualized models applied to neuroimaging data can better account for the subject-specific organization of domain-general cognitive systems in the human brain. Finally, the accumulation of functional brain signatures brings the possibility to clarify relationships among tasks and create a univocal link between brain systems and mental functions through: (1) the development of ontologies proposing an organization of cognitive processes; and (2) brain-network taxonomies describing functional specialization. To this end, tools to improve commensurability in cognitive science are necessary, such as public repositories, ontology-based platforms and automated meta-analysis tools. I will thus discuss some brain-atlasing resources currently under development, and their applicability in cognitive as well as clinical neuroscience.
This presentation explores a brief idea about the structural and functional attributes of nucleotides, the structure and function of genetic materials along with the impact of UV rays and pH upon them.
Richard's aventures in two entangled wonderlandsRichard Gill
Since the loophole-free Bell experiments of 2020 and the Nobel prizes in physics of 2022, critics of Bell's work have retreated to the fortress of super-determinism. Now, super-determinism is a derogatory word - it just means "determinism". Palmer, Hance and Hossenfelder argue that quantum mechanics and determinism are not incompatible, using a sophisticated mathematical construction based on a subtle thinning of allowed states and measurements in quantum mechanics, such that what is left appears to make Bell's argument fail, without altering the empirical predictions of quantum mechanics. I think however that it is a smoke screen, and the slogan "lost in math" comes to my mind. I will discuss some other recent disproofs of Bell's theorem using the language of causality based on causal graphs. Causal thinking is also central to law and justice. I will mention surprising connections to my work on serial killer nurse cases, in particular the Dutch case of Lucia de Berk and the current UK case of Lucy Letby.
Earliest Galaxies in the JADES Origins Field: Luminosity Function and Cosmic ...Sérgio Sacani
We characterize the earliest galaxy population in the JADES Origins Field (JOF), the deepest
imaging field observed with JWST. We make use of the ancillary Hubble optical images (5 filters
spanning 0.4−0.9µm) and novel JWST images with 14 filters spanning 0.8−5µm, including 7 mediumband filters, and reaching total exposure times of up to 46 hours per filter. We combine all our data
at > 2.3µm to construct an ultradeep image, reaching as deep as ≈ 31.4 AB mag in the stack and
30.3-31.0 AB mag (5σ, r = 0.1” circular aperture) in individual filters. We measure photometric
redshifts and use robust selection criteria to identify a sample of eight galaxy candidates at redshifts
z = 11.5 − 15. These objects show compact half-light radii of R1/2 ∼ 50 − 200pc, stellar masses of
M⋆ ∼ 107−108M⊙, and star-formation rates of SFR ∼ 0.1−1 M⊙ yr−1
. Our search finds no candidates
at 15 < z < 20, placing upper limits at these redshifts. We develop a forward modeling approach to
infer the properties of the evolving luminosity function without binning in redshift or luminosity that
marginalizes over the photometric redshift uncertainty of our candidate galaxies and incorporates the
impact of non-detections. We find a z = 12 luminosity function in good agreement with prior results,
and that the luminosity function normalization and UV luminosity density decline by a factor of ∼ 2.5
from z = 12 to z = 14. We discuss the possible implications of our results in the context of theoretical
models for evolution of the dark matter halo mass function.
What is greenhouse gasses and how many gasses are there to affect the Earth.moosaasad1975
What are greenhouse gasses how they affect the earth and its environment what is the future of the environment and earth how the weather and the climate effects.
The ability to recreate computational results with minimal effort and actionable metrics provides a solid foundation for scientific research and software development. When people can replicate an analysis at the touch of a button using open-source software, open data, and methods to assess and compare proposals, it significantly eases verification of results, engagement with a diverse range of contributors, and progress. However, we have yet to fully achieve this; there are still many sociotechnical frictions.
Inspired by David Donoho's vision, this talk aims to revisit the three crucial pillars of frictionless reproducibility (data sharing, code sharing, and competitive challenges) with the perspective of deep software variability.
Our observation is that multiple layers — hardware, operating systems, third-party libraries, software versions, input data, compile-time options, and parameters — are subject to variability that exacerbates frictions but is also essential for achieving robust, generalizable results and fostering innovation. I will first review the literature, providing evidence of how the complex variability interactions across these layers affect qualitative and quantitative software properties, thereby complicating the reproduction and replication of scientific studies in various fields.
I will then present some software engineering and AI techniques that can support the strategic exploration of variability spaces. These include the use of abstractions and models (e.g., feature models), sampling strategies (e.g., uniform, random), cost-effective measurements (e.g., incremental build of software configurations), and dimensionality reduction methods (e.g., transfer learning, feature selection, software debloating).
I will finally argue that deep variability is both the problem and solution of frictionless reproducibility, calling the software science community to develop new methods and tools to manage variability and foster reproducibility in software systems.
Exposé invité Journées Nationales du GDR GPL 2024
Professional air quality monitoring systems provide immediate, on-site data for analysis, compliance, and decision-making.
Monitor common gases, weather parameters, particulates.
Comparing Evolved Extractive Text Summary Scores of Bidirectional Encoder Rep...University of Maribor
Slides from:
11th International Conference on Electrical, Electronics and Computer Engineering (IcETRAN), Niš, 3-6 June 2024
Track: Artificial Intelligence
https://www.etran.rs/2024/en/home-english/
Comparing Evolved Extractive Text Summary Scores of Bidirectional Encoder Rep...
2017: Prototype-based models in unsupervised and supervised machine learning
1. Michael Biehl, Aleke Nolte
Johann Bernoulli Institute for
Mathematics and Computer Science
University of Groningen, NL
SUNDIAL H2020 Network
www.cs.rug.nl/~biehl
www.astro.rug.nl/~sundial/
pre- reprints, available code
Prototype-based models in unsupervised
and supervised machine learning
Lingyu Wang
Kapteyn Astronomical Inst.
and SRON Groningen
Astrophysics Science Group
Groningen, NL
3. Astroinformatics, Cape Town, November 2017
Introduction
prototypes, exemplars:
representation of information in terms of
typical representatives (e.g. of a class of objects),
much debated concept in cognitive psychology
machine learning: prototype- (and distance-) based systems
- easy to implement, highly flexible, online training
- white box: parameterization in the space of observed data
- yield interpretable classifiers/regression systems
- help to detect bias in training data, other artifacts
- provide insights into data set / problem at hand
Accuracy is not enough! [Paulo Lisboa]
4. Astroinformatics, Cape Town, November 2017
4
Introduction
neural interpretation: activation and learning in a shallow network
external stimulus to a network of neurons
response according to weights (= expected inputs)
activation: BMU - best matching unit (and neighbors)
learning -> even stronger response to the same stimulus in future
weights represent different expected stimuli (prototypes)
5. Astroinformatics, Cape Town, November 2017
based on dis-similarity/distance measure
assignment to prototypes: e.g. Nearest Prototype Scheme
given vector xμ , determine winner
(BMU)
→ assign xμ to prototype w*
most popular example: (squared) Euclidean distance
Vector Quantization (VQ)
VQ system: set of prototypes
data: set of feature vectors
Vector Quantization: identify typical representatives of data
which capture essential features
6. Astroinformatics, Cape Town, November 2017 6
random sequential (repeated) presentation of data
… the Winner Takes it All (WTA):
initially: randomized wk, e.g. in randomly selected data points
Competitive Learning
η (<1): learning rate, step size of update
competitive VQ: competition without neighborhood cooperativeness
stochastic gradient descent minimization of the
Quantization Error
(here: sq. Euclidean)
7. Astroinformatics, Cape Town, November 2017
Self-Organizing Map (SOM)
T. Kohonen. Self-Organizing Maps (Springer 1995, 1997, 2001)
neighborhood cooperativeness on a pre-defined low-dim. lattice
d-dim. lattice A of
neurons (prototypes)
- update BMU and lattice neighborhood:
where
range ρ w.r.t. distances in lattice A
upon presentation of xμ :
- determine the Best Matching Unit
at position s in the lattice
10. Astroinformatics, Cape Town, November 2017
10
11
12
.
.
41
Illustration: Galaxy Characteristics
Numerical features describing a catalogue of galaxies
work in progress - details not (yet) disclosed
GAMA: Galaxy and
Mass Assembly Survey
www.gama-survey.org
reduced
set of 10
selected
features
full set
of 41
features
(semi-major)
(semi-minor)
logistic normalization:
11. Astroinformatics, Cape Town, November 2017
class 1
class 3
class 4
7
class 5
class 6
class 2 8,9
1 - elliptical E0-E6
3 – “early type spirals”
4 – “early type barred spirals”
5 – “intermediate type spirals”
6 – “intermediate type, barred”
7 – “late type spirals & irregulars”
Illustration: Galaxy Classification
2 - Little Blue Spheroids (LBS)
“
8,9 – artefacts, stars
Kelvin et al., MNRAS 439: 1245-1269, 2014.
13. Astroinformatics, Cape Town, November 2017 13
SOM (rectangular grid, ‘medium size’)
unsupervised clustering
pie-charts:
percentage at which classes
are assigned to a particular unit
Self-Organizing Map
observations / suggestions:
- LBS appear well separated
- overlap of 1 / 3 and 5 / 7
with smooth transtions
- 6 and 5 mix/overlap
- “small classes” 4,8,9 hardly
represented
to do: inspect prototypes, U-matrix, ...
meta-clustering
14. Astroinformatics, Cape Town, November 2017
∙ identification of prototype vectors from labeled example data
∙ distance based classification (e.g. Euclidean)
Supervised Competitive Learning
N-dimensional data, feature vectors
• initialize prototype vectors
for different classes
Learning Vector Quantization here: heuristic LVQ1 [Kohonen, 1990]
• identify the winner
(closest prototype)
• present a single example
• move the winner
- closer towards the data (same class)
- away from the data (different class)
Alternatives: cost function based training
e.g. Generalized LVQ [ GLVQ: Sato and Yamada, 1995]
15. Astroinformatics, Cape Town, November 2017
∙ identification of prototype vectors from labeled example data
∙ distance based classification (e.g. Euclidean)
Learning Vector Quantization
N-dimensional data, feature vectors
∙ distance-based classification
[here: Euclidean distances]
∙ generalization ability
correct classification of new data
∙ aim: discrimination of classes
( ≠ vector quantization
or density estimation )
Nearest Prototype Classifier
16. Astroinformatics, Cape Town, November 2017 16
Distance Measures
fixed distance measures:
- select distance measures (prior knowledge, pre-processing)
- compare performance of various measures
relevance learning: adaptive distance measures
- fix only parametric form of distance measure
- data driven adaptation:
determine prototypes and distance parameters
in the same training process (e.g. cost function based GLVQ)
Example: Generalized Matrix Relevance LVQ
(Adaptive)
[Schneider, Biehl, Hammer, 2009]
17. Astroinformatics, Cape Town, November 2017
Generalized Relevance Matrix LVQ (GMLVQ)
adaptive quadratic distance in LVQ:
normalization:
summarizes
- the contribution of the original dimension j
- relevance of original features for the classification
standard (squared) Euclidean distance for
linearly transformed features
: relevance of pairs (i,j) of features
18. Astroinformatics, Cape Town, November 2017 18
- restriction to classes with significant number of samples
- sub-sampling in order to achieve balanced training sets (5×743)
- use of all 41 features
- avgerages over random splits in 90% training, 10% test set
GMLVQ analysis
one prototype
per class
1 2
3 5
7
confusion matrix of the NPC
61.3 10.4 20.1 7.5 0.7
3.1 90.5 0 1.9 4.5
16.5 1.7 68.0 13.6 0.2
1.6 7.8 10.0 73.6 7.0
1.3 13.0 0.3 13.8 71.
6
predicted
trueclass
19. Astroinformatics, Cape Town, November 2017 19
diagonal of the relevance matrix:
continuous weights
- alternative set of features ?
projection of the data set on leading
eigenvectors of Λ: discriminative
low-dim. representation:
e.g. strong overlap of classes 1 / 3
(elliptical / early type spirals)
- agrees only partially
with hand-crafted set ()
correlations between features?
GMLVQ analysis
20. Astroinformatics, Cape Town, November 2017 20
Summary
Prototype-based systems in machine learning:
represent data in terms of exemplars, white box
parameterization of clustering / classification / regression
Unsupervised Learning
data reduction, vector quantization, clustering
low-dimensional representation, topology preserving SOM
Supervised Learning
example: LVQ for classification with adaptive distance
Generalized Matrix Relevance LVQ (GMLVQ) *
white box, transparent, intuitive, powerful
accuracy is not enough: insight into problem / data set
e.g. with respect to feature selection / weighting
* GMLVQ (matlab) toolboxes: www.cs.rug.nl/~biehl
21. Astroinformatics, Cape Town, November 2017 21
Unsupervised Learning
Neural Gas (NG)
Generative Topographic Map (GTM)
Relevance learning in dimension reduction
Regression
Ordinal Regresssion in GMVLQ
Radial Basis Function networks (RBF)
Probabilistic classification
likelihood-based classifiers (Robust Soft LVQ)
Distances / Similarities
unconventional, problem-specific similarity measures
e.g. functional data (time series, spectra, histograms...)
non-vectorial data, relational data
relevances: weak/strong, bounds
...
there is a lot more...
1 – elliptical 5 – intermediate type spirals
2 - Little Blue Spheroids 6 - intermediate type barred
3 - early type spirals 7- irregular
4 – early type barred spirals
1 – elliptical 5 – intermediate type spirals
2 - Little Blue Spheroids 6 - intermediate type barred
3 - early type spirals 7- irregular
4 – early type barred spirals
1 – elliptical 5 – intermediate type spirals
2 - Little Blue Spheroids 6 - intermediate type barred
3 - early type spirals 7- irregular
4 – early type barred spirals
1 – elliptical 5 – intermediate type spirals
2 - Little Blue Spheroids 6 - intermediate type barred
3 - early type spirals 7- irregular
4 – early type barred spirals