1) The document discusses using kernel methods to test statistical dependence between random variables. It proposes a measure called constrained covariance (COCO) that finds the maximum covariance between two random variables after transforming them using reproducing kernel Hilbert spaces.
2) COCO is defined as the supremum of the covariance between smooth transformations of the random variables, where the transformations live in unit balls of RKHSs. This captures dependencies that may not be detected by simple correlation.
3) Empirically, COCO is estimated as the largest singular value of the covariance between kernel embeddings of samples from the two random variables. This provides a way to compute the proposed dependence measure in practice from data samples.
FPGA Implementation of A New Chien Search Block for Reed-Solomon Codes RS (25...IJERA Editor
The Reed-Solomon codes RS are widely used in communication systems, in particular forming part of the specification for the ETSI digital terrestrial television standard. In this paper a simple algorithm for error detection in the Chien Search block is proposed. This algorithm is based on a simple factorization of the error locator polynomial, which allows reducing the number of components required to implement the proposed algorithm on FPGA board. Consequently, it reduces the power consumption with a percentage which can reach 50 % compared to the basic RS decoder. First, we developed the design of Chien Search Block Second, we generated and simulated the hardware description language source code using Quartus software tools,finally we implemented the proposed algorithm of Chien search block for Reed-Solomon codesRS (255, 239) on FPGA board to show both the reduced hardware resources and low complexity compared to the basic algorithm.
This paper presents a novel SAT-based approach for the computation of extensions in abstract argumentation, with focus on preferred semantics, and an empirical evaluation of its performances. The approach is based on the idea of reducing the problem of computing complete extensions to a SAT problem and then using a depth-first search method to derive preferred extensions. The proposed approach has been tested using two distinct SAT solvers and compared with three state-of-the-art systems for preferred extension computation. It turns out that the proposed approach delivers significantly better performances in the large majority of the considered cases.
FPGA Implementation of A New Chien Search Block for Reed-Solomon Codes RS (25...IJERA Editor
The Reed-Solomon codes RS are widely used in communication systems, in particular forming part of the specification for the ETSI digital terrestrial television standard. In this paper a simple algorithm for error detection in the Chien Search block is proposed. This algorithm is based on a simple factorization of the error locator polynomial, which allows reducing the number of components required to implement the proposed algorithm on FPGA board. Consequently, it reduces the power consumption with a percentage which can reach 50 % compared to the basic RS decoder. First, we developed the design of Chien Search Block Second, we generated and simulated the hardware description language source code using Quartus software tools,finally we implemented the proposed algorithm of Chien search block for Reed-Solomon codesRS (255, 239) on FPGA board to show both the reduced hardware resources and low complexity compared to the basic algorithm.
This paper presents a novel SAT-based approach for the computation of extensions in abstract argumentation, with focus on preferred semantics, and an empirical evaluation of its performances. The approach is based on the idea of reducing the problem of computing complete extensions to a SAT problem and then using a depth-first search method to derive preferred extensions. The proposed approach has been tested using two distinct SAT solvers and compared with three state-of-the-art systems for preferred extension computation. It turns out that the proposed approach delivers significantly better performances in the large majority of the considered cases.
A New Enhanced Method of Non Parametric power spectrum Estimation.CSCJournals
The spectral analysis of non uniform sampled data sequences using Fourier Periodogram method is the classical approach. In view of data fitting and computational standpoints why the Least squares periodogram(LSP) method is preferable than the “classical” Fourier periodogram and as well as to the frequently-used form of LSP due to Lomb and Scargle is explained. Then a new method of spectral analysis of nonuniform data sequences can be interpreted as an iteratively weighted LSP that makes use of a data-dependent weighting matrix built from the most recent spectral estimate. It is iterative and it makes use of an adaptive (i.e., data-dependent) weighting, we refer to it as the iterative adaptive approach (IAA).LSP and IAA are nonparametric methods that can be used for the spectral analysis of general data sequences with both continuous and discrete spectra. However, they are most suitable for data sequences with discrete spectra (i.e., sinusoidal data), which is the case we emphasize in this paper. Of the existing methods for nonuniform sinusoidal data, Welch, MUSIC and ESPRIT methods appear to be the closest in spirit to the IAA proposed here. Indeed, all these methods make use of the estimated covariance matrix that is computed in the first iteration of IAA from LSP. MUSIC and ESPRIT, on the other hand, are parametric methods that require a guess of the number of sinusoidal components present in the data, otherwise they cannot be used; furthermore.
TMPA-2017: Generating Cost Aware Covering Arrays For Free Iosif Itkin
TMPA-2017: Tools and Methods of Program Analysis
3-4 March, 2017, Hotel Holiday Inn Moscow Vinogradovo, Moscow
Generating Cost Aware Covering Arrays For Free
Mustafa Kemal Tas, Hanefi Mercan, Gülşen Demiröz, Kamer Kaya, Cemal Yilmaz, Sabanci University
For video follow the link: https://youtu.be/Wkdd4A0rRjE
Would like to know more?
Visit our website:
www.tmpaconf.org
www.exactprosystems.com/events/tmpa
Follow us:
https://www.linkedin.com/company/exactpro-systems-llc?trk=biz-companies-cym
https://twitter.com/exactpro
Bayesian Inference and Uncertainty Quantification for Inverse ProblemsMatt Moores
So-called “inverse” problems arise when the parameters of a physical system cannot be directly observed. The mapping between these latent parameters and the space of noisy observations is represented as a mathematical model, often involving a system of differential equations. We seek to infer the parameter values that best fit our observed data. However, it is also vital to obtain accurate quantification of the uncertainty involved with these parameters, particularly when the output of the model will be used for forecasting. Bayesian inference provides well-calibrated uncertainty estimates, represented by the posterior distribution over the parameters. In this talk, I will give a brief introduction to Markov chain Monte Carlo (MCMC) algorithms for sampling from the posterior distribution and describe how they can be combined with numerical solvers for the forward model. We apply these methods to two examples of ODE models: growth curves in ecology, and thermogravimetric analysis (TGA) in chemistry. This is joint work with Matthew Berry, Mark Nelson, Brian Monaghan and Raymond Longbottom.
bayesImageS: Bayesian computation for medical Image Segmentation using a hidd...Matt Moores
There are many approaches to Bayesian computation with intractable likelihoods, including the exchange algorithm, approximate Bayesian computation (ABC), thermodynamic integration, and composite likelihood. These approaches vary in accuracy as well as scalability for datasets of significant size. The Potts model is an example where such methods are required, due to its intractable normalising constant. This model is a type of Markov random field, which is commonly used for image segmentation. The dimension of its parameter space increases linearly with the number of pixels in the image, making this a challenging application for scalable Bayesian computation. My talk will introduce various algorithms in the context of the Potts model and describe their implementation in C++, using OpenMP for parallelism.
R package 'bayesImageS': a case study in Bayesian computation using Rcpp and ...Matt Moores
There are many approaches to Bayesian computation with intractable likelihoods, including the exchange algorithm, approximate Bayesian computation (ABC), thermodynamic integration, and composite likelihood. These approaches vary in accuracy as well as scalability for datasets of significant size. The Potts model is an example where such methods are required, due to its intractable normalising constant. This model is a type of Markov random field, which is commonly used for image segmentation. The dimension of its parameter space increases linearly with the number of pixels in the image, making this a challenging application for scalable Bayesian computation. My talk will introduce various algorithms in the context of the Potts model and describe their implementation in C++, using OpenMP for parallelism. I will also discuss the process of releasing this software as an open source R package on the CRAN repository.
This is concerned with designing an exact exponential time algorithm that is better than the well-known 2^n algorithm for the problem Path Contraction. This answers an open question of van't Hof et. al [TCS 2009]. This is based on the article that appeared in ICALP 2019.
A new transformation into State Transition Algorithm for finding the global m...Michael_Chou
To promote the global search ability of the original state transition algorithm, a new operator called axesion is suggested, which aims to search along the axes and strengthen single dimensional search. Several benchmark minimization
problems are used to illustrate the advantages of the improved algorithm over other random search methods. The results of
numerical experiments show that the new transformation can enhance the performance of the state transition algorithm and the new strategy is effective and reliable.
new optimization algorithm for topology optimizationSeonho Park
authors devise new convex approximation called DQA which utilizes information of two consecutive points at iterates. Also, to guarantee global convergence, filter method is illustrated.
We approach the screening problem - i.e. detecting which inputs of a computer model significantly impact the output - from a formal Bayesian model selection point of view. That is, we place a Gaussian process prior on the computer model and consider the $2^p$ models that result from assuming that each of the subsets of the $p$ inputs affect the response. The goal is to obtain the posterior probabilities of each of these models. In this talk, we focus on the specification of objective priors on the model-specific parameters and on convenient ways to compute the associated marginal likelihoods. These two problems that normally are seen as unrelated, have challenging connections since the priors proposed in the literature are specifically designed to have posterior modes in the boundary of the parameter space, hence precluding the application of approximate integration techniques based on e.g. Laplace approximations. We explore several ways of circumventing this difficulty, comparing different methodologies with synthetic examples taken from the literature.
Authors: Gonzalo Garcia-Donato (Universidad de Castilla-La Mancha) and Rui Paulo (Universidade de Lisboa)
A New Enhanced Method of Non Parametric power spectrum Estimation.CSCJournals
The spectral analysis of non uniform sampled data sequences using Fourier Periodogram method is the classical approach. In view of data fitting and computational standpoints why the Least squares periodogram(LSP) method is preferable than the “classical” Fourier periodogram and as well as to the frequently-used form of LSP due to Lomb and Scargle is explained. Then a new method of spectral analysis of nonuniform data sequences can be interpreted as an iteratively weighted LSP that makes use of a data-dependent weighting matrix built from the most recent spectral estimate. It is iterative and it makes use of an adaptive (i.e., data-dependent) weighting, we refer to it as the iterative adaptive approach (IAA).LSP and IAA are nonparametric methods that can be used for the spectral analysis of general data sequences with both continuous and discrete spectra. However, they are most suitable for data sequences with discrete spectra (i.e., sinusoidal data), which is the case we emphasize in this paper. Of the existing methods for nonuniform sinusoidal data, Welch, MUSIC and ESPRIT methods appear to be the closest in spirit to the IAA proposed here. Indeed, all these methods make use of the estimated covariance matrix that is computed in the first iteration of IAA from LSP. MUSIC and ESPRIT, on the other hand, are parametric methods that require a guess of the number of sinusoidal components present in the data, otherwise they cannot be used; furthermore.
TMPA-2017: Generating Cost Aware Covering Arrays For Free Iosif Itkin
TMPA-2017: Tools and Methods of Program Analysis
3-4 March, 2017, Hotel Holiday Inn Moscow Vinogradovo, Moscow
Generating Cost Aware Covering Arrays For Free
Mustafa Kemal Tas, Hanefi Mercan, Gülşen Demiröz, Kamer Kaya, Cemal Yilmaz, Sabanci University
For video follow the link: https://youtu.be/Wkdd4A0rRjE
Would like to know more?
Visit our website:
www.tmpaconf.org
www.exactprosystems.com/events/tmpa
Follow us:
https://www.linkedin.com/company/exactpro-systems-llc?trk=biz-companies-cym
https://twitter.com/exactpro
Bayesian Inference and Uncertainty Quantification for Inverse ProblemsMatt Moores
So-called “inverse” problems arise when the parameters of a physical system cannot be directly observed. The mapping between these latent parameters and the space of noisy observations is represented as a mathematical model, often involving a system of differential equations. We seek to infer the parameter values that best fit our observed data. However, it is also vital to obtain accurate quantification of the uncertainty involved with these parameters, particularly when the output of the model will be used for forecasting. Bayesian inference provides well-calibrated uncertainty estimates, represented by the posterior distribution over the parameters. In this talk, I will give a brief introduction to Markov chain Monte Carlo (MCMC) algorithms for sampling from the posterior distribution and describe how they can be combined with numerical solvers for the forward model. We apply these methods to two examples of ODE models: growth curves in ecology, and thermogravimetric analysis (TGA) in chemistry. This is joint work with Matthew Berry, Mark Nelson, Brian Monaghan and Raymond Longbottom.
bayesImageS: Bayesian computation for medical Image Segmentation using a hidd...Matt Moores
There are many approaches to Bayesian computation with intractable likelihoods, including the exchange algorithm, approximate Bayesian computation (ABC), thermodynamic integration, and composite likelihood. These approaches vary in accuracy as well as scalability for datasets of significant size. The Potts model is an example where such methods are required, due to its intractable normalising constant. This model is a type of Markov random field, which is commonly used for image segmentation. The dimension of its parameter space increases linearly with the number of pixels in the image, making this a challenging application for scalable Bayesian computation. My talk will introduce various algorithms in the context of the Potts model and describe their implementation in C++, using OpenMP for parallelism.
R package 'bayesImageS': a case study in Bayesian computation using Rcpp and ...Matt Moores
There are many approaches to Bayesian computation with intractable likelihoods, including the exchange algorithm, approximate Bayesian computation (ABC), thermodynamic integration, and composite likelihood. These approaches vary in accuracy as well as scalability for datasets of significant size. The Potts model is an example where such methods are required, due to its intractable normalising constant. This model is a type of Markov random field, which is commonly used for image segmentation. The dimension of its parameter space increases linearly with the number of pixels in the image, making this a challenging application for scalable Bayesian computation. My talk will introduce various algorithms in the context of the Potts model and describe their implementation in C++, using OpenMP for parallelism. I will also discuss the process of releasing this software as an open source R package on the CRAN repository.
This is concerned with designing an exact exponential time algorithm that is better than the well-known 2^n algorithm for the problem Path Contraction. This answers an open question of van't Hof et. al [TCS 2009]. This is based on the article that appeared in ICALP 2019.
A new transformation into State Transition Algorithm for finding the global m...Michael_Chou
To promote the global search ability of the original state transition algorithm, a new operator called axesion is suggested, which aims to search along the axes and strengthen single dimensional search. Several benchmark minimization
problems are used to illustrate the advantages of the improved algorithm over other random search methods. The results of
numerical experiments show that the new transformation can enhance the performance of the state transition algorithm and the new strategy is effective and reliable.
new optimization algorithm for topology optimizationSeonho Park
authors devise new convex approximation called DQA which utilizes information of two consecutive points at iterates. Also, to guarantee global convergence, filter method is illustrated.
We approach the screening problem - i.e. detecting which inputs of a computer model significantly impact the output - from a formal Bayesian model selection point of view. That is, we place a Gaussian process prior on the computer model and consider the $2^p$ models that result from assuming that each of the subsets of the $p$ inputs affect the response. The goal is to obtain the posterior probabilities of each of these models. In this talk, we focus on the specification of objective priors on the model-specific parameters and on convenient ways to compute the associated marginal likelihoods. These two problems that normally are seen as unrelated, have challenging connections since the priors proposed in the literature are specifically designed to have posterior modes in the boundary of the parameter space, hence precluding the application of approximate integration techniques based on e.g. Laplace approximations. We explore several ways of circumventing this difficulty, comparing different methodologies with synthetic examples taken from the literature.
Authors: Gonzalo Garcia-Donato (Universidad de Castilla-La Mancha) and Rui Paulo (Universidade de Lisboa)
Nec 602 unit ii Random Variables and Random processDr Naim R Kidwai
The presentation explains concept of Probability, random variable, statistical averages, correlation, sum of random Variables, Central Limit Theorem,
random process, classification of random processes, power spectral density, multiple random processes.
We are interested in finding a permutation of the entries of a given square matrix so that the maximum number of its nonzero entries are moved to one of the corners in a L-shaped fashion.
If we interpret the nonzero entries of the matrix as the edges of a graph, this problem boils down to the so-called core–periphery structure, consisting of two sets: the core, a set of nodes that is highly connected across the whole graph, and the periphery, a set of nodes that is well connected only to the nodes that are in the core.
Matrix reordering problems have applications in sparse factorizations and preconditioning, while revealing core–periphery structures in networks has applications in economic, social and communication networks.
Abstract: An enhanced hybrid approach to OWL query answering that combines an RDF triple-store with an OWL reasoner in order to provide scaleable pay-as-you-go performance. The enhancements presented here include an extension to deal with arbitary OWL ontologies and optimisations that significantly improve scalability. We have implemented these techniques in a prototype system, a preliminary evaluation of which has produced very encouraging results.
Robust Image Denoising in RKHS via Orthogonal Matching PursuitPantelis Bouboulis
We present a robust method for the image denoising task based on kernel ridge regression and sparse modeling. Added noise is assumed to consist of two parts. One part is impulse noise assumed to be sparse (outliers), while the other part is bounded noise. The noisy image is divided into small regions of interest, whose pixels are regarded as points of a two-dimensional surface. A kernel based ridge regression method, whose parameters are selected adaptively, is employed to fit the data, whereas the outliers are detected via the use of the increasingly popular orthogonal matching pursuit (OMP) algorithm. To this end, a new variant of the OMP rationale is employed that has the additional advantage to automatically terminate, when all outliers have been selected.
Machine Learning and Counterfactual Reasoning for "Personalized" Decision- ...MLReview
Suchi Saria
Assistant Professor
Computer Science, Applied Math & Stats and Health Policy
Institute for Computational Medicine
Hossein Soleimani
Postdoctoral Fellow Computer Science
2017 Tutorial - Deep Learning for Dialogue SystemsMLReview
In the past decade, goal-oriented spoken dialogue systems (SDS) have been the most promi-nent component in today’s virtual personal assistants (VPAs). Among these VPAs, Microsoft’s Cortana, Apple’s Siri, Amazon Alexa, Google Assistant, and Facebook’s M, have incorporated SDS modules in various devices, which allow users to speak naturally in order to finish tasks more efficiently. The traditional conversational systems have rather complex and/or modular pipelines. The advance of deep learning technologies has recently risen the applicatins of neural models to dialogue modeling. Nevertheless, applying deep learning technologies for building robust and scalable dialogue systems is still a challenging task and an open research area as it requires deeper understanding of the classic pipelines as well as detailed knowledge on the benchmark of the models of the prior work and the recent state-of-the-art work. Thus, this tutorial is designed to focus on an overview of the dialogue system development while describing most recent research for building dialogue systems, and summarizing the challenges. We target an audience of students and practitioners who have some deep learning background and want to get more familiar with conversational dialog systems.
Tutorial on Theory and Application of Generative Adversarial NetworksMLReview
Description
Generative adversarial network (GAN) has recently emerged as a promising generative modeling approach. It consists of a generative network and a discriminative network. Through the competition between the two networks, it learns to model the data distribution. In addition to modeling the image/video distribution in computer vision problems, the framework finds use in defining visual concept using examples. To a large extent, it eliminates the need of hand-crafting objective functions for various computer vision problems. In this tutorial, we will present an overview of generative adversarial network research. We will cover several recent theoretical studies as well as training techniques and will also cover several vision applications of generative adversarial networks.
The use of Nauplii and metanauplii artemia in aquaculture (brine shrimp).pptxMAGOTI ERNEST
Although Artemia has been known to man for centuries, its use as a food for the culture of larval organisms apparently began only in the 1930s, when several investigators found that it made an excellent food for newly hatched fish larvae (Litvinenko et al., 2023). As aquaculture developed in the 1960s and ‘70s, the use of Artemia also became more widespread, due both to its convenience and to its nutritional value for larval organisms (Arenas-Pardo et al., 2024). The fact that Artemia dormant cysts can be stored for long periods in cans, and then used as an off-the-shelf food requiring only 24 h of incubation makes them the most convenient, least labor-intensive, live food available for aquaculture (Sorgeloos & Roubach, 2021). The nutritional value of Artemia, especially for marine organisms, is not constant, but varies both geographically and temporally. During the last decade, however, both the causes of Artemia nutritional variability and methods to improve poorquality Artemia have been identified (Loufi et al., 2024).
Brine shrimp (Artemia spp.) are used in marine aquaculture worldwide. Annually, more than 2,000 metric tons of dry cysts are used for cultivation of fish, crustacean, and shellfish larva. Brine shrimp are important to aquaculture because newly hatched brine shrimp nauplii (larvae) provide a food source for many fish fry (Mozanzadeh et al., 2021). Culture and harvesting of brine shrimp eggs represents another aspect of the aquaculture industry. Nauplii and metanauplii of Artemia, commonly known as brine shrimp, play a crucial role in aquaculture due to their nutritional value and suitability as live feed for many aquatic species, particularly in larval stages (Sorgeloos & Roubach, 2021).
Professional air quality monitoring systems provide immediate, on-site data for analysis, compliance, and decision-making.
Monitor common gases, weather parameters, particulates.
Seminar of U.V. Spectroscopy by SAMIR PANDASAMIR PANDA
Spectroscopy is a branch of science dealing the study of interaction of electromagnetic radiation with matter.
Ultraviolet-visible spectroscopy refers to absorption spectroscopy or reflect spectroscopy in the UV-VIS spectral region.
Ultraviolet-visible spectroscopy is an analytical method that can measure the amount of light received by the analyte.
ESR spectroscopy in liquid food and beverages.pptxPRIYANKA PATEL
With increasing population, people need to rely on packaged food stuffs. Packaging of food materials requires the preservation of food. There are various methods for the treatment of food to preserve them and irradiation treatment of food is one of them. It is the most common and the most harmless method for the food preservation as it does not alter the necessary micronutrients of food materials. Although irradiated food doesn’t cause any harm to the human health but still the quality assessment of food is required to provide consumers with necessary information about the food. ESR spectroscopy is the most sophisticated way to investigate the quality of the food and the free radicals induced during the processing of the food. ESR spin trapping technique is useful for the detection of highly unstable radicals in the food. The antioxidant capability of liquid food and beverages in mainly performed by spin trapping technique.
Toxic effects of heavy metals : Lead and Arsenicsanjana502982
Heavy metals are naturally occuring metallic chemical elements that have relatively high density, and are toxic at even low concentrations. All toxic metals are termed as heavy metals irrespective of their atomic mass and density, eg. arsenic, lead, mercury, cadmium, thallium, chromium, etc.
Salas, V. (2024) "John of St. Thomas (Poinsot) on the Science of Sacred Theol...Studia Poinsotiana
I Introduction
II Subalternation and Theology
III Theology and Dogmatic Declarations
IV The Mixed Principles of Theology
V Virtual Revelation: The Unity of Theology
VI Theology as a Natural Science
VII Theology’s Certitude
VIII Conclusion
Notes
Bibliography
All the contents are fully attributable to the author, Doctor Victor Salas. Should you wish to get this text republished, get in touch with the author or the editorial committee of the Studia Poinsotiana. Insofar as possible, we will be happy to broker your contact.
Earliest Galaxies in the JADES Origins Field: Luminosity Function and Cosmic ...Sérgio Sacani
We characterize the earliest galaxy population in the JADES Origins Field (JOF), the deepest
imaging field observed with JWST. We make use of the ancillary Hubble optical images (5 filters
spanning 0.4−0.9µm) and novel JWST images with 14 filters spanning 0.8−5µm, including 7 mediumband filters, and reaching total exposure times of up to 46 hours per filter. We combine all our data
at > 2.3µm to construct an ultradeep image, reaching as deep as ≈ 31.4 AB mag in the stack and
30.3-31.0 AB mag (5σ, r = 0.1” circular aperture) in individual filters. We measure photometric
redshifts and use robust selection criteria to identify a sample of eight galaxy candidates at redshifts
z = 11.5 − 15. These objects show compact half-light radii of R1/2 ∼ 50 − 200pc, stellar masses of
M⋆ ∼ 107−108M⊙, and star-formation rates of SFR ∼ 0.1−1 M⊙ yr−1
. Our search finds no candidates
at 15 < z < 20, placing upper limits at these redshifts. We develop a forward modeling approach to
infer the properties of the evolving luminosity function without binning in redshift or luminosity that
marginalizes over the photometric redshift uncertainty of our candidate galaxies and incorporates the
impact of non-detections. We find a z = 12 luminosity function in good agreement with prior results,
and that the luminosity function normalization and UV luminosity density decline by a factor of ∼ 2.5
from z = 12 to z = 14. We discuss the possible implications of our results in the context of theoretical
models for evolution of the dark matter halo mass function.
Nutraceutical market, scope and growth: Herbal drug technologyLokesh Patil
As consumer awareness of health and wellness rises, the nutraceutical market—which includes goods like functional meals, drinks, and dietary supplements that provide health advantages beyond basic nutrition—is growing significantly. As healthcare expenses rise, the population ages, and people want natural and preventative health solutions more and more, this industry is increasing quickly. Further driving market expansion are product formulation innovations and the use of cutting-edge technology for customized nutrition. With its worldwide reach, the nutraceutical industry is expected to keep growing and provide significant chances for research and investment in a number of categories, including vitamins, minerals, probiotics, and herbal supplements.
This presentation explores a brief idea about the structural and functional attributes of nucleotides, the structure and function of genetic materials along with the impact of UV rays and pH upon them.
Deep Behavioral Phenotyping in Systems Neuroscience for Functional Atlasing a...Ana Luísa Pinho
Functional Magnetic Resonance Imaging (fMRI) provides means to characterize brain activations in response to behavior. However, cognitive neuroscience has been limited to group-level effects referring to the performance of specific tasks. To obtain the functional profile of elementary cognitive mechanisms, the combination of brain responses to many tasks is required. Yet, to date, both structural atlases and parcellation-based activations do not fully account for cognitive function and still present several limitations. Further, they do not adapt overall to individual characteristics. In this talk, I will give an account of deep-behavioral phenotyping strategies, namely data-driven methods in large task-fMRI datasets, to optimize functional brain-data collection and improve inference of effects-of-interest related to mental processes. Key to this approach is the employment of fast multi-functional paradigms rich on features that can be well parametrized and, consequently, facilitate the creation of psycho-physiological constructs to be modelled with imaging data. Particular emphasis will be given to music stimuli when studying high-order cognitive mechanisms, due to their ecological nature and quality to enable complex behavior compounded by discrete entities. I will also discuss how deep-behavioral phenotyping and individualized models applied to neuroimaging data can better account for the subject-specific organization of domain-general cognitive systems in the human brain. Finally, the accumulation of functional brain signatures brings the possibility to clarify relationships among tasks and create a univocal link between brain systems and mental functions through: (1) the development of ontologies proposing an organization of cognitive processes; and (2) brain-network taxonomies describing functional specialization. To this end, tools to improve commensurability in cognitive science are necessary, such as public repositories, ontology-based platforms and automated meta-analysis tools. I will thus discuss some brain-atlasing resources currently under development, and their applicability in cognitive as well as clinical neuroscience.
What is greenhouse gasses and how many gasses are there to affect the Earth.moosaasad1975
What are greenhouse gasses how they affect the earth and its environment what is the future of the environment and earth how the weather and the climate effects.
hematic appreciation test is a psychological assessment tool used to measure an individual's appreciation and understanding of specific themes or topics. This test helps to evaluate an individual's ability to connect different ideas and concepts within a given theme, as well as their overall comprehension and interpretation skills. The results of the test can provide valuable insights into an individual's cognitive abilities, creativity, and critical thinking skills
DERIVATION OF MODIFIED BERNOULLI EQUATION WITH VISCOUS EFFECTS AND TERMINAL V...Wasswaderrick3
In this book, we use conservation of energy techniques on a fluid element to derive the Modified Bernoulli equation of flow with viscous or friction effects. We derive the general equation of flow/ velocity and then from this we derive the Pouiselle flow equation, the transition flow equation and the turbulent flow equation. In the situations where there are no viscous effects , the equation reduces to the Bernoulli equation. From experimental results, we are able to include other terms in the Bernoulli equation. We also look at cases where pressure gradients exist. We use the Modified Bernoulli equation to derive equations of flow rate for pipes of different cross sectional areas connected together. We also extend our techniques of energy conservation to a sphere falling in a viscous medium under the effect of gravity. We demonstrate Stokes equation of terminal velocity and turbulent flow equation. We look at a way of calculating the time taken for a body to fall in a viscous medium. We also look at the general equation of terminal velocity.
3. Statistical model criticism
MMD@P;QA a kf £k2 a supkf kF 1‘EQ f Epf “
-4 -2 2 4
-0.3
-0.2
-0.1
0.1
0.2
0.3
0.4
p(x)
q(x)
f *
(x)
f £@xA is the witness function
Can we compute MMD with samples from Q and a model P?
Problem: usualy can’t compute Epf in closed form.
3/52
4. Stein idea
To get rid of Epf in
sup
kf kF 1
‘Eq f Epf “
we define the Stein operator
‘Tpf “@xA a 1
p@xA
d
dx
@f @xAp@xAA
Then
EP TP f a 0
subject to appropriate boundary conditions. (Oates, Girolami, Chopin, 2016)
4/52
5. Stein idea: proof
Ep ‘Tpf “ a
Z
1
p@xA
d
dx
@f @xAp@xAA
p@xAdx
Z
d
dx
@f @xAp@xAA
dx
a ‘f @xAp@xA“I
I
a 0
5/52
6. Stein idea: proof
Ep ‘Tpf “ a
Z
1
¨¨¨p@xA
d
dx
@f @xAp@xAA
¨
¨¨p@xAdx
Z
d
dx
@f @xAp@xAA
dx
a ‘f @xAp@xA“I
I
a 0
5/52
7. Stein idea: proof
Ep ‘Tpf “ a
Z
1
¨¨¨p@xA
d
dx
@f @xAp@xAA
¨
¨¨p@xAdx
Z
d
dx
@f @xAp@xAA
dx
a ‘f @xAp@xA“I
I
a 0
5/52
8. Stein idea: proof
Ep ‘Tpf “ a
Z
1
¨¨¨p@xA
d
dx
@f @xAp@xAA
¨
¨¨p@xAdx
Z
d
dx
@f @xAp@xAA
dx
a ‘f @xAp@xA“I
I
a 0
5/52
9. Stein idea: proof
Ep ‘Tpf “ a
Z
1
¨¨¨p@xA
d
dx
@f @xAp@xAA
¨
¨¨p@xAdx
Z
d
dx
@f @xAp@xAA
dx
a ‘f @xAp@xA“I
I
a 0
5/52
10. Kernel Stein Discrepancy
Stein operator
Tpf a @x f Cf @x @log pA
Kernel Stein Discrepancy (KSD)
KSD@p;q;pA a sup
kgkF 1
Eq Tpg EpTpg
6/52
11. Kernel Stein Discrepancy
Stein operator
Tpf a @x f Cf @x @log pA
Kernel Stein Discrepancy (KSD)
KSD@p;q;pA a sup
kgkF 1
Eq Tpg $$$$EpTpg a sup
kgkF 1
Eq Tpg
6/52
12. Kernel Stein Discrepancy
Stein operator
Tpf a @x f Cf @x @log pA
Kernel Stein Discrepancy (KSD)
KSD@p;q;pA a sup
kgkF 1
Eq Tpg $$$$EpTpg a sup
kgkF 1
Eq Tpg
-4 -2 2 4
-0.6
-0.4
-0.2
0.2
0.4
p(x)
q(x)
g*
(x)
6/52
13. Kernel Stein Discrepancy
Stein operator
Tpf a @x f Cf @x @log pA
Kernel Stein Discrepancy (KSD)
KSD@p;q;pA a sup
kgkF 1
Eq Tpg $$$$EpTpg a sup
kgkF 1
Eq Tpg
-4 -2 2 4
0.1
0.2
0.3
0.4
p(x)
q(x)
g*
(x)
6/52
14. Kernel stein discrepancy
Closed-form expression for KSD: given Z;ZH $ q, then
(Chwialkowski, Strathmann, G., ICML 2016) (Liu, Lee, Jordan ICML 2016)
KSD@p;q;pA a Eq hp@Z;ZHA
where
hp@x;yA Xa @x log p@xA@x log p@yAk@x;yA
C@y log p@yA@x k@x;yA
C@x log p@xA@y k@x;yA
C@x @y k@x;yA
and k is RKHS kernel for p
Only depends on kernel and @x log p(x). Do not need to
normalize p, or sample from it.
7/52
19. Statistical model criticism
Chicago crime data
Model is Gaussian mixture with ten components
Stein witness function
Code: https://github.com/karlnapf/kernel_goodness_of_fit 8/52
20. Kernel stein discrepancy
Further applications:
Evaluation of approximate MCMC methods.
(Chwialkowski, Strathmann, G., ICML 2016; Gorham, Mackey, ICML 2017)
What kernel to use?
The inverse multiquadric kernel,
k@x;yA a
c Ckx yk2
2
24. Dependence testing
Given: Samples from a distribution PX Y
Goal: Are X and Y independent?
Their noses guide them
through life, and they're
never happier than when
following an interesting scent.
A large animal who slings slobber,
exudes a distinctive houndy odor,
and wants nothing more than to
follow his nose.
Text from dogtime.com and petfinder.com
A responsive, interactive
pet, one that will blow in
your ear and follow you
everywhere.
YX
11/52
25. MMD as a dependence measure?
Could we use MMD?
MMD@PXY
| {z }
P
;PX PY
| {z }
Q
;rA
We don’t have samples from Q Xa PX PY , only pairs
f@xi ;yi gn
i=1
i:i:d:
$ PXY
Solution: simulate Q with pairs (xi ;yj ) for j T= i.
What kernel to use for the RKHS r?
12/52
26. MMD as a dependence measure?
Could we use MMD?
MMD@PXY
| {z }
P
;PX PY
| {z }
Q
;rA
We don’t have samples from Q Xa PX PY , only pairs
f@xi ;yi gn
i=1
i:i:d:
$ PXY
Solution: simulate Q with pairs (xi ;yj ) for j T= i.
What kernel to use for the RKHS r?
12/52
27. MMD as a dependence measure?
Could we use MMD?
MMD@PXY
| {z }
P
;PX PY
| {z }
Q
;rA
We don’t have samples from Q Xa PX PY , only pairs
f@xi ;yi gn
i=1
i:i:d:
$ PXY
Solution: simulate Q with pairs (xi ;yj ) for j T= i.
What kernel to use for the RKHS r?
12/52
28. MMD as a dependence measure
Kernel k on images with feature space p,
Kernel l on captions with feature space q,
13/52
29. MMD as a dependence measure
Kernel k on images with feature space p,
Kernel l on captions with feature space q,
Kernel on image-text pairs: are images and captions similar?
13/52
30. MMD as a dependence measure
Given: Samples from a distribution PX Y
Goal: Are X and Y independent?
MMD2
@bPXY ; bPX
bPY ;rA Xa 1
n2
trace@KLA
( K, L column centered)
14/52
31. MMD as a dependence measure
Given: Samples from a distribution PX Y
Goal: Are X and Y independent?
MMD2
@bPXY ; bPX
bPY ;rA Xa 1
n2
trace@KLA
14/52
32. MMD as a dependence measure
Two questions:
Why the product kernel? Many ways to combine kernels - why not
eg a sum?
Is there a more interpretable way of defining this dependence
measure?
15/52
33. Finding covariance with smooth transformations
Illustration: two variables with no correlation but strong dependence.
-2 -1 0 1 2
-1.5
-1
-0.5
0
0.5
1
1.5
Correlation: 0.00
16/52
34. Finding covariance with smooth transformations
Illustration: two variables with no correlation but strong dependence.
-2 -1 0 1 2
-1.5
-1
-0.5
0
0.5
1
1.5
Correlation: 0.00
-2 0 2
-1
-0.5
0
0.5
-2 0 2
-1
-0.5
0
0.5
16/52
36. Define two spaces, one for each witness
Function in p
f @xA a
IX
j =1
fj 'j @xA
Feature map
Kernel for RKHS p on ˆ:
k@x;xHA a h'@xA;'@xHAip
Function in q
g@yA a
IX
j =1
gj j @yA
Feature map
Kernel for RKHS q on ‰:
l@x;xHA a h@yA;@yHAiq
17/52
38. The constrained covariance
The constrained covariance is
COCO@PXY A a sup
kf kp 1
kgkq 1
cov
2
4
0
@
IX
j =1
fj 'j @xA
1
A
0
@
IX
j =1
gj j @yA
1
A
3
5
18/52
39. The constrained covariance
The constrained covariance is
COCO@PXY A a sup
kf kp 1
kgkq 1
Exy
2
4
0
@
IX
j =1
fj 'j @xA
1
A
0
@
IX
j =1
gj j @yA
1
A
3
5
Fine print: feature mappings '(x) and (y) assumed to have zero mean.
18/52
40. The constrained covariance
The constrained covariance is
COCO@PXY A a sup
kf kp 1
kgkq 1
Exy
2
4
0
@
IX
j =1
fj 'j @xA
1
A
0
@
IX
j =1
gj j @yA
1
A
3
5
Fine print: feature mappings '(x) and (y) assumed to have zero mean.
Rewriting:
Exy ‘f @xAg@yA“
a
2
6
6
4
f1
f2
...
3
7
7
5
b
Exy
0
B
B
@
2
6
6
4
'1@xA
'2@xA
...
3
7
7
5
h
1@yA 2@yA :::
i
1
C
C
A
| {z }
C'(x)(y)
2
6
6
4
g1
g2
...
3
7
7
5
18/52
41. The constrained covariance
The constrained covariance is
COCO@PXY A a sup
kf kp 1
kgkq 1
Exy
2
4
0
@
IX
j =1
fj 'j @xA
1
A
0
@
IX
j =1
gj j @yA
1
A
3
5
Fine print: feature mappings '(x) and (y) assumed to have zero mean.
Rewriting:
Exy ‘f @xAg@yA“
a
2
6
6
4
f1
f2
...
3
7
7
5
b
Exy
0
B
B
@
2
6
6
4
'1@xA
'2@xA
...
3
7
7
5
h
1@yA 2@yA :::
i
1
C
C
A
| {z }
C'(x)(y)
2
6
6
4
g1
g2
...
3
7
7
5
COCO: max singular value of feature covariance C'(x)(y)
18/52
42. Computing COCO in practice
Given sample f@xi ;yi Agn
i=1
i:i:d:
$ PXY , what is empirical COCO ?
19/52
43. Computing COCO in practice
Given sample f@xi ;yi Agn
i=1
i:i:d:
$ PXY , what is empirical COCO ?
COCO is largest eigenvalue
max of
0 1
n KL
1
n LK 0
#
45. #
:
Kij a k@xi ;xj A and Lij a l@yi ;yj A.
Fine print: kernels are computed with empirically centered features '(x) 1
n
Pn
i=1
'(xi )
and (y) 1
n
Pn
i=1
(yi ).
AG., A. Smola., O. Bousquet, R. Herbrich, A. Belitski, M. Augath, Y. Murayama, J. Pauls, B.
Schoelkopf, and N. Logothetis, AISTATS’05
19/52
46. Computing COCO in practice
Given sample f@xi ;yi Agn
i=1
i:i:d:
$ PXY , what is empirical COCO ?
COCO is largest eigenvalue
max of
0 1
n KL
1
n LK 0
#
48. #
:
Kij a k@xi ;xj A and Lij a l@yi ;yj A.
Witness functions (singular vectors):
f @xA G
mX
i=1
i k@xi ;xA g@yA G
nX
i=1
49. i l@yi ;yA
Fine print: kernels are computed with empirically centered features '(x) 1
n
Pn
i=1
'(xi )
and (y) 1
n
Pn
i=1
(yi ).
AG., A. Smola., O. Bousquet, R. Herbrich, A. Belitski, M. Augath, Y. Murayama, J. Pauls, B.
Schoelkopf, and N. Logothetis, AISTATS’05
19/52
50. What is a large dependence with COCO?
−2 0 2
−3
−2
−1
0
1
2
3
X
Y
Smooth density
−4 −2 0 2 4
−4
−2
0
2
4
X
Y
500 Samples, smooth density
−2 0 2
−3
−2
−1
0
1
2
3
X
Y
Rough density
−4 −2 0 2 4
−4
−2
0
2
4
X
Y
500 samples, rough density
Density takes the form:
PXY G 1Csin@!xAsin@!yA
Which of these is the more “dependent”?
20/52
57. Dependence largest when at “low” frequencies
As dependence is encoded at higher frequencies, the smooth
mappings f ;g achieve lower linear dependence.
Even for independent variables, COCO will not be zero at finite
sample sizes, since some mild linear dependence will be found by f,g
(bias)
This bias will decrease with increasing sample size.
27/52
58. Can we do better than COCO?
A second example with zero correlation.
First singular value of feature covariance C'(x)(y):
-1 -0.5 0 0.5 1
-1
-0.5
0
0.5
1
Correlation: 0.00
-2 0 2
-1
-0.5
0
0.5
-2 0 2
-1
-0.5
0
0.5
-1 -0.5 0 0.5
-0.8
-0.6
-0.4
-0.2
0
0.2
0.4
Correlation: 0.80 COCO1
: 0.11
28/52
59. Can we do better than COCO?
A second example with zero correlation.
Second singular value of feature covariance C'(x)(y):
-1 -0.5 0 0.5 1
-1
-0.5
0
0.5
1
Correlation: 0.00
-2 0 2
-1
-0.5
0
0.5
1
-2 0 2
-1
-0.5
0
0.5
1
28/52
60. Can we do better than COCO?
A second example with zero correlation.
Second singular value of feature covariance C'(x)(y):
-1 -0.5 0 0.5 1
-1
-0.5
0
0.5
1
Correlation: 0.00
-2 0 2
-1
-0.5
0
0.5
1
-2 0 2
-1
-0.5
0
0.5
1
-1 -0.5 0 0.5 1
-0.5
0
0.5
Correlation: 0.37 COCO2
: 0.06
28/52
61. The Hilbert-Schmidt Independence Criterion
Writing the ith singular value of the feature covariance C'(x)(y) as
i Xa COCOi @PXY Yp;qA;
define Hilbert-Schmidt Independence Criterion (HSIC)
HSIC2
@PXY Yp;qA a
IX
i=1
2
i :
AG, O. Bousquet , A. Smola., and B. Schoelkopf, ALT2005; AG,.,K. Fukumizu„C.H. Teo., L. Song., B.
Schoelkopf., and A. Smola, NIPS 2007,.
29/52
62. The Hilbert-Schmidt Independence Criterion
Writing the ith singular value of the feature covariance C'(x)(y) as
i Xa COCOi @PXY Yp;qA;
define Hilbert-Schmidt Independence Criterion (HSIC)
HSIC2
@PXY Yp;qA a
IX
i=1
2
i :
AG, O. Bousquet , A. Smola., and B. Schoelkopf, ALT2005; AG,.,K. Fukumizu„C.H. Teo., L. Song., B.
Schoelkopf., and A. Smola, NIPS 2007,.
HSIC is MMD with product kernel!
HSIC2
@PXY Yp;qA a MMD2
@PXY ;PX PY YrA
where @@x;yA;@xH;yHAA a k@x;xHAl@y;yHA.
29/52
63. Asymptotics of HSIC under independence
Given sample f@xi ;yi gn
i=1
i:i:d:
$ PXY , what is empirical HSIC?
Empirical HSIC (biased)
HSIC a 1
n2
trace@KLA
Kij a k@xi ;xj A and Lij a l@yi yj A (K and L computed with
empirically centered features)
Statistical testing: given PXY a PX PY , what is the threshold c
such that P@HSIC cA for small ?
Asymptotics of HSIC when PXY a PX PY :
n HSIC
D
3
IX
l=1
l z2
l ; zl $ x@0;1Ai:i:d:
where l l (zj ) =
R
hijqr l (zi )dFi;q;r ; hijqr = 1
4!
P(i;j ;q;r)
(t;u;v;w)
ktu ltu + ktu lvw 2ktu ltv
30/52
64. Asymptotics of HSIC under independence
Given sample f@xi ;yi gn
i=1
i:i:d:
$ PXY , what is empirical HSIC?
Empirical HSIC (biased)
HSIC a 1
n2
trace@KLA
Kij a k@xi ;xj A and Lij a l@yi yj A (K and L computed with
empirically centered features)
Statistical testing: given PXY a PX PY , what is the threshold c
such that P@HSIC cA for small ?
Asymptotics of HSIC when PXY a PX PY :
n HSIC
D
3
IX
l=1
l z2
l ; zl $ x@0;1Ai:i:d:
where l l (zj ) =
R
hijqr l (zi )dFi;q;r ; hijqr = 1
4!
P(i;j ;q;r)
(t;u;v;w)
ktu ltu + ktu lvw 2ktu ltv
30/52
65. Asymptotics of HSIC under independence
Given sample f@xi ;yi gn
i=1
i:i:d:
$ PXY , what is empirical HSIC?
Empirical HSIC (biased)
HSIC a 1
n2
trace@KLA
Kij a k@xi ;xj A and Lij a l@yi yj A (K and L computed with
empirically centered features)
Statistical testing: given PXY a PX PY , what is the threshold c
such that P@HSIC cA for small ?
Asymptotics of HSIC when PXY a PX PY :
n HSIC
D
3
IX
l=1
l z2
l ; zl $ x@0;1Ai:i:d:
where l l (zj ) =
R
hijqr l (zi )dFi;q;r ; hijqr = 1
4!
P(i;j ;q;r)
(t;u;v;w)
ktu ltu + ktu lvw 2ktu ltv
30/52
66. Asymptotics of HSIC under independence
Given sample f@xi ;yi gn
i=1
i:i:d:
$ PXY , what is empirical HSIC?
Empirical HSIC (biased)
HSIC a 1
n2
trace@KLA
Kij a k@xi ;xj A and Lij a l@yi yj A (K and L computed with
empirically centered features)
Statistical testing: given PXY a PX PY , what is the threshold c
such that P@HSIC cA for small ?
Asymptotics of HSIC when PXY a PX PY :
n HSIC
D
3
IX
l=1
l z2
l ; zl $ x@0;1Ai:i:d:
where l l (zj ) =
R
hijqr l (zi )dFi;q;r ; hijqr = 1
4!
P(i;j ;q;r)
(t;u;v;w)
ktu ltu + ktu lvw 2ktu ltv
30/52
67. A statistical test
Given PXY a PX PY , what is the threshold c such that
P@HSIC cA for small (prob. of false positive)?
Original time series:
X1 X2 X3 X4 X5 X6 X7 X8 X9 X10
Y1 Y2 Y3 Y4 Y5 Y6 Y7 Y8 Y9 Y10
Permutation:
X1 X2 X3 X4 X5 X6 X7 X8 X9 X10
Y7 Y3 Y9 Y2 Y4 Y8 Y5 Y1 Y6 Y10
Null distribution via permutation
Compute HSIC for fxi ;y(i)gn
i=1 for random permutation of indices
f1;:::;ng. This gives HSIC for independent variables.
Repeat for many different permutations, get empirical CDF
Threshold c is 1 quantile of empirical CDF 31/52
68. A statistical test
Given PXY a PX PY , what is the threshold c such that
P@HSIC cA for small (prob. of false positive)?
Original time series:
X1 X2 X3 X4 X5 X6 X7 X8 X9 X10
Y1 Y2 Y3 Y4 Y5 Y6 Y7 Y8 Y9 Y10
Permutation:
X1 X2 X3 X4 X5 X6 X7 X8 X9 X10
Y7 Y3 Y9 Y2 Y4 Y8 Y5 Y1 Y6 Y10
Null distribution via permutation
Compute HSIC for fxi ;y(i)gn
i=1 for random permutation of indices
f1;:::;ng. This gives HSIC for independent variables.
Repeat for many different permutations, get empirical CDF
Threshold c is 1 quantile of empirical CDF 31/52
69. A statistical test
Given PXY a PX PY , what is the threshold c such that
P@HSIC cA for small (prob. of false positive)?
Original time series:
X1 X2 X3 X4 X5 X6 X7 X8 X9 X10
Y1 Y2 Y3 Y4 Y5 Y6 Y7 Y8 Y9 Y10
Permutation:
X1 X2 X3 X4 X5 X6 X7 X8 X9 X10
Y7 Y3 Y9 Y2 Y4 Y8 Y5 Y1 Y6 Y10
Null distribution via permutation
Compute HSIC for fxi ;y(i)gn
i=1 for random permutation of indices
f1;:::;ng. This gives HSIC for independent variables.
Repeat for many different permutations, get empirical CDF
Threshold c is 1 quantile of empirical CDF 31/52
70. Application: dependence detection across languages
Testing task: detect dependence between English and French text
Les ordres de gouvernements
provinciaux et municipaux
subissent de fortes pressions
Honourable senators, I have a
question for the Leader of the
Government in the Senate
Text from the aligned hansards of the 36th parliament of canada,
https://www.isi.edu/natural-language/download/hansard/
YX
Honorables sénateurs, ma question
s’adresse au leader du
gouvernement au Sénat
Au contraire, nous avons augmenté
le financement fédéral pour le
développement des jeunes
No doubt there is great pressure
on provincial and municipal
governments
In fact, we have increased
federal investments for early
childhood development.
...
...
32/52
71. Application: dependence detection across languages
Testing task: detect dependence between English and French text
k-spectrum kernel, k a 10, sample size n a 10
HSIC a 1
n2
trace@KLA
(K and L column centered) 33/52
72. Application:Dependence detection across languages
Results (for a 0:05)
k-spectrum kernel: average Type II error 0
Bag of words kernel: average Type II error 0.18
Settings: Five line extracts, averaged over 300 repetitions, for
“Agriculture” transcripts. Similar results for Fisheries and
Immigration transcripts.
34/52
74. Detecting higher order interaction
How to detect V-structures with pairwise weak individual
dependence?
X Y
Z
36/52
75. Detecting higher order interaction
How to detect V-structures with pairwise weak individual
dependence?
36/52
76. Detecting higher order interaction
How to detect V-structures with pairwise weak individual
dependence?
X cc Y ;Y cc Z;X cc Z
X1 vs Y1 Y1 vs Z1
X1 vs Z1 X1*Y1 vs Z1
X Y
Z
X ;Y
i:i:d:
$ x@0;1A
Zj X ;Y $ sign@XY AExp@ 1p
2
A
Fine print: Faithfulness violated here!
36/52
77. V-structure discovery
X Y
Z
Assume X cc Y has been established.
V-structure can then be detected by:
Consistent CI test: H0 X X cc Y jZ [Fukumizu et al. 2008, Zhang et al. 2011]
Factorisation test: H0 X @X ;Y A cc Z • @X ;ZA cc Y • @Y ;ZA cc X
(multiple standard two-variable tests)
How well do these work?
37/52
78. Detecting higher order interaction
Generalise earlier example to p dimensions
X cc Y ;Y cc Z;X cc Z
X1 vs Y1 Y1 vs Z1
X1 vs Z1 X1*Y1 vs Z1
X Y
Z
X ;Y
i:i:d:
$ x@0;1A
Zj X ;Y $ sign@XY AExp@ 1p
2
A
X2:p;Y2:p;Z2:p
i:i:d:
$ x@0;Ip 1A
Fine print: Faithfulness violated here!
38/52
80. Lancaster interaction measure
Lancaster interaction measure of @X1;:::;XD A $ P is a signed
measure ¡P that vanishes whenever P can be factorised non-trivially.
D a 2 X ¡LP a PXY PX PY
40/52
81. Lancaster interaction measure
Lancaster interaction measure of @X1;:::;XD A $ P is a signed
measure ¡P that vanishes whenever P can be factorised non-trivially.
D a 2 X ¡LP a PXY PX PY
D a 3 X ¡LP a PXYZ PX PYZ PY PXZ PZ PXY C2PX PY PZ
40/52
82. Lancaster interaction measure
Lancaster interaction measure of @X1;:::;XD A $ P is a signed
measure ¡P that vanishes whenever P can be factorised non-trivially.
D a 2 X ¡LP a PXY PX PY
D a 3 X ¡LP a PXYZ PX PYZ PY PXZ PZ PXY C2PX PY PZ
X Y
Z
X Y
Z
X Y
Z
X Y
Z
PXY Z −PXPY Z −PY PXZ −PZPXY +2PXPY PZ
∆LP =
40/52
83. Lancaster interaction measure
Lancaster interaction measure of @X1;:::;XD A $ P is a signed
measure ¡P that vanishes whenever P can be factorised non-trivially.
D a 2 X ¡LP a PXY PX PY
D a 3 X ¡LP a PXYZ PX PYZ PY PXZ PZ PXY C2PX PY PZ
X Y
Z
X Y
Z
X Y
Z
X Y
Z
PXY Z −PXPY Z −PXZPY −PXY PZ +2PXPY PZ
∆LP = 0
Case of PX cc PYZ
40/52
84. Lancaster interaction measure
Lancaster interaction measure of @X1;:::;XD A $ P is a signed
measure ¡P that vanishes whenever P can be factorised non-trivially.
D a 2 X ¡LP a PXY PX PY
D a 3 X ¡LP a PXYZ PX PYZ PY PXZ PZ PXY C2PX PY PZ
@X ;Y A cc Z • @X ;ZA cc Y • @Y ;ZA cc X A ¡LP a 0:
...so what might be missed?
40/52
85. Lancaster interaction measure
Lancaster interaction measure of @X1;:::;XD A $ P is a signed
measure ¡P that vanishes whenever P can be factorised non-trivially.
D a 2 X ¡LP a PXY PX PY
D a 3 X ¡LP a PXYZ PX PYZ PY PXZ PZ PXY C2PX PY PZ
¡LP a 0 ;@X ;Y A cc Z • @X ;ZA cc Y • @Y ;ZA cc X
Example:
P(0;0;0) = 0:2 P(0;0;1) = 0:1 P(1;0;0) = 0:1 P(1;0;1) = 0:1
P(0;1;0) = 0:1 P(0;1;1) = 0:1 P(1;1;0) = 0:1 P(1;1;1) = 0:2
40/52
86. A kernel test statistic using Lancaster Measure
Construct a test by estimating k @¡LPAk2
r ; where a k l m:
k@PXYZ PXY PZ ¡¡¡Ak2
r a
hPXYZ ;PXYZ ir 2 hPXYZ ;PXY PZ ir ¡¡¡
41/52
87. A kernel test statistic using Lancaster Measure
Table: V -statistic estimators of h;Hir
(without terms PX PY PZ ). H
is centering matrix I n 1
Lancaster interaction statistic: D. Sejdinovic, AG, W. Bergsma, NIPS13
k @¡LPAk2
r a 1
n2
@HKH HLH HMHA++ :
Empirical joint central moment in the feature space
42/52
88. A kernel test statistic using Lancaster Measure
Table: V -statistic estimators of h;Hir
(without terms PX PY PZ ). H
is centering matrix I n 1
Lancaster interaction statistic: D. Sejdinovic, AG, W. Bergsma, NIPS13
k @¡LPAk2
r a 1
n2
@HKH HLH HMHA++ :
Empirical joint central moment in the feature space
42/52
90. Interaction for D 4
Interaction measure valid for all D:
(Streitberg, 1990)
¡S P a
X
@ 1Ajj 1
@jj 1A3JP
For a partition , J associates to the
joint the corresponding factorisation,
e.g., J13j2j4P = PX1X3
PX2
PX4
:
44/52
91. Interaction for D 4
Interaction measure valid for all D:
(Streitberg, 1990)
¡S P a
X
@ 1Ajj 1
@jj 1A3JP
For a partition , J associates to the
joint the corresponding factorisation,
e.g., J13j2j4P = PX1X3
PX2
PX4
:
44/52
92. Interaction for D 4
Interaction measure valid for all D:
(Streitberg, 1990)
¡S P a
X
@ 1Ajj 1
@jj 1A3JP
For a partition , J associates to the
joint the corresponding factorisation,
e.g., J13j2j4P = PX1X3
PX2
PX4
:
1e+04
1e+09
1e+14
1e+19
1 3 5 7 9 11 13 15 17 19 21 23 25
D
Numberofpartitionsof{1,...,D}
Bell numbers growth
44/52