The document discusses limitations of Shannon entropy and computable measures. It describes an algorithm for creating a simple graph by iteratively adding nodes to maximize the number of "core nodes". The degree sequence of labeled nodes forms the Champernowne constant in base 10. The sequence of number of edges is a function of core and supportive nodes. Maximum entropy and algorithmic complexity principles are discussed. The Coding Theorem Method is introduced for computing the uncomputable. Computability mediates challenges in causality via algorithmic information theory. The Block Decomposition Method joins Shannon entropy and algorithmic complexity by allowing a tighter upper bound on complexity than analyzing parts separately. Examples are given of applying these methods.
The document discusses the planted clique problem in graph theory. It introduces the problem and describes how previous research has found polynomial-time algorithms to solve the problem when the size of the planted clique k is O(√n). The document then summarizes two algorithms - Kucera's algorithm and the Low Degree Removal (LDR) algorithm - that have been used to approach the problem. It describes implementing the algorithms in a C++ program to simulate random graphs with planted cliques and test the ability of the algorithms to recover the planted clique.
The document describes various adaptive methods for numerical integration or cubature of functions, including Monte Carlo methods, low-discrepancy sampling, and Bayesian cubature. It discusses approaches to choose sample sizes and weights to guarantee the integral estimate is within a given tolerance of the true integral with high probability. Specific examples discussed include multidimensional Gaussian integrals and estimating Sobol' sensitivity indices.
Moment Preserving Approximation of Independent Components for the Reconstruct...rahulmonikasharma
The application of Independent Component Analysis (ICA) has found considerable success in problems where sets of observed time series may be considered as results of linearly mixed instantaneous source signals. The Independent Components (IC’s) or features can be used in the reconstruction of observed multivariate time seriesfollowing an optimal ordering process. For trend discovery and forecasting, the generated IC’s can be approximated for the purpose of noise removal and for the lossy compression of the signals.We propose a moment-preserving (MP) methodology for approximating IC’s for the reconstruction of multivariate time series.The methodologyis based on deriving the approximation in the signal domain while preserving a finite number of geometric moments in its Fourier domain.Experimental results are presented onthe approximation of both artificial time series and actual time series of currency exchange rates. Our results show that the moment-preserving (MP) approximations of time series are superior to other usual interpolation approximation methods, particularly when the signals contain significant noise components. The results also indicate that the present MP approximations have significantly higher reconstruction accuracy and can be used successfully for signal denoising while achieving in the same time high packing ratios. Moreover, we find that quite acceptable reconstructions of observed multivariate time series can be obtained with only the first few MP approximated IC’s.
In this talk we will describe a methodology to handle the causality to make inference on common-cause failure in a situation of missing data. The data are collected in the form of contingency table but the available information are only the numbers of CCF of different orders and the numbers of failure due to a given cause. Therefore only the margins of the contingency table are observed; thefrequencies in each cell are unknown. Assuming a Poisson model for the count, we suggest a Bayesian approach and we use the inverse Bayes formula (IBF) combined with a Metropolis-Hastings algorithm to make inference on the rate of occurrence for the different combination cause, order. The performance of the resulting algorithm is evaluated through simulations. A comparison is made with results obtained from the _-composition approach to deal with causality suggested by Zheng et al. (2013).
This document discusses challenges and recent advances in Approximate Bayesian Computation (ABC) methods. ABC methods are used when the likelihood function is intractable or unavailable in closed form. The core ABC algorithm involves simulating parameters from the prior and simulating data, retaining simulations where the simulated and observed data are close according to a distance measure on summary statistics. The document outlines key issues like scalability to large datasets, assessment of uncertainty, and model choice, and discusses advances such as modified proposals, nonparametric methods, and perspectives that include summary construction in the framework. Validation of ABC model choice and selection of summary statistics remains an open challenge.
The document proposes using random forests (RF), a machine learning tool, for approximate Bayesian computation (ABC) model choice rather than estimating model posterior probabilities. RF improves on existing ABC model choice methods by having greater discriminative power among models, being robust to the choice and number of summary statistics, requiring less computation, and providing an error rate to evaluate confidence in the model choice. The authors illustrate the power of the RF-based ABC methodology on controlled experiments and real population genetics datasets.
The document discusses the planted clique problem in graph theory. It introduces the problem and describes how previous research has found polynomial-time algorithms to solve the problem when the size of the planted clique k is O(√n). The document then summarizes two algorithms - Kucera's algorithm and the Low Degree Removal (LDR) algorithm - that have been used to approach the problem. It describes implementing the algorithms in a C++ program to simulate random graphs with planted cliques and test the ability of the algorithms to recover the planted clique.
The document describes various adaptive methods for numerical integration or cubature of functions, including Monte Carlo methods, low-discrepancy sampling, and Bayesian cubature. It discusses approaches to choose sample sizes and weights to guarantee the integral estimate is within a given tolerance of the true integral with high probability. Specific examples discussed include multidimensional Gaussian integrals and estimating Sobol' sensitivity indices.
Moment Preserving Approximation of Independent Components for the Reconstruct...rahulmonikasharma
The application of Independent Component Analysis (ICA) has found considerable success in problems where sets of observed time series may be considered as results of linearly mixed instantaneous source signals. The Independent Components (IC’s) or features can be used in the reconstruction of observed multivariate time seriesfollowing an optimal ordering process. For trend discovery and forecasting, the generated IC’s can be approximated for the purpose of noise removal and for the lossy compression of the signals.We propose a moment-preserving (MP) methodology for approximating IC’s for the reconstruction of multivariate time series.The methodologyis based on deriving the approximation in the signal domain while preserving a finite number of geometric moments in its Fourier domain.Experimental results are presented onthe approximation of both artificial time series and actual time series of currency exchange rates. Our results show that the moment-preserving (MP) approximations of time series are superior to other usual interpolation approximation methods, particularly when the signals contain significant noise components. The results also indicate that the present MP approximations have significantly higher reconstruction accuracy and can be used successfully for signal denoising while achieving in the same time high packing ratios. Moreover, we find that quite acceptable reconstructions of observed multivariate time series can be obtained with only the first few MP approximated IC’s.
In this talk we will describe a methodology to handle the causality to make inference on common-cause failure in a situation of missing data. The data are collected in the form of contingency table but the available information are only the numbers of CCF of different orders and the numbers of failure due to a given cause. Therefore only the margins of the contingency table are observed; thefrequencies in each cell are unknown. Assuming a Poisson model for the count, we suggest a Bayesian approach and we use the inverse Bayes formula (IBF) combined with a Metropolis-Hastings algorithm to make inference on the rate of occurrence for the different combination cause, order. The performance of the resulting algorithm is evaluated through simulations. A comparison is made with results obtained from the _-composition approach to deal with causality suggested by Zheng et al. (2013).
This document discusses challenges and recent advances in Approximate Bayesian Computation (ABC) methods. ABC methods are used when the likelihood function is intractable or unavailable in closed form. The core ABC algorithm involves simulating parameters from the prior and simulating data, retaining simulations where the simulated and observed data are close according to a distance measure on summary statistics. The document outlines key issues like scalability to large datasets, assessment of uncertainty, and model choice, and discusses advances such as modified proposals, nonparametric methods, and perspectives that include summary construction in the framework. Validation of ABC model choice and selection of summary statistics remains an open challenge.
The document proposes using random forests (RF), a machine learning tool, for approximate Bayesian computation (ABC) model choice rather than estimating model posterior probabilities. RF improves on existing ABC model choice methods by having greater discriminative power among models, being robust to the choice and number of summary statistics, requiring less computation, and providing an error rate to evaluate confidence in the model choice. The authors illustrate the power of the RF-based ABC methodology on controlled experiments and real population genetics datasets.
The International Journal of Engineering & Science is aimed at providing a platform for researchers, engineers, scientists, or educators to publish their original research results, to exchange new ideas, to disseminate information in innovative designs, engineering experiences and technological skills. It is also the Journal's objective to promote engineering and technology education. All papers submitted to the Journal will be blind peer-reviewed. Only original articles will be published.
The papers for publication in The International Journal of Engineering& Science are selected through rigorous peer reviews to ensure originality, timeliness, relevance, and readability.
The document discusses different perspectives on simulating the mean of a function, including deterministic, randomized, and Bayesian approaches. It summarizes Monte Carlo methods using the central limit theorem and Berry-Esseen inequality to estimate error bounds. Low-discrepancy sampling and cubature methods are described which use Fourier coefficients to bound integration errors. Bayesian cubature is outlined, which assumes the function is drawn from a Gaussian process prior to perform optimal quadrature. Maximum likelihood is used to estimate the kernel hyperparameters.
This document summarizes and compares two popular Python libraries for graph neural networks - Spektral and PyTorch Geometric. It begins by providing an overview of the basic functionality and architecture of each library. It then discusses how each library handles data loading and mini-batching of graph data. The document reviews several common message passing layer types implemented in both libraries. It provides an example comparison of using each library for a node classification task on the Cora dataset. Finally, it discusses a graph classification comparison in PyTorch Geometric using different message passing and pooling layers on the IMDB-binary dataset.
Since the advent of the horseshoe priors for regularization, global-local shrinkage methods have proved to be a fertile ground for the development of Bayesian theory and methodology in machine learning. They have achieved remarkable success in computation, and enjoy strong theoretical support. Much of the existing literature has focused on the linear Gaussian case. The purpose of the current talk is to demonstrate that the horseshoe priors are useful more broadly, by reviewing both methodological and computational developments in complex models that are more relevant to machine learning applications. Specifically, we focus on methodological challenges in horseshoe regularization in nonlinear and non-Gaussian models; multivariate models; and deep neural networks. We also outline the recent computational developments in horseshoe shrinkage for complex models along with a list of available software implementations that allows one to venture out beyond the comfort zone of the canonical linear regression problems.
This document summarizes a research paper about using hierarchical deterministic quadrature methods for option pricing under the rough Bergomi model. It discusses the rough Bergomi model and challenges in pricing options under this model numerically. It then describes the methodology used, which involves analytic smoothing, adaptive sparse grids quadrature, quasi Monte Carlo, and coupling these with hierarchical representations and Richardson extrapolation. Several figures are included to illustrate the adaptive construction of sparse grids and simulation of the rough Bergomi dynamics.
Hierarchical Deterministic Quadrature Methods for Option Pricing under the Ro...Chiheb Ben Hammouda
Conference talk at the SIAM Conference on Financial Mathematics and Engineering, held in virtual format, June 1-4 2021, about our recently published work "Hierarchical adaptive sparse grids and quasi-Monte Carlo for option pricing under the rough Bergomi model".
- Link of the paper: https://www.tandfonline.com/doi/abs/10.1080/14697688.2020.1744700
This document provides an overview of discrete choice analysis and nested logit models. It begins with a review of binary and multinomial logit models and the independence from irrelevant alternatives property. It then introduces nested logit models as a way to address IIA violations when alternatives are correlated or choices are multidimensional. The document provides an example of a nested logit specification and calculation of choice probabilities. It concludes with extensions like mixed logit models and an appendix on additional model specifications.
This document discusses iterative Monte Carlo methods for solving linear systems. It begins by motivating the use of stochastic approaches for solving highly dimensional systems due to the occurrence of faults in parallel computing environments. It then provides the mathematical setting for standard Monte Carlo linear solvers using both forward and adjoint methods. It discusses how the solution can be written as a power series and evaluated using random walks. It also discusses the necessary statistical constraints for the estimators to converge, such as the spectral radii of the preconditioned matrix being less than one. Finally, it addresses the choice of transition probabilities for the random walks to guarantee the constraints.
This document discusses the origins and meaning of regression analysis in econometrics. It describes Francis Galton's studies in the late 1800s of genetic inheritance, where he observed that offspring traits tended to regress towards the mean trait of the general population. This concept of regression led to the development of the regression model in statistics. The document provides examples of regression analysis using data on heights from Galton's studies and discusses how regression relates to the concept of correlation.
The document proposes a new approach called successively quadratic interpolation for polynomial interpolation that is more efficient than Neville's and Aitken's algorithms. It involves iteratively computing quadratic interpolation polynomials using three points rather than linear interpolation using two points. The new algorithm reduces computational costs by about 20% compared to Neville's algorithm. Numerical experiments on test functions show the new algorithm has lower CPU time than Neville's algorithm while achieving the same solutions, demonstrating its improved efficiency.
We provide a review of the recent literature on statistical risk bounds for deep neural networks. We also discuss some theoretical results that compare the performance of deep ReLU networks to other methods such as wavelets and spline-type methods. The talk will moreover highlight some open problems and sketch possible new directions.
The document discusses methods for efficiently and accurately estimating integrals, including Monte Carlo simulation, low-discrepancy sampling, and Bayesian cubature. It notes that product rules for estimating high-dimensional integrals become prohibitively expensive as dimension increases. Adaptive low-discrepancy sampling is proposed as a method that uses Sobol' or lattice points and normally doubles the number of points until a tolerance is reached.
Information-theoretic clustering with applicationsFrank Nielsen
Information-theoretic clustering with applications
Abstract: Clustering is a fundamental and key primitive to discover structural groups of homogeneous data in data sets, called clusters. The most famous clustering technique is the celebrated k-means clustering that seeks to minimize the sum of intra-cluster variances. k-Means is NP-hard as soon as the dimension and the number of clusters are both greater than 1. In the first part of the talk, we first present a generic dynamic programming method to compute the optimal clustering of n scalar elements into k pairwise disjoint intervals. This case includes 1D Euclidean k-means but also other kinds of clustering algorithms like the k-medoids, the k-medians, the k-centers, etc.
We extend the method to incorporate cluster size constraints and show how to choose the appropriate number of clusters using model selection. We then illustrate and refine the method on two case studies: 1D Bregman clustering and univariate statistical mixture learning maximizing the complete likelihood. In the second part of the talk, we introduce a generalization of k-means to cluster sets of histograms that has become an important ingredient of modern information processing due to the success of the bag-of-word modelling paradigm.
Clustering histograms can be performed using the celebrated k-means centroid-based algorithm. We consider the Jeffreys divergence that symmetrizes the Kullback-Leibler divergence, and investigate the computation of Jeffreys centroids. We prove that the Jeffreys centroid can be expressed analytically using the Lambert W function for positive histograms. We then show how to obtain a fast guaranteed approximation when dealing with frequency histograms and conclude with some remarks on the k-means histogram clustering.
References: - Optimal interval clustering: Application to Bregman clustering and statistical mixture learning IEEE ISIT 2014 (recent result poster) http://arxiv.org/abs/1403.2485
- Jeffreys Centroids: A Closed-Form Expression for Positive Histograms and a Guaranteed Tight Approximation for Frequency Histograms.
IEEE Signal Process. Lett. 20(7): 657-660 (2013) http://arxiv.org/abs/1303.7286
http://www.i.kyoto-u.ac.jp/informatics-seminar/
Differential analyses of structures in HiC datatuxette
When Hi-C matrices are collected from two different conditions, methods can compare the matrices to identify regions with significant structural differences between conditions. TADpole and TADcompare are two available methods. TADpole represents hierarchical TAD structures and detects differences by computing a difference index between normalized binarized matrices. TADcompare represents Hi-C matrices as networks and uses the eigenvectors of the graph Laplacian and gap scores to define boundaries and detect differential boundaries between conditions. Both methods were shown to recover known breakpoints and have boundaries enriched for biological marks.
Here we overview the problem of clustering, why it's so different from supervised learning problems, how to select the number of clusters. We discuss main approaches to perform clustering: k-Means, hierarchical clustering, DBSCAN.
Mini useR! in Melbourne https://www.meetup.com/fr-FR/MelbURN-Melbourne-Users-of-R-Network/events/251933078/
MelbURN (Melbourne useR group) https://www.meetup.com/fr-FR/MelbURN-Melbourne-Users-of-R-Network
July 16th, 2018
Melbourne, Australia
1. The document discusses the author's research in three areas: graph-based clustering methods, approximate Bayesian computation (ABC), and Bayesian computation using empirical likelihood.
2. For graph-based clustering, the author presents asymptotic results for spectral clustering as the number of data points and bandwidth approach infinity.
3. For ABC, the author discusses sequential ABC algorithms and challenges of model choice and high-dimensional summary statistics. Machine learning methods are proposed to analyze simulated ABC data.
4. For empirical likelihood, the author proposes using it for Bayesian computation when the likelihood is intractable and simulation is infeasible, as it provides correct confidence intervals unlike composite likelihoods.
A Maximum Entropy Approach to the Loss Data Aggregation ProblemErika G. G.
This document discusses using a maximum entropy approach to model loss distributions for operational risk modeling. It begins by motivating the need to accurately model loss distributions given challenges with limited datasets, including heavy tails and dependence between risks. It then provides an overview of the loss distribution approach commonly used in operational risk modeling and its limitations. The document introduces the maximum entropy approach, which frames the problem as maximizing entropy subject to moment constraints. It discusses using the Laplace transform of loss distributions to compress information into moments and how these can be estimated from sample data or fitted distributions to serve as constraints for the maximum entropy approach.
This document summarizes the work of Working Group II on Probabilistic Numerics from the SAMSI QMC Transition Workshop. The working group aims to develop probabilistic numerical methods that provide a richer probabilistic quantification of numerical error in outputs, allowing for better statistical inference. Members of the working group have published several papers on topics like Bayesian probabilistic numerical methods for solving differential equations and performing integral approximations, and applying these methods to problems in mathematical epidemiology and industrial process monitoring. The group has also organized workshops and reading groups to discuss the development of probabilistic numerical methods.
Global Bilateral Symmetry Detection Using Multiscale Mirror HistogramsMohamed Elawady
M. ELAWADY, C. BARAT, C. DUCOTTET and P. COLANTONI
Laboratoire Hubert Curien, Saint-Etienne, FR
Conference "Advanced Concepts for Intelligent Vision Systems
" 2016
The International Journal of Engineering & Science is aimed at providing a platform for researchers, engineers, scientists, or educators to publish their original research results, to exchange new ideas, to disseminate information in innovative designs, engineering experiences and technological skills. It is also the Journal's objective to promote engineering and technology education. All papers submitted to the Journal will be blind peer-reviewed. Only original articles will be published.
The papers for publication in The International Journal of Engineering& Science are selected through rigorous peer reviews to ensure originality, timeliness, relevance, and readability.
The document discusses different perspectives on simulating the mean of a function, including deterministic, randomized, and Bayesian approaches. It summarizes Monte Carlo methods using the central limit theorem and Berry-Esseen inequality to estimate error bounds. Low-discrepancy sampling and cubature methods are described which use Fourier coefficients to bound integration errors. Bayesian cubature is outlined, which assumes the function is drawn from a Gaussian process prior to perform optimal quadrature. Maximum likelihood is used to estimate the kernel hyperparameters.
This document summarizes and compares two popular Python libraries for graph neural networks - Spektral and PyTorch Geometric. It begins by providing an overview of the basic functionality and architecture of each library. It then discusses how each library handles data loading and mini-batching of graph data. The document reviews several common message passing layer types implemented in both libraries. It provides an example comparison of using each library for a node classification task on the Cora dataset. Finally, it discusses a graph classification comparison in PyTorch Geometric using different message passing and pooling layers on the IMDB-binary dataset.
Since the advent of the horseshoe priors for regularization, global-local shrinkage methods have proved to be a fertile ground for the development of Bayesian theory and methodology in machine learning. They have achieved remarkable success in computation, and enjoy strong theoretical support. Much of the existing literature has focused on the linear Gaussian case. The purpose of the current talk is to demonstrate that the horseshoe priors are useful more broadly, by reviewing both methodological and computational developments in complex models that are more relevant to machine learning applications. Specifically, we focus on methodological challenges in horseshoe regularization in nonlinear and non-Gaussian models; multivariate models; and deep neural networks. We also outline the recent computational developments in horseshoe shrinkage for complex models along with a list of available software implementations that allows one to venture out beyond the comfort zone of the canonical linear regression problems.
This document summarizes a research paper about using hierarchical deterministic quadrature methods for option pricing under the rough Bergomi model. It discusses the rough Bergomi model and challenges in pricing options under this model numerically. It then describes the methodology used, which involves analytic smoothing, adaptive sparse grids quadrature, quasi Monte Carlo, and coupling these with hierarchical representations and Richardson extrapolation. Several figures are included to illustrate the adaptive construction of sparse grids and simulation of the rough Bergomi dynamics.
Hierarchical Deterministic Quadrature Methods for Option Pricing under the Ro...Chiheb Ben Hammouda
Conference talk at the SIAM Conference on Financial Mathematics and Engineering, held in virtual format, June 1-4 2021, about our recently published work "Hierarchical adaptive sparse grids and quasi-Monte Carlo for option pricing under the rough Bergomi model".
- Link of the paper: https://www.tandfonline.com/doi/abs/10.1080/14697688.2020.1744700
This document provides an overview of discrete choice analysis and nested logit models. It begins with a review of binary and multinomial logit models and the independence from irrelevant alternatives property. It then introduces nested logit models as a way to address IIA violations when alternatives are correlated or choices are multidimensional. The document provides an example of a nested logit specification and calculation of choice probabilities. It concludes with extensions like mixed logit models and an appendix on additional model specifications.
This document discusses iterative Monte Carlo methods for solving linear systems. It begins by motivating the use of stochastic approaches for solving highly dimensional systems due to the occurrence of faults in parallel computing environments. It then provides the mathematical setting for standard Monte Carlo linear solvers using both forward and adjoint methods. It discusses how the solution can be written as a power series and evaluated using random walks. It also discusses the necessary statistical constraints for the estimators to converge, such as the spectral radii of the preconditioned matrix being less than one. Finally, it addresses the choice of transition probabilities for the random walks to guarantee the constraints.
This document discusses the origins and meaning of regression analysis in econometrics. It describes Francis Galton's studies in the late 1800s of genetic inheritance, where he observed that offspring traits tended to regress towards the mean trait of the general population. This concept of regression led to the development of the regression model in statistics. The document provides examples of regression analysis using data on heights from Galton's studies and discusses how regression relates to the concept of correlation.
The document proposes a new approach called successively quadratic interpolation for polynomial interpolation that is more efficient than Neville's and Aitken's algorithms. It involves iteratively computing quadratic interpolation polynomials using three points rather than linear interpolation using two points. The new algorithm reduces computational costs by about 20% compared to Neville's algorithm. Numerical experiments on test functions show the new algorithm has lower CPU time than Neville's algorithm while achieving the same solutions, demonstrating its improved efficiency.
We provide a review of the recent literature on statistical risk bounds for deep neural networks. We also discuss some theoretical results that compare the performance of deep ReLU networks to other methods such as wavelets and spline-type methods. The talk will moreover highlight some open problems and sketch possible new directions.
The document discusses methods for efficiently and accurately estimating integrals, including Monte Carlo simulation, low-discrepancy sampling, and Bayesian cubature. It notes that product rules for estimating high-dimensional integrals become prohibitively expensive as dimension increases. Adaptive low-discrepancy sampling is proposed as a method that uses Sobol' or lattice points and normally doubles the number of points until a tolerance is reached.
Information-theoretic clustering with applicationsFrank Nielsen
Information-theoretic clustering with applications
Abstract: Clustering is a fundamental and key primitive to discover structural groups of homogeneous data in data sets, called clusters. The most famous clustering technique is the celebrated k-means clustering that seeks to minimize the sum of intra-cluster variances. k-Means is NP-hard as soon as the dimension and the number of clusters are both greater than 1. In the first part of the talk, we first present a generic dynamic programming method to compute the optimal clustering of n scalar elements into k pairwise disjoint intervals. This case includes 1D Euclidean k-means but also other kinds of clustering algorithms like the k-medoids, the k-medians, the k-centers, etc.
We extend the method to incorporate cluster size constraints and show how to choose the appropriate number of clusters using model selection. We then illustrate and refine the method on two case studies: 1D Bregman clustering and univariate statistical mixture learning maximizing the complete likelihood. In the second part of the talk, we introduce a generalization of k-means to cluster sets of histograms that has become an important ingredient of modern information processing due to the success of the bag-of-word modelling paradigm.
Clustering histograms can be performed using the celebrated k-means centroid-based algorithm. We consider the Jeffreys divergence that symmetrizes the Kullback-Leibler divergence, and investigate the computation of Jeffreys centroids. We prove that the Jeffreys centroid can be expressed analytically using the Lambert W function for positive histograms. We then show how to obtain a fast guaranteed approximation when dealing with frequency histograms and conclude with some remarks on the k-means histogram clustering.
References: - Optimal interval clustering: Application to Bregman clustering and statistical mixture learning IEEE ISIT 2014 (recent result poster) http://arxiv.org/abs/1403.2485
- Jeffreys Centroids: A Closed-Form Expression for Positive Histograms and a Guaranteed Tight Approximation for Frequency Histograms.
IEEE Signal Process. Lett. 20(7): 657-660 (2013) http://arxiv.org/abs/1303.7286
http://www.i.kyoto-u.ac.jp/informatics-seminar/
Differential analyses of structures in HiC datatuxette
When Hi-C matrices are collected from two different conditions, methods can compare the matrices to identify regions with significant structural differences between conditions. TADpole and TADcompare are two available methods. TADpole represents hierarchical TAD structures and detects differences by computing a difference index between normalized binarized matrices. TADcompare represents Hi-C matrices as networks and uses the eigenvectors of the graph Laplacian and gap scores to define boundaries and detect differential boundaries between conditions. Both methods were shown to recover known breakpoints and have boundaries enriched for biological marks.
Here we overview the problem of clustering, why it's so different from supervised learning problems, how to select the number of clusters. We discuss main approaches to perform clustering: k-Means, hierarchical clustering, DBSCAN.
Mini useR! in Melbourne https://www.meetup.com/fr-FR/MelbURN-Melbourne-Users-of-R-Network/events/251933078/
MelbURN (Melbourne useR group) https://www.meetup.com/fr-FR/MelbURN-Melbourne-Users-of-R-Network
July 16th, 2018
Melbourne, Australia
1. The document discusses the author's research in three areas: graph-based clustering methods, approximate Bayesian computation (ABC), and Bayesian computation using empirical likelihood.
2. For graph-based clustering, the author presents asymptotic results for spectral clustering as the number of data points and bandwidth approach infinity.
3. For ABC, the author discusses sequential ABC algorithms and challenges of model choice and high-dimensional summary statistics. Machine learning methods are proposed to analyze simulated ABC data.
4. For empirical likelihood, the author proposes using it for Bayesian computation when the likelihood is intractable and simulation is infeasible, as it provides correct confidence intervals unlike composite likelihoods.
A Maximum Entropy Approach to the Loss Data Aggregation ProblemErika G. G.
This document discusses using a maximum entropy approach to model loss distributions for operational risk modeling. It begins by motivating the need to accurately model loss distributions given challenges with limited datasets, including heavy tails and dependence between risks. It then provides an overview of the loss distribution approach commonly used in operational risk modeling and its limitations. The document introduces the maximum entropy approach, which frames the problem as maximizing entropy subject to moment constraints. It discusses using the Laplace transform of loss distributions to compress information into moments and how these can be estimated from sample data or fitted distributions to serve as constraints for the maximum entropy approach.
This document summarizes the work of Working Group II on Probabilistic Numerics from the SAMSI QMC Transition Workshop. The working group aims to develop probabilistic numerical methods that provide a richer probabilistic quantification of numerical error in outputs, allowing for better statistical inference. Members of the working group have published several papers on topics like Bayesian probabilistic numerical methods for solving differential equations and performing integral approximations, and applying these methods to problems in mathematical epidemiology and industrial process monitoring. The group has also organized workshops and reading groups to discuss the development of probabilistic numerical methods.
Global Bilateral Symmetry Detection Using Multiscale Mirror HistogramsMohamed Elawady
M. ELAWADY, C. BARAT, C. DUCOTTET and P. COLANTONI
Laboratoire Hubert Curien, Saint-Etienne, FR
Conference "Advanced Concepts for Intelligent Vision Systems
" 2016
This document summarizes research on using particle swarm optimization to reconstruct microwave images of two-dimensional dielectric scatterers. It formulates the inverse scattering problem as an optimization problem to find the dielectric parameter distribution that minimizes the difference between measured and simulated scattered field data. Numerical results show that a particle swarm optimization approach can accurately reconstruct the shape and dielectric properties of a test cylindrical scatterer, with lower background reconstruction error than a genetic algorithm approach. The research demonstrates that particle swarm optimization is a suitable technique for high-dimensional microwave imaging problems.
Special Plenary Lecture at the International Conference on VIBRATION ENGINEERING AND TECHNOLOGY OF MACHINERY (VETOMAC), Lisbon, Portugal, September 10 - 13, 2018
http://www.conf.pt/index.php/v-speakers
Propagation of uncertainties in complex engineering dynamical systems is receiving increasing attention. When uncertainties are taken into account, the equations of motion of discretised dynamical systems can be expressed by coupled ordinary differential equations with stochastic coefficients. The computational cost for the solution of such a system mainly depends on the number of degrees of freedom and number of random variables. Among various numerical methods developed for such systems, the polynomial chaos based Galerkin projection approach shows significant promise because it is more accurate compared to the classical perturbation based methods and computationally more efficient compared to the Monte Carlo simulation based methods. However, the computational cost increases significantly with the number of random variables and the results tend to become less accurate for a longer length of time. In this talk novel approaches will be discussed to address these issues. Reduced-order Galerkin projection schemes in the frequency domain will be discussed to address the problem of a large number of random variables. Practical examples will be given to illustrate the application of the proposed Galerkin projection techniques.
A Non Parametric Estimation Based Underwater Target ClassifierCSCJournals
Underwater noise sources constitute a prominent class of input signal in most underwater signal processing systems. The problem of identification of noise sources in the ocean is of great importance because of its numerous practical applications. In this paper, a methodology is presented for the detection and identification of underwater targets and noise sources based on non parametric indicators. The proposed system utilizes Cepstral coefficient analysis and the Kruskal-Wallis H statistic along with other statistical indicators like F-test statistic for the effective detection and classification of noise sources in the ocean. Simulation results for typical underwater noise data and the set of identified underwater targets are also presented in this paper.
Information theory deals with quantifying and transmitting information. It answers questions about data compression and transmission rates. Shannon showed that transmission rates can exceed the channel capacity as long as they remain below it, and that random processes have an irreducible complexity below which they cannot be compressed. Information theory relates to fields like computer science, probability theory, and communication theory. It had its beginnings in the early 20th century but was developed further after World War II by scientists like Shannon and Wiener.
This document summarizes Robert Fry's presentation on computation and design of autonomous intelligent systems. It outlines a computational theory of intelligence based on defining questions and answers within a system. Key points include:
- Intelligent systems acquire information to make decisions to achieve goals.
- Questions are defined as sets of possible answers. Boolean algebra is used to represent questions and assertions.
- Probability and entropy theories are derived from this logical framework.
- A simple protozoan system is used to illustrate how a system maps information to decisions.
- Neural computation is modeled using this theory, with neurons posing questions and making optimal decisions.
- Hebbian learning allows neural systems to adapt optimally via dual-matching.
Wavelet-based Reflection Symmetry Detection via Textural and Color HistogramsMohamed Elawady
This document presents a methodology for wavelet-based reflection symmetry detection using textural and color histograms. It extracts multiscale edge segments using Log-Gabor filters and measures symmetry based on edge orientations, local texture histograms, and color histograms. Evaluation on public datasets shows it outperforms previous methods in detecting single and multiple symmetries, with quantitative and qualitative results presented. Future work could improve the detection using continuous maximal-seeking.
Research Inventy : International Journal of Engineering and Science is published by the group of young academic and industrial researchers with 12 Issues per year. It is an online as well as print version open access journal that provides rapid publication (monthly) of articles in all areas of the subject such as: civil, mechanical, chemical, electronic and computer engineering as well as production and information technology. The Journal welcomes the submission of manuscripts that meet the general criteria of significance and scientific excellence. Papers will be published by rapid process within 20 days after acceptance and peer review process takes only 7 days. All articles published in Research Inventy will be peer-reviewed.
This document discusses using information theory quantifiers to characterize chaotic systems and distinguish between chaotic and stochastic time series. It describes how chaos and noise share characteristics like broad power spectra and delta-like autocorrelation functions that make them difficult to distinguish. However, chaotic dynamics is not stochastic. Information theory quantifiers like entropy measures, Fisher information, and statistical complexity can be applied to probability distribution functions derived from time series data using methods like Bandt-Pompe patterns or visibility graphs in order to characterize the systems and distinguish between chaos and noise.
Multiple estimators for Monte Carlo approximationsChristian Robert
This document discusses multiple estimators that can be used to approximate integrals using Monte Carlo simulations. It begins by introducing concepts like multiple importance sampling, Rao-Blackwellisation, and delayed acceptance that allow combining multiple estimators to improve accuracy. It then discusses approaches like mixtures as proposals, global adaptation, and nonparametric maximum likelihood estimation (NPMLE) that frame Monte Carlo estimation as a statistical estimation problem. The document notes various advantages of the statistical formulation, like the ability to directly estimate simulation error from the Fisher information. Overall, the document presents an overview of different techniques for combining Monte Carlo simulations to obtain more accurate integral approximations.
The thesis aimed to advance knowledge in decentralized detection in wireless sensor networks. It developed efficient algorithms for designing decision rules at sensors to minimize error probability. It proved conditions where balanced rate allocation is optimal and applied this to sensor network models. It also formulated decentralized detection problems for energy harvesting sensor networks, developing analytical bounds and numerical design methods. The work provided computational and theoretical advances for optimal inference in distributed sensing applications.
The document analyzes the use of the Tent map as a source of pseudorandom bits for generating binary codes. It evaluates the Tent map's period length, discrimination value, and merit factor. The Tent map is proposed as an alternative to traditional low-complexity pseudorandom bit generators. Different window functions are applied to the binary codes generated from the Tent map to reduce side lobes and improve performance. Results show discrimination increases with sequence length and some window functions perform better than others at different lengths.
A NOVEL ANT COLONY ALGORITHM FOR MULTICAST ROUTING IN WIRELESS AD HOC NETWORKS cscpconf
The Steiner tree is the underlying model for multicast communication. This paper presents a
novel ant colony algorithm guided by problem relaxation for unconstrained Steiner tree in static
wireless ad hoc networks. The framework of the proposed algorithm is based on ant colony
system (ACS). In the first step, the ants probabilistically construct the path from the source tothe terminal nodes. These paths are then merged together to generate a Steiner tree rooted at the source. The problem is relaxed to incorporate the structural information into the heuristic value for the selection of nodes. The effectiveness of the algorithm is tested on the benchmark problems of the OR-library. Simulation results show that our algorithm can find optimal Steiner tree with high success rate.
Local Outlier Detection with InterpretationDaiki Tanaka
This paper proposes a method called Local Outlier Detection with Interpretation (LODI) that detects outliers and explains their anomalousness simultaneously. LODI first selects a neighboring set for each outlier candidate using entropy measures. It then computes an anomaly degree for each object based on its deviation from neighbors in a learned 1D subspace. Finally, LODI interprets outliers by identifying a small set of influential features. Experiments on synthetic and real-world data show LODI outperforms other methods in outlier detection and provides intuitive feature-based explanations. However, LODI's computation is expensive and it assumes linear separability, which are limitations for future work.
α Nearness ant colony system with adaptive strategies for the traveling sales...ijfcstjournal
On account of ant colony algorithm easy to fall into local optimum, this paper presents an improved ant
colony optimization called α-AACS and reports its performance. At first, we provide an concise description
of the original ant colony system(ACS) and introduce α-nearness based on the minimum 1-tree for ACS’s
disadvantage, which better reflects the chances of a given link being a member of an optimal tour. Then, we
improve α-nearness by computing a lower bound and propose other adaptations for ACS. Finally, we
conduct a fair competition between our algorithm and others. The results clearly show that α-AACS has a
better global searching ability in finding the best solutions, which indicates that α-AACS is an effective
approach for solving the traveling salesman problem.
PICO presentation at EGU 2014 about the use of measures from information theory to visualise uncertainty in kinematic structural models - and to estimate where additional data would help reduce uncertainties. Some nice counter-intuitive results ;-)
High-Dimensional Methods: Examples for Inference on Structural EffectsNBER
This document describes a study that uses high-dimensional methods to estimate the effect of 401(k) eligibility on measures of accumulated assets. It begins by outlining the baseline model and notes areas for improvement, such as controlling for income. It then discusses using regularization like LASSO for variable selection in high-dimensional settings. The document explores more flexible specifications by generating many interaction and polynomial terms but notes the need for dimension reduction. It describes using LASSO to select important variables from a large set. The results select a parsimonious set of variables and estimate similar 401(k) effects as the baseline.
Graph theory is the study of graphs, which are mathematical structures used to model pairwise relations between objects. A graph consists of vertices and edges connecting pairs of vertices. There are many types of graphs including trees, which are connected acyclic graphs. Spanning trees are subgraphs of a graph that connect all vertices using the minimum number of edges. Key concepts in graph theory include paths, connectedness, cycles, and isomorphism between graphs.
This document discusses dynamical systems. It defines a dynamical system as a system that changes over time according to fixed rules determining how its state changes from one time to the next. It then covers:
- The two parts of a dynamical system: state space and function determining next state.
- Classification of systems as deterministic/stochastic, discrete/continuous, linear/nonlinear, and autonomous/nonautonomous.
- Examples of discrete and continuous models, differential equations, and linear vs nonlinear systems.
- Terminology including phase space, phase curve, phase portrait, and attractors.
- Analysis methods including fixed points, stability, and perturbation analysis.
- Examples of harmonic oscillator,
The document discusses the Infinite Monkey Theorem. It provides references to papers and authors that are relevant to the theorem, including foundational works by Solomonoff, Kolmogorov, Levin, and Chaitin on algorithmic information theory and inductive inference. The references cover the original formulation of the theorem, its properties, applications to artificial intelligence, and relationships to concepts of information, randomness, and probability theory.
The document discusses the Algorithmic Coding Theorem proposed by Dr. Hector Zenil. The theorem concerns Kolmogorov complexity and algorithmic probability, showing that the algorithmic probability of a string is equal to its complexity up to a computable and additive constant term. The theorem unifies concepts from algorithmic information theory and inductive inference, and helps characterize randomness and structure in strings. It is supported by previous work from Solomonoff, Kolmogorov, Chaitin, and others in the fields of algorithmic information theory and inductive inference.
4.13: Bennett's Logical Depth: A Measure of Sophistication Hector Zenil
Bennett's Logical Depth is a measure of the sophistication of a string or program. It quantifies how hard it would be for a simple program to randomly generate the string. The deeper a string's logical depth, the more sophisticated or complex it is considered to be. Dr. Hector Zenil explores Bennett's Logical Depth as a way to analyze complexity and sophistication.
4.10: Algorithmic Probability & the Universal Distribution Hector Zenil
The universal distribution is a formalization of Ockham's razor that provides a semi-computable probability distribution over all possible strings. It produces the digits of mathematical constants like pi at random, with the probability of producing longer strings of correct digits falling exponentially, for example the probability of randomly producing the first 2400 digits of pi is 1/10^2400. The universal distribution and algorithmic probability were discovered by Solomonoff and further developed by others in the field of algorithmic information theory.
The Chaitin-Leibniz Medal is an award given to Dr. Hector Zenil for his contributions to algorithmic information theory and computational mechanics. Dr. Zenil is recognized for developing new methods to quantify complexity in networks and systems using algorithmic information theory. He has also made advances in applying algorithmic information theory to study biological and cognitive systems.
4.8: Epistemological Aspects of Infinite WisdomHector Zenil
This 3 sentence document discusses epistemological aspects of infinite wisdom. It mentions Dr. Hector Zenil and Jorge Luis Borges twice, suggesting the document is about their views on infinite wisdom from an epistemological perspective.
Chaitin's Omega number is a real number between 0 and 1 that represents the probability that a randomly generated computer program will halt. It is defined using prefix-free codes and Turing machines to avoid counting spurious program variations multiple times. The document then provides an example numeric value for Chaitin's Omega number.
Convergence of Definitions argues that incompressibility, unpredictability, and typicality converge and are essentially equivalent definitions. The paper explores how incompressibility relates to unpredictability and typicality by showing they are interchangeable concepts.
The document discusses the Invariance Theorem, which states that the behavior of a system is independent of the representation used to describe it. The theorem suggests that two different models that describe the same system will result in identical behaviors, even if the representations are different. The document examines the implications of the Invariance Theorem as it relates to modeling complex systems.
4.4: Algorithmic Complexity and CompressibilityHector Zenil
This document discusses the properties of algorithmic complexity and compressibility, including unpredictability which refers to the impossibility of predicting an outcome, incompressibility meaning something random cannot be described simply or shortly, and typicality where a random event has nothing special about it. It also examines program-size characterization and its most relevant properties in algorithmic complexity.
This document discusses pseudo-randomness and statistical tests used to analyze random number generators. It explains an algorithm that takes a digit number, squares it, and uses the middle digits as the random number for the next iteration. It also lists several popular statistical tests used to analyze randomness, including normality, compressibility, linear complexity, autocorrelation, and Fourier coefficient tests.
Randomness intuitively refers to events or objects that have no causal explanation, pattern, or predictability. However, classical probability has limitations in capturing true randomness as even simple repetitive sequences can be considered random. True randomness may come from quantum mechanical phenomena, which can generate unpredictable random numbers.
This document discusses classical information theory and concepts like entropy. It defines entropy as quantifying the apparent disorder of particles in a room. Different examples of sequences are provided, from a highly ordered repeating sequence to a more random typical sequence. Functions are defined to calculate the string length and characters of a given sequence. The document also shows how to calculate the entropy of a sequence and compares the results of different entropy calculation functions.
This document discusses Shannon Entropy and Meaning. It repeats the phrase "Shannon Entropy and Meaning" multiple times without providing any additional context or information about the topic. The document does not have a clear main point or message to summarize in 3 sentences or less.
Unit 3: Joint, Conditional, Mutual Information, & a Case StudyHector Zenil
This document discusses joint entropy, conditional entropy, and mutual information. It defines joint entropy as the entropy of two random variables together. Conditional entropy is defined as the entropy of one random variable given knowledge of the other. It provides examples of when conditional entropy would be 0 or equal to the entropy without conditioning. The document also mentions a case study to demonstrate these information theory concepts.
Unit 3: Entropy rate, languages and multidimensional dataHector Zenil
This document discusses entropy rate and its application to multidimensional objects. It first covers entropy rate and then examines how coarse-graining, or reducing the level of detail, can be used to calculate entropy for multidimensional systems represented as 6x6 grids. The document aims to explain how entropy rate extends the concept of entropy to capture the randomness within sequences and multidimensional structures over time.
Unit 3: Redundancy, noise, and biological informationHector Zenil
Information in biological systems is redundant and able to withstand noise. Repetition of key data allows organisms to survive errors and remain able to pass genetic information between generations. Redundancy and error correction are thus essential features of biological systems that help preserve life.
Algorithmic Dynamics of Cellular AutomataHector Zenil
Original presentation prepared for the opening keynote of the meeting in celebration of Prof. Harold McIntosh. This talk covers aspects of the complexity and behaviour of cellular automata and their emergent dynamic patterns and information dynamics of events such as particle collisions.
Microbial interaction
Microorganisms interacts with each other and can be physically associated with another organisms in a variety of ways.
One organism can be located on the surface of another organism as an ectobiont or located within another organism as endobiont.
Microbial interaction may be positive such as mutualism, proto-cooperation, commensalism or may be negative such as parasitism, predation or competition
Types of microbial interaction
Positive interaction: mutualism, proto-cooperation, commensalism
Negative interaction: Ammensalism (antagonism), parasitism, predation, competition
I. Mutualism:
It is defined as the relationship in which each organism in interaction gets benefits from association. It is an obligatory relationship in which mutualist and host are metabolically dependent on each other.
Mutualistic relationship is very specific where one member of association cannot be replaced by another species.
Mutualism require close physical contact between interacting organisms.
Relationship of mutualism allows organisms to exist in habitat that could not occupied by either species alone.
Mutualistic relationship between organisms allows them to act as a single organism.
Examples of mutualism:
i. Lichens:
Lichens are excellent example of mutualism.
They are the association of specific fungi and certain genus of algae. In lichen, fungal partner is called mycobiont and algal partner is called
II. Syntrophism:
It is an association in which the growth of one organism either depends on or improved by the substrate provided by another organism.
In syntrophism both organism in association gets benefits.
Compound A
Utilized by population 1
Compound B
Utilized by population 2
Compound C
utilized by both Population 1+2
Products
In this theoretical example of syntrophism, population 1 is able to utilize and metabolize compound A, forming compound B but cannot metabolize beyond compound B without co-operation of population 2. Population 2is unable to utilize compound A but it can metabolize compound B forming compound C. Then both population 1 and 2 are able to carry out metabolic reaction which leads to formation of end product that neither population could produce alone.
Examples of syntrophism:
i. Methanogenic ecosystem in sludge digester
Methane produced by methanogenic bacteria depends upon interspecies hydrogen transfer by other fermentative bacteria.
Anaerobic fermentative bacteria generate CO2 and H2 utilizing carbohydrates which is then utilized by methanogenic bacteria (Methanobacter) to produce methane.
ii. Lactobacillus arobinosus and Enterococcus faecalis:
In the minimal media, Lactobacillus arobinosus and Enterococcus faecalis are able to grow together but not alone.
The synergistic relationship between E. faecalis and L. arobinosus occurs in which E. faecalis require folic acid
Anti-Universe And Emergent Gravity and the Dark UniverseSérgio Sacani
Recent theoretical progress indicates that spacetime and gravity emerge together from the entanglement structure of an underlying microscopic theory. These ideas are best understood in Anti-de Sitter space, where they rely on the area law for entanglement entropy. The extension to de Sitter space requires taking into account the entropy and temperature associated with the cosmological horizon. Using insights from string theory, black hole physics and quantum information theory we argue that the positive dark energy leads to a thermal volume law contribution to the entropy that overtakes the area law precisely at the cosmological horizon. Due to the competition between area and volume law entanglement the microscopic de Sitter states do not thermalise at sub-Hubble scales: they exhibit memory effects in the form of an entropy displacement caused by matter. The emergent laws of gravity contain an additional ‘dark’ gravitational force describing the ‘elastic’ response due to the entropy displacement. We derive an estimate of the strength of this extra force in terms of the baryonic mass, Newton’s constant and the Hubble acceleration scale a0 = cH0, and provide evidence for the fact that this additional ‘dark gravity force’ explains the observed phenomena in galaxies and clusters currently attributed to dark matter.
Discovery of An Apparent Red, High-Velocity Type Ia Supernova at 𝐳 = 2.9 wi...Sérgio Sacani
We present the JWST discovery of SN 2023adsy, a transient object located in a host galaxy JADES-GS
+
53.13485
−
27.82088
with a host spectroscopic redshift of
2.903
±
0.007
. The transient was identified in deep James Webb Space Telescope (JWST)/NIRCam imaging from the JWST Advanced Deep Extragalactic Survey (JADES) program. Photometric and spectroscopic followup with NIRCam and NIRSpec, respectively, confirm the redshift and yield UV-NIR light-curve, NIR color, and spectroscopic information all consistent with a Type Ia classification. Despite its classification as a likely SN Ia, SN 2023adsy is both fairly red (
�
(
�
−
�
)
∼
0.9
) despite a host galaxy with low-extinction and has a high Ca II velocity (
19
,
000
±
2
,
000
km/s) compared to the general population of SNe Ia. While these characteristics are consistent with some Ca-rich SNe Ia, particularly SN 2016hnk, SN 2023adsy is intrinsically brighter than the low-
�
Ca-rich population. Although such an object is too red for any low-
�
cosmological sample, we apply a fiducial standardization approach to SN 2023adsy and find that the SN 2023adsy luminosity distance measurement is in excellent agreement (
≲
1
�
) with
Λ
CDM. Therefore unlike low-
�
Ca-rich SNe Ia, SN 2023adsy is standardizable and gives no indication that SN Ia standardized luminosities change significantly with redshift. A larger sample of distant SNe Ia is required to determine if SN Ia population characteristics at high-
�
truly diverge from their low-
�
counterparts, and to confirm that standardized luminosities nevertheless remain constant with redshift.
Candidate young stellar objects in the S-cluster: Kinematic analysis of a sub...Sérgio Sacani
Context. The observation of several L-band emission sources in the S cluster has led to a rich discussion of their nature. However, a definitive answer to the classification of the dusty objects requires an explanation for the detection of compact Doppler-shifted Brγ emission. The ionized hydrogen in combination with the observation of mid-infrared L-band continuum emission suggests that most of these sources are embedded in a dusty envelope. These embedded sources are part of the S-cluster, and their relationship to the S-stars is still under debate. To date, the question of the origin of these two populations has been vague, although all explanations favor migration processes for the individual cluster members. Aims. This work revisits the S-cluster and its dusty members orbiting the supermassive black hole SgrA* on bound Keplerian orbits from a kinematic perspective. The aim is to explore the Keplerian parameters for patterns that might imply a nonrandom distribution of the sample. Additionally, various analytical aspects are considered to address the nature of the dusty sources. Methods. Based on the photometric analysis, we estimated the individual H−K and K−L colors for the source sample and compared the results to known cluster members. The classification revealed a noticeable contrast between the S-stars and the dusty sources. To fit the flux-density distribution, we utilized the radiative transfer code HYPERION and implemented a young stellar object Class I model. We obtained the position angle from the Keplerian fit results; additionally, we analyzed the distribution of the inclinations and the longitudes of the ascending node. Results. The colors of the dusty sources suggest a stellar nature consistent with the spectral energy distribution in the near and midinfrared domains. Furthermore, the evaporation timescales of dusty and gaseous clumps in the vicinity of SgrA* are much shorter ( 2yr) than the epochs covered by the observations (≈15yr). In addition to the strong evidence for the stellar classification of the D-sources, we also find a clear disk-like pattern following the arrangements of S-stars proposed in the literature. Furthermore, we find a global intrinsic inclination for all dusty sources of 60 ± 20◦, implying a common formation process. Conclusions. The pattern of the dusty sources manifested in the distribution of the position angles, inclinations, and longitudes of the ascending node strongly suggests two different scenarios: the main-sequence stars and the dusty stellar S-cluster sources share a common formation history or migrated with a similar formation channel in the vicinity of SgrA*. Alternatively, the gravitational influence of SgrA* in combination with a massive perturber, such as a putative intermediate mass black hole in the IRS 13 cluster, forces the dusty objects and S-stars to follow a particular orbital arrangement. Key words. stars: black holes– stars: formation– Galaxy: center– galaxies: star formation
Mechanics:- Simple and Compound PendulumPravinHudge1
a compound pendulum is a physical system with a more complex structure than a simple pendulum, incorporating its mass distribution and dimensions into its oscillatory motion around a fixed axis. Understanding its dynamics involves principles of rotational mechanics and the interplay between gravitational potential energy and kinetic energy. Compound pendulums are used in various scientific and engineering applications, such as seismology for measuring earthquakes, in clocks to maintain accurate timekeeping, and in mechanical systems to study oscillatory motion dynamics.
Sexuality - Issues, Attitude and Behaviour - Applied Social Psychology - Psyc...PsychoTech Services
A proprietary approach developed by bringing together the best of learning theories from Psychology, design principles from the world of visualization, and pedagogical methods from over a decade of training experience, that enables you to: Learn better, faster!
Evidence of Jet Activity from the Secondary Black Hole in the OJ 287 Binary S...Sérgio Sacani
Wereport the study of a huge optical intraday flare on 2021 November 12 at 2 a.m. UT in the blazar OJ287. In the binary black hole model, it is associated with an impact of the secondary black hole on the accretion disk of the primary. Our multifrequency observing campaign was set up to search for such a signature of the impact based on a prediction made 8 yr earlier. The first I-band results of the flare have already been reported by Kishore et al. (2024). Here we combine these data with our monitoring in the R-band. There is a big change in the R–I spectral index by 1.0 ±0.1 between the normal background and the flare, suggesting a new component of radiation. The polarization variation during the rise of the flare suggests the same. The limits on the source size place it most reasonably in the jet of the secondary BH. We then ask why we have not seen this phenomenon before. We show that OJ287 was never before observed with sufficient sensitivity on the night when the flare should have happened according to the binary model. We also study the probability that this flare is just an oversized example of intraday variability using the Krakow data set of intense monitoring between 2015 and 2023. We find that the occurrence of a flare of this size and rapidity is unlikely. In machine-readable Tables 1 and 2, we give the full orbit-linked historical light curve of OJ287 as well as the dense monitoring sample of Krakow.
SDSS1335+0728: The awakening of a ∼ 106M⊙ black hole⋆Sérgio Sacani
Context. The early-type galaxy SDSS J133519.91+072807.4 (hereafter SDSS1335+0728), which had exhibited no prior optical variations during the preceding two decades, began showing significant nuclear variability in the Zwicky Transient Facility (ZTF) alert stream from December 2019 (as ZTF19acnskyy). This variability behaviour, coupled with the host-galaxy properties, suggests that SDSS1335+0728 hosts a ∼ 106M⊙ black hole (BH) that is currently in the process of ‘turning on’. Aims. We present a multi-wavelength photometric analysis and spectroscopic follow-up performed with the aim of better understanding the origin of the nuclear variations detected in SDSS1335+0728. Methods. We used archival photometry (from WISE, 2MASS, SDSS, GALEX, eROSITA) and spectroscopic data (from SDSS and LAMOST) to study the state of SDSS1335+0728 prior to December 2019, and new observations from Swift, SOAR/Goodman, VLT/X-shooter, and Keck/LRIS taken after its turn-on to characterise its current state. We analysed the variability of SDSS1335+0728 in the X-ray/UV/optical/mid-infrared range, modelled its spectral energy distribution prior to and after December 2019, and studied the evolution of its UV/optical spectra. Results. From our multi-wavelength photometric analysis, we find that: (a) since 2021, the UV flux (from Swift/UVOT observations) is four times brighter than the flux reported by GALEX in 2004; (b) since June 2022, the mid-infrared flux has risen more than two times, and the W1−W2 WISE colour has become redder; and (c) since February 2024, the source has begun showing X-ray emission. From our spectroscopic follow-up, we see that (i) the narrow emission line ratios are now consistent with a more energetic ionising continuum; (ii) broad emission lines are not detected; and (iii) the [OIII] line increased its flux ∼ 3.6 years after the first ZTF alert, which implies a relatively compact narrow-line-emitting region. Conclusions. We conclude that the variations observed in SDSS1335+0728 could be either explained by a ∼ 106M⊙ AGN that is just turning on or by an exotic tidal disruption event (TDE). If the former is true, SDSS1335+0728 is one of the strongest cases of an AGNobserved in the process of activating. If the latter were found to be the case, it would correspond to the longest and faintest TDE ever observed (or another class of still unknown nuclear transient). Future observations of SDSS1335+0728 are crucial to further understand its behaviour. Key words. galaxies: active– accretion, accretion discs– galaxies: individual: SDSS J133519.91+072807.4
TOPIC OF DISCUSSION: CENTRIFUGATION SLIDESHARE.pptxshubhijain836
Centrifugation is a powerful technique used in laboratories to separate components of a heterogeneous mixture based on their density. This process utilizes centrifugal force to rapidly spin samples, causing denser particles to migrate outward more quickly than lighter ones. As a result, distinct layers form within the sample tube, allowing for easy isolation and purification of target substances.
4. To illustrate the limitations of entropy
we created a very simple graph:
1. Let 1 → 2 be a starting graph G connecting a
node with label 1 to a node with label 2. If a node
with label n has degree n, we call it a core node;
otherwise, we call it a supportive node.
2. Iteratively add a node n + 1 to G such that the
number of core nodes in G is maximised.
5.
6. Algorithm:
The degree sequence d of the labeled nodes
d = 1,2, 3, 4, . . . , n
is the Champernowne constant in base 10, a transcendental real
whose decimal expansion is Borel normal.
7. Algorithm:
The sequence of number of edges is a function of core and supportive nodes, defined by
[1/r] + [2/r] +···+ [n/r],
where r = (1 + √5)/2 is the golden ratio and [· ] the floor
function (sequence A183136 in the OEIS), whose values are
1, 2, 4, 7, 10, 14, 18, 23, 29, 35, 42, 50, 58, 67, 76, 86, 97, 108,
120, 132, 145,....
12. Maximum Entropy (MaxEnt)
Model Principle
The principle of maximum entropy states that the
probability distribution which best represents the
current state of knowledge is the one with largest
entropy…
E.T. Jaynes
15. K(s) = min{ |p| | U(p) = s}
Computability, Algorithmic
Complexity & Causality
(Un)Computability mediates in the challenge of causality
by way of Algorithmic Information Theory:
16. Computability, Algorithmic
Complexity & Causality
(Un)Computability mediates in the challenge of causality
by way of Algorithmic Information Theory:
Generating
mechanism
Observation
K(s) = min{ |p| | U(p) = s}
17. Generating
mechanism
Observation
K(s) = min{ |p| | U(p) = s}
Most likely
mechanism
according
to Ockham’s
Computability, Algorithmic
Complexity & Causality
(Un)Computability mediates in the challenge of causality
by way of Algorithmic Information Theory:
18. Constructor
Generating
mechanism
Observation
K(s) = min{ |p| | U(p) = s}
Most likely
mechanism
according
to Ockham’s
Computability, Algorithmic
Complexity & Causality
(Un)Computability mediates in the challenge of causality
by way of Algorithmic Information Theory:
21. Algorithmic Data Analytics, Small Data Matters and Correlation
versus Causation. In Computability of the World? Philosophy and
Science in the Age of Big Data), Springer Verlag, 2017
suggestive
methods
Coding Theorem Method
(CTM)
22. Algorithmic Data Analytics, Small Data Matters and Correlation
versus Causation. In Computability of the World? Philosophy and
Science in the Age of Big Data), Springer Verlag, 2017
Coding Theorem Method
(CTM)
suggestive
methods
39. Divide and Conquer
sample
sample
sample
sample
sample
If any part of the whole system (samples) is of high m(x) and low K(x), then that part can
be generated by mechanistic/algorithmic means and thus is causal. The lower BDM the
more causal.
Rube
Goldberg
machine
40. Block Decomposition Method
(BDM)
sample
sample
sample
sample
If any part of the whole system (samples) is of high m(x) and low K(x), then that part can
be generated by mechanistic/algorithmic means and thus is causal. The lower BDM the
more causal.
Rube
Goldberg
machine
41. Formally:
Block Decomposition Method
(BDM)
H Zenil, F Soler-Toscano, N.A. Kiani, S. Hernández-Orozco, A. Rueda-Toicen
A Decomposition Method for Global Evaluation of Shannon Entropy and Local Estimations of Algorithmic
Complexity, Entropy 20(8), 605, 2018.
42. Block Decomposition Method
(BDM)
H Zenil, F Soler-Toscano, N.A. Kiani, S. Hernández-Orozco, A. Rueda-Toicen
A Decomposition Method for Global Evaluation of Shannon Entropy and Local Estimations of Algorithmic
Complexity, Entropy 20(8), 605, 2018.
43. Block Decomposition Method
(BDM)
H Zenil, F Soler-Toscano, N.A. Kiani, S. Hernández-Orozco, A. Rueda-Toicen
A Decomposition Method for Global Evaluation of Shannon Entropy and Local Estimations of Algorithmic
Complexity, Entropy 20(8), 605, 2018.
44. Block Decomposition Method
(BDM)
H Zenil, F Soler-Toscano, N.A. Kiani, S. Hernández-Orozco, A. Rueda-Toicen
A Decomposition Method for Global Evaluation of Shannon Entropy and Local Estimations of Algorithmic
Complexity, Entropy 20(8), 605, 2018.
45. Block Decomposition Method
(BDM)
For example:
If 2 strings are 10111 and 10111, a tighter upper bound of their
algorithmic complexity K is not:
|K(U(p)=10111)| + |K(U(p)=10111)| but
|K(U(p)=10111 2 times)|
H Zenil, F Soler-Toscano, N.A. Kiani, S. Hernández-Orozco, A. Rueda-Toicen
A Decomposition Method for Global Evaluation of Shannon Entropy and Local Estimations of Algorithmic
Complexity, Entropy 20(8), 605, 2018.
46. Hybrid Measure:
Shannon Entropy +
Local Algorithmic Complexity
Because clearly:
|K(U(p)=10111)| + |K(U(p)=10111)| >|K(U(p)=10111 2 times)|
H Zenil, F Soler-Toscano, N.A. Kiani, S. Hernández-Orozco, A. Rueda-Toicen
A Decomposition Method for Global Evaluation of Shannon Entropy and Local Estimations of Algorithmic
Complexity, Entropy 20(8), 605, 2018.
47. Hybrid Measure:
Shannon Entropy +
Local Algorithmic Complexity
H Zenil, F Soler-Toscano, N.A. Kiani, S. Hernández-Orozco, A. Rueda-Toicen
A Decomposition Method for Global Evaluation of Shannon Entropy and Local Estimations of Algorithmic
Complexity, Entropy 20(8), 605, 2018.
48. e.g. These 2 strings are of high Entropy (and high Entropy rate)
but low K:
010101110100, 011011010010
Gaining Causation Power
H Zenil, F Soler-Toscano, N.A. Kiani, S. Hernández-Orozco, A. Rueda-Toicen
A Decomposition Method for Global Evaluation of Shannon Entropy and Local Estimations of Algorithmic
Complexity, Entropy 20(8), 605, 2018.
49. Examples
H Zenil, F Soler-Toscano, N.A. Kiani, S. Hernández-Orozco, A. Rueda-Toicen
A Decomposition Method for Global Evaluation of Shannon Entropy and Local Estimations of Algorithmic
Complexity, Entropy 20(8), 605, 2018.
50. CTM and BDM conform with the
theoretical expectation
H Zenil, F Soler-Toscano, N.A. Kiani, S. Hernández-Orozco, A. Rueda-Toicen
A Decomposition Method for Global Evaluation of Shannon Entropy and Local Estimations of Algorithmic
Complexity, Entropy 20(8), 605, 2018.
51. Example of a generating
mechanism/model
https://journals.plos.org/plosone/article?id=10.1371/journal.pone.0096223
52. Algorithmic Sequence Model
Sequential aggregation of small models produce a model able to explain larger
data. TMs: 2205 <> 1351 <> 1447 <> 599063 can produce:
S = 00000101111111111111111111110
+ +
+
54. Version history and future
of the OACC
We keep expanding the calculator to wider
horizons of methodological and numerical capabilities:
• Version 1: Estimations of K for short binary strings
• Version 2: Expanded CTM and BDM capabilities for non-binary
strings and arrays
• Version 2.5: Estimations of Bennett's logical depth based on
CTM
• Version 3 (current version): Algorithmic Information dynamics
of strings, arrays and networks
• Version 4: Algorithmic complexity of models, algorithmic feature
selection, algorithmic dimensionality reduction and model
generator. (BETA version ready!)
complexitycalculator.com
55. We have seen how important AIT and
the role of Computability theory is in
the challenge of causality in science in
the 4th revolution of causal discover
towards model-driven approaches
https://www.youtube.com/watch?v=BEaXyDS_1Cw&t=30s
63. Boundary Conditions
H Zenil, F Soler-Toscano, N.A. Kiani, S. Hernández-Orozco, A. Rueda-Toicen
A Decomposition Method for Global Evaluation of Shannon Entropy and Local Estimations of Algorithmic
Complexity, Entropy 20(8), 605, 2018.
87. The 2D CA Game of Life (GoL)
Density diagram showing the
persistence of structures:
The Game of Life
88. The Game of Life Space-time
evolution
H. Zenil et al., Algorithmic-information Properties of Dynamic Persistent
Patterns and Colliding Particles in the Game of Life.
In A. Adamatzky (ed), From Parallel to Emergent Computing (book),
Taylor & Francis / CRC Press, 2018
89. The Game of Life Space-time
evolution
H. Zenil et al., Algorithmic-information Properties of Dynamic Persistent
Patterns and Colliding Particles in the Game of Life.
In A. Adamatzky (ed), From Parallel to Emergent Computing (book),
Taylor & Francis / CRC Press, 2018
90. GoL’s glider
(a ‘moving’ particle)
A piece of information moving through the grid from one side to another
92. Characterization of evolving
patterns
H. Zenil et al., Algorithmic-information Properties of Dynamic Persistent
Patterns and Colliding Particles in the Game of Life.
In A. Adamatzky (ed), From Parallel to Emergent Computing (book),
Taylor & Francis / CRC Press, 2018
95. Dynamics of a 4-particle
collision
New particles
Fixed point
H. Zenil et al., Algorithmic-information Properties of Dynamic Persistent Patterns and Colliding Particles in the Game
of Life. In A. Adamatzky (ed), From Parallel to Emergent Computin, Taylor & Francis / CRC, 2018
96. Algorithmic dynamics of a stable
‘near miss’
Zenil et al., Algorithmic-information Properties of Dynamic Persistent
Patterns and Colliding Particles in the Game of Life.
In A. Adamatzky (ed), From Parallel to Emergent Computing (book),
Taylor & Francis / CRC Press, 2018
97. Sensitivity of dynamics to initial conditions
Periodic fixed point
Zenil et al., Algorithmic-information Properties of Dynamic Persistent
Patterns and Colliding Particles in the Game of Life.
In A. Adamatzky (ed), From Parallel to Emergent Computing (book),
Taylor & Francis / CRC Press, 2018
98. Particle annihilation
H. Zenil, N.A. Kiani and J. Tegnér, Algorithmic-
information Properties of Dynamic Persistent
Patterns and Colliding Particles in the Game of
Life, In A. Adamatzky (ed), From Parallel to
Emergent Computing (book), Taylor & Francis /
CRC Press, 2018
100. All cases can be reduced to 4 density
plots (+ annihilation)
Algorithmicdynamics
ofglidercollisions
0 50 100 150 200
50
100
500
1000
BDM all cases
102. A Causal Calculus Based on
Algorithmic Information Dynamics
Let S be a non-random binary file containing a string in binary:
S= 10101010101010101010101010101010101010101010101010101010101010
Clearly, S is algorithmically compressible, with pS = “Print(01) 31 times” a short
computer program generating S. Let S’ be equal to S except for a bitwise operation
(bitwise NOT, or complement) in, say, position 23:
S’ = 10101010101010101010100010101010101010101010101010101010101010
We have that the relationship of their algorithmic complexity denoted by C is:
C(S) < C(S’)
for any single-bit mutation of S, where C(S) = |pS| and C(S’) = |pS’| are the lengths
of the shortest computer programs generating S and S’.
103. The algorithmic information dynamics of a string (step 0) pushed towards and away
from randomness by digit removal using BDM.
The initial string consists of a simple segment of ten 1s followed by a random-looking
segment of 10 digits. The resulting strings mostly extract each of the respective
segments.
Moving a string
104. Neutral Deletion
Neutral digit deletion maximises the preservation of the elements contributing to
the algorithmic information content of the original string thus the most important
(computable) features (of which statistical regularities are a subset). Here applied
to a string (step 0) after
removal of 10 digits.
105. Random data is of
different nature
Now consider S to be a binary file of the same size but consisting of, say,
random data:
01101100100011001101010001110110001100011010011100010000110
So pS = Print(S) is the shortest possible generating program, then, in a
random object:
All perturbations are neutral!
H Zenil, N.A. Kiani, F. Marabita, Y. Deng, S. Elias, A. Schmidt, G. Ball, J. Tegnér,
An Algorithmic Information Calculus for Causal Discovery and Reprogramming
Systems, bioaRXiv DOI: https://doi.org/10.1101/185637
106. Identifying random objects by
the effect of perturbations
01101100100011001101010001110110001100011010011100010000110
H Zenil, N.A. Kiani, F. Marabita, Y. Deng, S.
Elias, A. Schmidt, G. Ball, J. Tegnér, An
Algorithmic Information Calculus for Causal
Discovery and Reprogramming
Systems, bioaRXiv DOI:
https://doi.org/10.1101/185637
107. The Principle of Algorithmic
Information Dynamics
C(rule 30)
C(rule 30 after 1 step) = C(rule 30 after 2 steps) + log(2) = …
C(rule 30 after 100 steps) + log(100) … = C(rule 30 after 1000 steps) + log(1000)
In a deterministic isolated dynamical system algorithmic complexity never
changes if it did, then it is neither deterministic or isolated and we can
identify those elements
108. Networks as programs
H Zenil, N.A. Kiani, F. Marabita, Y. Deng, S. Elias, A. Schmidt, G. Ball, J. Tegnér,
An Algorithmic Information Calculus for Causal Discovery and Reprogramming Systems, bioaRXiv DOI:
https://doi.org/10.1101/185637
109. Information Ranking and
Information Spectra
H Zenil, N.A. Kiani, F. Marabita, Y. Deng, S. Elias, A. Schmidt, G. Ball, J. Tegnér,
An Algorithmic Information Calculus for Causal Discovery and Reprogramming Systems, bioaRXiv DOI:
https://doi.org/10.1101/185637
110. Information Signatures
σ(G)
H Zenil, N.A. Kiani, F. Marabita, Y.
Deng, S. Elias, A. Schmidt, G. Ball,
J. Tegnér,
An Algorithmic Information Calculus
for Causal Discovery and
Reprogramming Systems, bioaRXiv
DOI: https://doi.org/10.1101/185637
111. H Zenil, N.A. Kiani, F. Marabita, Y. Deng, S. Elias, A. Schmidt, G. Ball, J. Tegnér,
An Algorithmic Information Calculus for Causal Discovery and Reprogramming Systems, bioaRXiv DOI:
https://doi.org/10.1101/185637
112. Moving Networks
H Zenil, N.A. Kiani, F. Marabita, Y. Deng, S. Elias, A. Schmidt, G. Ball, J. Tegnér,
An Algorithmic Information Calculus for Causal Discovery and Reprogramming Systems, bioaRXiv DOI:
https://doi.org/10.1101/185637
114. AID Application to
Dynamical Systems
Original
candidate
model set M
with individual
models mi in M
After negative
perturbation:
candidate
models M’
grow in length (in
average):
Σi=1
|M’||m’i|>Σi=1
|M||mi|
Positive
perturbation:
candidate
models
decrease in length
Neutral
perturbation
|M| is the
cardinality of M
and |mi| the size
of model m in M.
115. Reconstructing space-time
dynamics of ECAs
original
reconstructed
H Zenil, N.A. Kiani, F. Marabita, Y. Deng, S. Elias, A. Schmidt, G. Ball, J. Tegnér,
An Algorithmic Information Calculus for Causal Discovery and Reprogramming Systems, bioaRXiv DOI:
https://doi.org/10.1101/185637
116. Reconstructing time in space-time
dynamics
We are generalizing these tools to continuous dynamical systems such as strange attractors
original
reconstructed
H Zenil, N.A. Kiani, F. Marabita, Y. Deng, S. Elias, A. Schmidt, G. Ball, J. Tegnér,
An Algorithmic Information Calculus for Causal Discovery and Reprogramming Systems, bioaRXiv DOI:
https://doi.org/10.1101/185637
117. Reconstructing space-phase dynamics of
ECA (more observations)
H Zenil, N.A. Kiani, F. Marabita, Y. Deng, S. Elias, A. Schmidt, G. Ball, J. Tegnér,
An Algorithmic Information Calculus for Causal Discovery and Reprogramming Systems, bioaRXiv DOI:
https://doi.org/10.1101/185637
118. Sensitivity to perturbations at
difference time steps
H Zenil, N.A. Kiani, F. Marabita, Y. Deng, S. Elias, A. Schmidt, G. Ball, J. Tegnér,
An Algorithmic Information Calculus for Causal Discovery and Reprogramming Systems, bioaRXiv DOI:
https://doi.org/10.1101/185637
121. Evidence of pervasive
Reprogrammability
Reprogrammability of ECA and
CA
J. Riedel, H. Zenil
Cross-boundary Behavioural Reprogrammability Reveals Evidence of Pervasive Universality
International Journal of Unconventional Computing, vol 13:14-15 pp. 309-357
122. J. Riedel, H. Zenil
Cross-boundary Behavioural Reprogrammability Reveals Evidence of Pervasive Universality
International Journal of Unconventional Computing, vol 13:14-15 pp. 309-357
Reprogrammability Networks
Evidence of pervasive
Turing universality
123. J. Riedel and H. Zenil
Rule Primality, Minimal Generating Sets and Turing-Universality in the Causal Decomposition of Elementary Cellular
Automata
Journal of Cellular Automata, vol. 13, pp. 479–497, 2018
New Universal
Cellular Automata
125. An Algorithmic Information Calculus for Causal Discovery and Reprogramming Systems
H Zenil, N.A. Kiani, F. Marabita, Y. Deng, S. Elias, A. Schmidt, G. Ball, J. Tegnér
doi: https://doi.org/10.1101/185637
Algorithmic space Dynamical phase space
Algorithmic/Dynamic
Landscape Relationship