This document discusses the average case complexity of binary tree sort compared to quicksort. It analyzes the robustness of binary tree sort's O(n log n) average case complexity for non-uniform input distributions like binomial, Poisson, discrete uniform, continuous uniform, and standard normal distributions. Through statistical modeling and simulation experiments on different sample sizes, it is shown that binary tree sort maintains O(n log n) average case complexity even for non-uniform inputs, demonstrating that it is more robust than quicksort in the average case. The document concludes that only algorithms with equal worst case and average case complexities can reliably depend on average complexity measures, otherwise robustness must be verified.
Linked CP Tensor Decomposition (presented by ICONIP2012)Tatsuya Yokota
This document proposes a new method called Linked Tensor Decomposition (LTD) to analyze common and individual factors from a group of tensor data. LTD combines the advantages of Individual Tensor Decomposition (ITD), which analyzes individual characteristics, and Simultaneous Tensor Decomposition (STD), which analyzes common factors in a group. LTD represents each tensor as the sum of a common factor and individual factors. An algorithm using Hierarchical Alternating Least Squares is developed to solve the LTD model. Experiments on toy problems and face reconstruction demonstrate LTD can extract both common and individual factors more effectively than ITD or STD alone. Future work will explore Tucker-based LTD and statistical independence in the LTD model
Introduction to Common Spatial Pattern Filters for EEG Motor Imagery Classifi...Tatsuya Yokota
This document introduces common spatial pattern (CSP) filters for EEG motor imagery classification. CSP filters aim to find spatial patterns in EEG data that maximize the difference between two classes. The document outlines several CSP algorithms including standard CSP, common spatially standardized CSP, and spatially constrained CSP. CSP filters extract discriminative features from EEG data that can improve classification accuracy for brain-computer interface applications involving motor imagery tasks.
A Mixed Binary-Real NSGA II Algorithm Ensuring Both Accuracy and Interpretabi...IJECEIAES
In this work, a Neuro-Fuzzy Controller network, called NFC that implements a Mamdani fuzzy inference system is proposed. This network includes neurons able to perform fundamental fuzzy operations. Connections between neurons are weighted through binary and real weights. Then a mixed binaryreal Non dominated Sorting Genetic Algorithm II (NSGA II) is used to perform both accuracy and interpretability of the NFC by minimizing two objective functions; one objective relates to the number of rules, for compactness, while the second is the mean square error, for accuracy. In order to preserve interpretability of fuzzy rules during the optimization process, some constraints are imposed. The approach is tested on two control examples: a single input single output (SISO) system and a multivariable (MIMO) system.
One of the fundamental issues in computer science is ordering a list of items. Although there is a number of sorting algorithms, sorting problem has attracted a great deal of research, because efficient sorting is important to optimize the use of other algorithms. This paper presents a new sorting algorithm which runs faster by decreasing the number of comparisons by taking some extra memory. In this algorithm we are using lists to sort the elements. This algorithm was analyzed, implemented and tested and the results are promising for a random data
This document presents an exponential-Lindley additive failure rate model (ELAFRM) by combining the hazard functions of an exponential distribution and a Lindley distribution. The key properties of the ELAFRM are derived, including the probability density function, cumulative distribution function, hazard function, moments, and graphical representations. Estimation of the model parameters is also discussed. The document proposes this new ELAFRM distribution and analyzes its mathematical properties.
Multilayer Backpropagation Neural Networks for Implementation of Logic GatesIJCSES Journal
ANN is a computational model that is composed of several processing elements (neurons) that tries to solve a specific problem. Like the human brain, it provides the ability to learn
from experiences without being explicitly programmed. This article is based on the implementation of artificial neural networks for logic gates. At first, the 3 layers Artificial Neural Network is
designed with 2 input neurons, 2 hidden neurons & 1 output neuron. after that model is trained
by using a backpropagation algorithm until the model satisfies the predefined error criteria (e)
which set 0.01 in this experiment. The learning rate (α) used for this experiment was 0.01. The
NN model produces correct output at iteration (p)= 20000 for AND, NAND & NOR gate. For
OR & XOR the correct output is predicted at iteration (p)=15000 & 80000 respectively
The document discusses finding initial parameters for neural networks used in data clustering. It presents a modified K-means fast learning artificial neural network (K-FLANN) algorithm that uses differential evolution to optimize the vigilance and tolerance parameters of K-FLANN. Differential evolution is used to select parameter values that yield good clustering performance. The modified K-FLANN algorithm is evaluated on several data sets and shown to provide promising results in terms of convergence rate and accuracy compared to other algorithms. Comparisons are also made between the original and modified versions of K-FLANN.
Linked CP Tensor Decomposition (presented by ICONIP2012)Tatsuya Yokota
This document proposes a new method called Linked Tensor Decomposition (LTD) to analyze common and individual factors from a group of tensor data. LTD combines the advantages of Individual Tensor Decomposition (ITD), which analyzes individual characteristics, and Simultaneous Tensor Decomposition (STD), which analyzes common factors in a group. LTD represents each tensor as the sum of a common factor and individual factors. An algorithm using Hierarchical Alternating Least Squares is developed to solve the LTD model. Experiments on toy problems and face reconstruction demonstrate LTD can extract both common and individual factors more effectively than ITD or STD alone. Future work will explore Tucker-based LTD and statistical independence in the LTD model
Introduction to Common Spatial Pattern Filters for EEG Motor Imagery Classifi...Tatsuya Yokota
This document introduces common spatial pattern (CSP) filters for EEG motor imagery classification. CSP filters aim to find spatial patterns in EEG data that maximize the difference between two classes. The document outlines several CSP algorithms including standard CSP, common spatially standardized CSP, and spatially constrained CSP. CSP filters extract discriminative features from EEG data that can improve classification accuracy for brain-computer interface applications involving motor imagery tasks.
A Mixed Binary-Real NSGA II Algorithm Ensuring Both Accuracy and Interpretabi...IJECEIAES
In this work, a Neuro-Fuzzy Controller network, called NFC that implements a Mamdani fuzzy inference system is proposed. This network includes neurons able to perform fundamental fuzzy operations. Connections between neurons are weighted through binary and real weights. Then a mixed binaryreal Non dominated Sorting Genetic Algorithm II (NSGA II) is used to perform both accuracy and interpretability of the NFC by minimizing two objective functions; one objective relates to the number of rules, for compactness, while the second is the mean square error, for accuracy. In order to preserve interpretability of fuzzy rules during the optimization process, some constraints are imposed. The approach is tested on two control examples: a single input single output (SISO) system and a multivariable (MIMO) system.
One of the fundamental issues in computer science is ordering a list of items. Although there is a number of sorting algorithms, sorting problem has attracted a great deal of research, because efficient sorting is important to optimize the use of other algorithms. This paper presents a new sorting algorithm which runs faster by decreasing the number of comparisons by taking some extra memory. In this algorithm we are using lists to sort the elements. This algorithm was analyzed, implemented and tested and the results are promising for a random data
This document presents an exponential-Lindley additive failure rate model (ELAFRM) by combining the hazard functions of an exponential distribution and a Lindley distribution. The key properties of the ELAFRM are derived, including the probability density function, cumulative distribution function, hazard function, moments, and graphical representations. Estimation of the model parameters is also discussed. The document proposes this new ELAFRM distribution and analyzes its mathematical properties.
Multilayer Backpropagation Neural Networks for Implementation of Logic GatesIJCSES Journal
ANN is a computational model that is composed of several processing elements (neurons) that tries to solve a specific problem. Like the human brain, it provides the ability to learn
from experiences without being explicitly programmed. This article is based on the implementation of artificial neural networks for logic gates. At first, the 3 layers Artificial Neural Network is
designed with 2 input neurons, 2 hidden neurons & 1 output neuron. after that model is trained
by using a backpropagation algorithm until the model satisfies the predefined error criteria (e)
which set 0.01 in this experiment. The learning rate (α) used for this experiment was 0.01. The
NN model produces correct output at iteration (p)= 20000 for AND, NAND & NOR gate. For
OR & XOR the correct output is predicted at iteration (p)=15000 & 80000 respectively
The document discusses finding initial parameters for neural networks used in data clustering. It presents a modified K-means fast learning artificial neural network (K-FLANN) algorithm that uses differential evolution to optimize the vigilance and tolerance parameters of K-FLANN. Differential evolution is used to select parameter values that yield good clustering performance. The modified K-FLANN algorithm is evaluated on several data sets and shown to provide promising results in terms of convergence rate and accuracy compared to other algorithms. Comparisons are also made between the original and modified versions of K-FLANN.
The document describes the formulation of a Linear-Strain Triangular (LST) finite element. The LST element has 6 nodes, 12 degrees of freedom, and a quadratic displacement function, offering advantages over the Constant Strain Triangular (CST) element. The procedure to derive the LST element stiffness equations is identical to that used for the CST element. Key steps include discretizing the element, selecting displacement functions, defining strain-displacement relationships, and deriving the element stiffness matrix using the total potential energy approach.
Minimum weekly temperature forecasting using anfisAlexander Decker
The document describes using an Adaptive Neuro-Fuzzy Inference System (ANFIS) to forecast minimum weekly temperatures. Mean weekly maximum and minimum temperature data from 1997-2006 in Dehradun, India was used. ANFIS combines neural networks and fuzzy logic to model relationships between inputs like maximum temperatures and predict minimum temperatures. The ANFIS model was trained on 400 weeks of data and tested on 117 weeks of data to predict minimum weekly temperatures one week ahead based on current and past maximum temperature inputs.
Bayesian system reliability and availability analysis underthe vague environm...ijsc
Reliability modeling is the most important discipline of reliable engineering. The main purpose of this paper is to provide a methodology for discussing the vague environment. Actually we discuss on Bayesian system reliability and availability analysis on the vague environment based on Exponential distribution
under squared error symmetric and precautionary asymmetric loss functions. In order to apply the Bayesian approach, model parameters are assumed to be vague random variables with vague prior distributions. This approach will be used to create the vague Bayes estimate of system reliability and availability by introducing and applying a theorem called “Resolution Identity” for vague sets. For this purpose, the original problem is transformed into a nonlinear programming problem which is then divided up into eight subproblems to simplify computations. Finally, the results obtained for the subproblems can
be used to determine the membership functions of the vague Bayes estimate of system reliability. Finally, the sub problems can be solved by using any commercial optimizers, e.g. GAMS or LINGO.
IRJET- Performance Analysis of Optimization Techniques by using ClusteringIRJET Journal
This document discusses optimization techniques for clustering algorithms. It introduces fuzzy bee colony optimization (FBCO) and compares its performance to other swarm algorithms like fuzzy c-means (FCM) and fuzzy particle swarm optimization (FPSO). FBCO is motivated by the natural behaviors of bee colonies and aims to avoid local minima problems. The document provides background on clustering, describes the FCM and FPSO algorithms, and proposes a FBCO algorithm to improve clustering performance.
The document discusses necessary conditions for minimal system configurations of typical Takagi-Sugeno (TS) fuzzy systems and Mamdani fuzzy systems as universal approximators. It establishes that for TS fuzzy systems using trapezoidal input fuzzy sets and linear rule consequents, the number of input fuzzy sets and rules needed depends on the number and locations of extrema of the function to be approximated. Specifically, functions with a few extrema may require only a handful of rules, while periodic or oscillatory functions would require many rules. It then compares these conditions to those previously established for Mamdani fuzzy systems, finding their minimal configurations are comparable. Finally, it proves the conditions can be reduced for TS systems using non-tra
This document summarizes a study that used fuzzy finite element analysis to solve a heat transfer problem where material properties were treated as fuzzy parameters. The problem considered heat transfer through the periphery of an insulated circular iron rod. Thermal conductivity and heat transfer coefficient were represented as triangular fuzzy numbers to account for uncertainty. The temperatures at various points along the rod were calculated using analytical, finite difference, and finite element methods. The finite element method was also used to solve the problem considering the uncertain material properties as fuzzy values. Results were presented as fuzzy plots showing the temperature ranges associated with different alpha cut values. The fuzzy finite element method was able to provide solutions that closely matched the traditional finite element method while incorporating uncertainty in the material properties.
On Continuous Approximate Solution of Ordinary Differential EquationsWaqas Tariq
In this work the problem of continuous approximate solution of the ordinary differential equations will be investigated. An approach to construct the continuous approximate solution, which is based on the discrete approximate solution and the spline interpolation, will be provided. The existence and uniqueness of such continuous approximate solution will be pointed out. Its error will be estimated and its convergence will be considered. Finally, with the aid of modern PC and nathematical software three practical computer approaches to perform above construction will be offered.
FUZZY IMAGE SEGMENTATION USING VALIDITY INDEXES CORRELATIONijcsit
This paper introduces an algorithm for image segmentation using a clustering technique; the technique is
based on the fuzzy c means algorithm (FCM) that is executed iteratively with different number of clusters.
Furthermore, simultaneously five validity indexes are calculated and their information is correlated to
determine the optimal number of clusters in order to segment an image, results and simulations are shown
in the paper.
Financial Time Series Analysis Based On Normalized Mutual Information FunctionsIJCI JOURNAL
A method of predictability analysis of future values of financial time series is described. The method is based on normalized mutual information functions. In the analysis, the use of these functions allowed to refuse any restrictions on the distributions of the parameters and on the correlations between parameters. A comparative analysis of the predictability of financial time series of Tel Aviv 25 stock exchange has been carried out.
This document presents two new Boolean combination methods called "orthogonalizing difference-building" and "orthogonal OR-ing". These methods calculate the difference and complement of functions, as well as the EXOR and EXNOR of minterms or functions, resulting in orthogonal forms. Orthogonal forms have advantages for further calculations. The methods are based on set theory and treat minterms and functions as ternary-vector lists. Equations are provided to define the difference and disjunction of minterms or functions in terms of their ternary-vector list representations.
Hybrid PSO-SA algorithm for training a Neural Network for ClassificationIJCSEA Journal
In this work, we propose a Hybrid particle swarm optimization-Simulated annealing algorithm and present a comparison with i) Simulated annealing algorithm and ii) Back propagation algorithm for training neural networks. These neural networks were then tested on a classification task. In particle swarm optimization behaviour of a particle is influenced by the experiential knowledge of the particle as well as socially exchanged information. Particle swarm optimization follows a parallel search strategy. In simulated annealing uphill moves are made in the search space in a stochastic fashion in addition to the downhill moves. Simulated annealing therefore has better scope of escaping local minima and reach a global minimum in the search space. Thus simulated annealing gives a selective randomness to the search. Back propagation algorithm uses gradient descent approach search for minimizing the error. Our goal of global minima in the task being done here is to come to lowest energy state, where energy state is being modelled as the sum of the squares of the error between the target and observed output values for all the training samples. We compared the performance of the neural networks of identical architectures trained by the i) Hybrid particle swarm optimization-simulated annealing, ii) Simulated annealing and iii) Back propagation algorithms respectively on a classification task and noted the results obtained. Neural network trained by Hybrid particle swarm optimization-simulated annealing has given better results compared to the neural networks trained by the Simulated annealing and Back propagation algorithms in the tests conducted by us.
Iterative Determinant Method for Solving Eigenvalue Problemsijceronline
International Journal of Computational Engineering Research (IJCER) is dedicated to protecting personal information and will make every reasonable effort to handle collected information appropriately. All information collected, as well as related requests, will be handled as carefully and efficiently as possible in accordance with IJCER standards for integrity and objectivity.
Matrix factorization techniques can be used to address some of the limitations of traditional collaborative filtering approaches for recommender systems. Matrix factorization decomposes the user-item rating matrix into the product of two lower-dimensional matrices, one representing latent factors for users and the other for items. This reduced dimensionality addresses data sparsity and scalability issues. Specifically, singular value decomposition is often used to perform this matrix factorization, which can approximate the original rating matrix while ignoring less important singular values and factor vectors. The decomposed matrices can then be multiplied to predict unknown user ratings.
Catching co occurrence information using word2vec-inspired matrix factorizationhyunsung lee
- Factorizing a PMI matrix involves using matrix factorization techniques to represent words in a latent space based on their co-occurrence information, similar to word2vec.
- Recommender systems use matrix factorization to represent users and items numerically in a latent space to predict target values like ratings. Known ratings are used to find latent vectors for users and items that best approximate the rating matrix.
- These latent vectors can then be used to predict unknown ratings by taking the dot product of the user and item vectors.
This document summarizes a study that used an adaptive neuro-fuzzy inference system (ANFIS) to model rainfall-runoff in the Nagwan watershed in India. Daily rainfall and runoff data from 1993-1999 were used to train ANFIS models with different combinations of rainfall and runoff as inputs and current day runoff as the output. The best performing model used only current day rainfall as the input and had 91 Gaussian membership functions. This model had low root mean square error and high correlation for both the training and testing periods. The study demonstrated that ANFIS is effective for hydrological rainfall-runoff modeling in small watersheds where runoff is highly dependent on current day rainfall.
This document proposes a neural collaborative filtering model called NeuMF that combines generalized matrix factorization and multi-layer perceptrons. It represents users and items as dense embedding vectors. NeuMF takes the concatenation of user and item embeddings as input and outputs a prediction through multiple nonlinear layers. The model is pretrained using GMF and MLP models and then jointly optimized. Experiments on MovieLens data show NeuMF achieves better performance than MF, MLP, and other baselines in terms of hit ratio and NDCG for top-K recommendation. It improves over SVD++ in terms of RMSE and MAE.
Information Theory and Programmable MedicineHector Zenil
This document is a slide presentation by Hector Zenil on information theory and programmable medicine. It discusses using information theory and a behavioral approach to computation to map how molecular stimuli relate to the conformational spaces of biological systems. It provides an example of simulating the self-assembly of porphyrin molecules under different binding energies to demonstrate how this approach could help program biological behaviors and inform treatments. The goal is to work towards programmable medicine by introducing molecular arrays that can release payloads based on certain biological conditions being met.
CHN and Swap Heuristic to Solve the Maximum Independent Set ProblemIJECEIAES
We describe a new approach to solve the problem to find the maximum independent set in a given Graph, known also as Max-Stable set problem (MSSP). In this paper, we show how Max-Stable problem can be reformulated into a linear problem under quadratic constraints, and then we resolve the QP result by a hybrid approach based Continuous Hopfeild Neural Network (CHN) and Local Search. In a manner that the solution given by the CHN will be the starting point of the local search. The new approach showed a good performance than the original one which executes a suite of CHN runs, at each execution a new leaner constraint is added into the resolved model. To prove the efficiency of our approach, we present some computational experiments of solving random generated problem and typical MSSP instances of real life problem.
Algorithmic Information Theory and Computational BiologyHector Zenil
I present cutting-edge concepts and tools drawn from algorithmic information theory (AIT) for new generation genetic sequencing, network biology and bioinformatics in general. AIT is the most advanced mathematical theory of information theory formally characterising the concepts and differences between simplicity, randomness and structure. Measures of AIT will empower computational medicine and systems biology to deal with big data, sophisticated analytics and a powerful new understanding framework.
PARTITION SORT REVISITED: RECONFIRMING THE ROBUSTNESS IN AVERAGE CASE AND MUC...IJCSEA Journal
In our previous work there was some indication that Partition Sort could be having a more robust average case O(nlogn) complexity than the popular Quick sort. In our first study in this paper, we reconfirm this through computer experiments for inputs from Cauchy distribution for which expectation theoretically does not exist. Additionally, the algorithm is found to be sensitive to parameters of the input probability distribution demanding further investigation on parameterized complexity. The results on this algorithm for Binomial inputs in our second study are very encouraging in that direction.
Partition Sort Revisited: Reconfirming the Robustness in Average Case and Muc...ijcsa
In our previous work there was some indication that Partition Sort could be having a more robust average case O(nlogn) complexity than the popular Quick sort. In our first study in this paper, we reconfirm this through computer experiments for inputs from Cauchy distribution for which expectation theoretically does not exist. Additionally, the algorithm is found to be sensitive to parameters of the input probability distribution demanding further investigation on parameterized complexity. The results on this algorithm for Binomial inputs in our second study are very encouraging in that direction.
The document describes the formulation of a Linear-Strain Triangular (LST) finite element. The LST element has 6 nodes, 12 degrees of freedom, and a quadratic displacement function, offering advantages over the Constant Strain Triangular (CST) element. The procedure to derive the LST element stiffness equations is identical to that used for the CST element. Key steps include discretizing the element, selecting displacement functions, defining strain-displacement relationships, and deriving the element stiffness matrix using the total potential energy approach.
Minimum weekly temperature forecasting using anfisAlexander Decker
The document describes using an Adaptive Neuro-Fuzzy Inference System (ANFIS) to forecast minimum weekly temperatures. Mean weekly maximum and minimum temperature data from 1997-2006 in Dehradun, India was used. ANFIS combines neural networks and fuzzy logic to model relationships between inputs like maximum temperatures and predict minimum temperatures. The ANFIS model was trained on 400 weeks of data and tested on 117 weeks of data to predict minimum weekly temperatures one week ahead based on current and past maximum temperature inputs.
Bayesian system reliability and availability analysis underthe vague environm...ijsc
Reliability modeling is the most important discipline of reliable engineering. The main purpose of this paper is to provide a methodology for discussing the vague environment. Actually we discuss on Bayesian system reliability and availability analysis on the vague environment based on Exponential distribution
under squared error symmetric and precautionary asymmetric loss functions. In order to apply the Bayesian approach, model parameters are assumed to be vague random variables with vague prior distributions. This approach will be used to create the vague Bayes estimate of system reliability and availability by introducing and applying a theorem called “Resolution Identity” for vague sets. For this purpose, the original problem is transformed into a nonlinear programming problem which is then divided up into eight subproblems to simplify computations. Finally, the results obtained for the subproblems can
be used to determine the membership functions of the vague Bayes estimate of system reliability. Finally, the sub problems can be solved by using any commercial optimizers, e.g. GAMS or LINGO.
IRJET- Performance Analysis of Optimization Techniques by using ClusteringIRJET Journal
This document discusses optimization techniques for clustering algorithms. It introduces fuzzy bee colony optimization (FBCO) and compares its performance to other swarm algorithms like fuzzy c-means (FCM) and fuzzy particle swarm optimization (FPSO). FBCO is motivated by the natural behaviors of bee colonies and aims to avoid local minima problems. The document provides background on clustering, describes the FCM and FPSO algorithms, and proposes a FBCO algorithm to improve clustering performance.
The document discusses necessary conditions for minimal system configurations of typical Takagi-Sugeno (TS) fuzzy systems and Mamdani fuzzy systems as universal approximators. It establishes that for TS fuzzy systems using trapezoidal input fuzzy sets and linear rule consequents, the number of input fuzzy sets and rules needed depends on the number and locations of extrema of the function to be approximated. Specifically, functions with a few extrema may require only a handful of rules, while periodic or oscillatory functions would require many rules. It then compares these conditions to those previously established for Mamdani fuzzy systems, finding their minimal configurations are comparable. Finally, it proves the conditions can be reduced for TS systems using non-tra
This document summarizes a study that used fuzzy finite element analysis to solve a heat transfer problem where material properties were treated as fuzzy parameters. The problem considered heat transfer through the periphery of an insulated circular iron rod. Thermal conductivity and heat transfer coefficient were represented as triangular fuzzy numbers to account for uncertainty. The temperatures at various points along the rod were calculated using analytical, finite difference, and finite element methods. The finite element method was also used to solve the problem considering the uncertain material properties as fuzzy values. Results were presented as fuzzy plots showing the temperature ranges associated with different alpha cut values. The fuzzy finite element method was able to provide solutions that closely matched the traditional finite element method while incorporating uncertainty in the material properties.
On Continuous Approximate Solution of Ordinary Differential EquationsWaqas Tariq
In this work the problem of continuous approximate solution of the ordinary differential equations will be investigated. An approach to construct the continuous approximate solution, which is based on the discrete approximate solution and the spline interpolation, will be provided. The existence and uniqueness of such continuous approximate solution will be pointed out. Its error will be estimated and its convergence will be considered. Finally, with the aid of modern PC and nathematical software three practical computer approaches to perform above construction will be offered.
FUZZY IMAGE SEGMENTATION USING VALIDITY INDEXES CORRELATIONijcsit
This paper introduces an algorithm for image segmentation using a clustering technique; the technique is
based on the fuzzy c means algorithm (FCM) that is executed iteratively with different number of clusters.
Furthermore, simultaneously five validity indexes are calculated and their information is correlated to
determine the optimal number of clusters in order to segment an image, results and simulations are shown
in the paper.
Financial Time Series Analysis Based On Normalized Mutual Information FunctionsIJCI JOURNAL
A method of predictability analysis of future values of financial time series is described. The method is based on normalized mutual information functions. In the analysis, the use of these functions allowed to refuse any restrictions on the distributions of the parameters and on the correlations between parameters. A comparative analysis of the predictability of financial time series of Tel Aviv 25 stock exchange has been carried out.
This document presents two new Boolean combination methods called "orthogonalizing difference-building" and "orthogonal OR-ing". These methods calculate the difference and complement of functions, as well as the EXOR and EXNOR of minterms or functions, resulting in orthogonal forms. Orthogonal forms have advantages for further calculations. The methods are based on set theory and treat minterms and functions as ternary-vector lists. Equations are provided to define the difference and disjunction of minterms or functions in terms of their ternary-vector list representations.
Hybrid PSO-SA algorithm for training a Neural Network for ClassificationIJCSEA Journal
In this work, we propose a Hybrid particle swarm optimization-Simulated annealing algorithm and present a comparison with i) Simulated annealing algorithm and ii) Back propagation algorithm for training neural networks. These neural networks were then tested on a classification task. In particle swarm optimization behaviour of a particle is influenced by the experiential knowledge of the particle as well as socially exchanged information. Particle swarm optimization follows a parallel search strategy. In simulated annealing uphill moves are made in the search space in a stochastic fashion in addition to the downhill moves. Simulated annealing therefore has better scope of escaping local minima and reach a global minimum in the search space. Thus simulated annealing gives a selective randomness to the search. Back propagation algorithm uses gradient descent approach search for minimizing the error. Our goal of global minima in the task being done here is to come to lowest energy state, where energy state is being modelled as the sum of the squares of the error between the target and observed output values for all the training samples. We compared the performance of the neural networks of identical architectures trained by the i) Hybrid particle swarm optimization-simulated annealing, ii) Simulated annealing and iii) Back propagation algorithms respectively on a classification task and noted the results obtained. Neural network trained by Hybrid particle swarm optimization-simulated annealing has given better results compared to the neural networks trained by the Simulated annealing and Back propagation algorithms in the tests conducted by us.
Iterative Determinant Method for Solving Eigenvalue Problemsijceronline
International Journal of Computational Engineering Research (IJCER) is dedicated to protecting personal information and will make every reasonable effort to handle collected information appropriately. All information collected, as well as related requests, will be handled as carefully and efficiently as possible in accordance with IJCER standards for integrity and objectivity.
Matrix factorization techniques can be used to address some of the limitations of traditional collaborative filtering approaches for recommender systems. Matrix factorization decomposes the user-item rating matrix into the product of two lower-dimensional matrices, one representing latent factors for users and the other for items. This reduced dimensionality addresses data sparsity and scalability issues. Specifically, singular value decomposition is often used to perform this matrix factorization, which can approximate the original rating matrix while ignoring less important singular values and factor vectors. The decomposed matrices can then be multiplied to predict unknown user ratings.
Catching co occurrence information using word2vec-inspired matrix factorizationhyunsung lee
- Factorizing a PMI matrix involves using matrix factorization techniques to represent words in a latent space based on their co-occurrence information, similar to word2vec.
- Recommender systems use matrix factorization to represent users and items numerically in a latent space to predict target values like ratings. Known ratings are used to find latent vectors for users and items that best approximate the rating matrix.
- These latent vectors can then be used to predict unknown ratings by taking the dot product of the user and item vectors.
This document summarizes a study that used an adaptive neuro-fuzzy inference system (ANFIS) to model rainfall-runoff in the Nagwan watershed in India. Daily rainfall and runoff data from 1993-1999 were used to train ANFIS models with different combinations of rainfall and runoff as inputs and current day runoff as the output. The best performing model used only current day rainfall as the input and had 91 Gaussian membership functions. This model had low root mean square error and high correlation for both the training and testing periods. The study demonstrated that ANFIS is effective for hydrological rainfall-runoff modeling in small watersheds where runoff is highly dependent on current day rainfall.
This document proposes a neural collaborative filtering model called NeuMF that combines generalized matrix factorization and multi-layer perceptrons. It represents users and items as dense embedding vectors. NeuMF takes the concatenation of user and item embeddings as input and outputs a prediction through multiple nonlinear layers. The model is pretrained using GMF and MLP models and then jointly optimized. Experiments on MovieLens data show NeuMF achieves better performance than MF, MLP, and other baselines in terms of hit ratio and NDCG for top-K recommendation. It improves over SVD++ in terms of RMSE and MAE.
Information Theory and Programmable MedicineHector Zenil
This document is a slide presentation by Hector Zenil on information theory and programmable medicine. It discusses using information theory and a behavioral approach to computation to map how molecular stimuli relate to the conformational spaces of biological systems. It provides an example of simulating the self-assembly of porphyrin molecules under different binding energies to demonstrate how this approach could help program biological behaviors and inform treatments. The goal is to work towards programmable medicine by introducing molecular arrays that can release payloads based on certain biological conditions being met.
CHN and Swap Heuristic to Solve the Maximum Independent Set ProblemIJECEIAES
We describe a new approach to solve the problem to find the maximum independent set in a given Graph, known also as Max-Stable set problem (MSSP). In this paper, we show how Max-Stable problem can be reformulated into a linear problem under quadratic constraints, and then we resolve the QP result by a hybrid approach based Continuous Hopfeild Neural Network (CHN) and Local Search. In a manner that the solution given by the CHN will be the starting point of the local search. The new approach showed a good performance than the original one which executes a suite of CHN runs, at each execution a new leaner constraint is added into the resolved model. To prove the efficiency of our approach, we present some computational experiments of solving random generated problem and typical MSSP instances of real life problem.
Algorithmic Information Theory and Computational BiologyHector Zenil
I present cutting-edge concepts and tools drawn from algorithmic information theory (AIT) for new generation genetic sequencing, network biology and bioinformatics in general. AIT is the most advanced mathematical theory of information theory formally characterising the concepts and differences between simplicity, randomness and structure. Measures of AIT will empower computational medicine and systems biology to deal with big data, sophisticated analytics and a powerful new understanding framework.
PARTITION SORT REVISITED: RECONFIRMING THE ROBUSTNESS IN AVERAGE CASE AND MUC...IJCSEA Journal
In our previous work there was some indication that Partition Sort could be having a more robust average case O(nlogn) complexity than the popular Quick sort. In our first study in this paper, we reconfirm this through computer experiments for inputs from Cauchy distribution for which expectation theoretically does not exist. Additionally, the algorithm is found to be sensitive to parameters of the input probability distribution demanding further investigation on parameterized complexity. The results on this algorithm for Binomial inputs in our second study are very encouraging in that direction.
Partition Sort Revisited: Reconfirming the Robustness in Average Case and Muc...ijcsa
In our previous work there was some indication that Partition Sort could be having a more robust average case O(nlogn) complexity than the popular Quick sort. In our first study in this paper, we reconfirm this through computer experiments for inputs from Cauchy distribution for which expectation theoretically does not exist. Additionally, the algorithm is found to be sensitive to parameters of the input probability distribution demanding further investigation on parameterized complexity. The results on this algorithm for Binomial inputs in our second study are very encouraging in that direction.
A STATISTICAL COMPARATIVE STUDY OF SOME SORTING ALGORITHMSijfcstjournal
This research paper is a statistical comparative study of a few average case asymptotically optimal sorting
algorithms namely, Quick sort, Heap sort and K- sort. The three sorting algorithms all with the same
average case complexity have been compared by obtaining the corresponding statistical bounds while
subjecting these procedures over the randomly generated data from some standard discrete and continuous
probability distributions such as Binomial distribution, Uniform discrete and continuous distribution and
Poisson distribution. The statistical analysis is well supplemented by the parameterized complexity
analysis
This research paper is a statistical comparative study of a few average case asymptotically optimal sorting algorithms namely, Quick sort, Heap sort and K- sort. The three sorting algorithms all with the same average case complexity have been compared by obtaining the corresponding statistical bounds while subjecting these procedures over the randomly generated data from some standard discrete and continuous
probability distributions such as Binomial distribution, Uniform discrete and continuous distribution and Poisson distribution. The statistical analysis is well supplemented by the parameterized complexity analysis.
K-Sort: A New Sorting Algorithm that Beats Heap Sort for n 70 Lakhs!idescitation
Sundararajan and Chakraborty [10] introduced a new
version of Quick sort removing the interchanges. Khreisat
[6] found this algorithm to be competing well with some other
versions of Quick sort. However, it uses an auxiliary array
thereby increasing the space complexity. Here, we provide a
second version of our new sort where we have removed the
auxiliary array. This second improved version of the algorithm,
which we call K-sort, is found to sort elements faster than
Heap sort for an appreciably large array size (n 70,00,000)
for uniform U[0, 1] inputs.
Cauchy’s Inequality based study of the Differential Equations and the Simple ...IRJET Journal
1. The paper studies the multivariate generalization of Cauchy's inequality 1 + x ≤ ex, where x is a non-negative real number. This generalization can help solve certain ordinary differential equations (ODEs) and population dynamics problems.
2. The paper proves the multivariate generalization of the inequality and shows it only holds when the values are all equal to 0. It also analyzes some qualitative properties of solutions to ODE Cauchy problems using this generalization.
3. Different approaches are taken to directly prove the multivariate inequality using notions of monotone functions, Beppo Levi theorem, and divided differences mean value theorem. Allowed repetitions in the variables are also considered.
Exponential State Observer Design for a Class of Uncertain Chaotic and Non-Ch...ijtsrd
The document presents an exponential state observer design for uncertain chaotic and non-chaotic systems. It introduces a class of uncertain nonlinear systems and explores the state observation problem for such systems. Using an approach with integral and differential equalities, an exponential state observer is established that guarantees global exponential stability of the error system. Numerical simulations exhibit the feasibility and effectiveness of the observer design.
The International Journal of Engineering & Science is aimed at providing a platform for researchers, engineers, scientists, or educators to publish their original research results, to exchange new ideas, to disseminate information in innovative designs, engineering experiences and technological skills. It is also the Journal's objective to promote engineering and technology education. All papers submitted to the Journal will be blind peer-reviewed. Only original articles will be published.
The papers for publication in The International Journal of Engineering& Science are selected through rigorous peer reviews to ensure originality, timeliness, relevance, and readability.
New Approach of Preprocessing For Numeral RecognitionIJERA Editor
The present paper proposes a new approach of preprocessing for handwritten, printed and isolated numeral
characters. The new approach reduces the size of the input image of each numeral by discarding the redundant
information. This method reduces also the number of features of the attribute vector provided by the extraction
features method. Numeral recognition is carried out in this work through k nearest neighbors and multilayer
perceptron techniques. The simulations have obtained a good rate of recognition in fewer running time.
This paper introduces a new comparison base stable sorting algorithm, named RA sort. The RA sort
involves only the comparison of pair of elements in an array which ultimately sorts the array and does not
involve the comparison of each element with every other element. It tries to build upon the relationship
established between the elements in each pass. Instead of going for a blind comparison we prefer a
selective comparison to get an efficient method. Sorting is a fundamental operation in computer science.
This algorithm is analysed both theoretically and empirically to get a robust average case result. We have
performed its Empirical analysis and compared its performance with the well-known quick sort for various
input types. Although the theoretical worst case complexity of RA sort is Yworst(n) = O(n√), the
experimental results suggest an empirical Oemp(nlgn)1.333 time complexity for typical input instances, where
the parameter n characterizes the input size. The theoretical complexity is given for comparison operation.
We emphasize that the theoretical complexity is operation specific whereas the empirical one represents the
overall algorithmic complexity.
DESPECKLING OF SAR IMAGES BY OPTIMIZING AVERAGED POWER SPECTRAL VALUE IN CURV...ijistjournal
The document describes a novel algorithm for despeckling synthetic aperture radar (SAR) images using particle swarm optimization (PSO) in the curvelet domain. The algorithm first identifies homogeneous regions in the speckled image using variance calculations. It then uses PSO to optimize the thresholding of curvelet coefficients, with the objective of minimizing the average power spectral value. This provides an optimized threshold to apply curvelet-based despeckling. The proposed method is tested on standard images and shown to outperform conventional filters like median and Lee filters in reducing speckle noise.
DESPECKLING OF SAR IMAGES BY OPTIMIZING AVERAGED POWER SPECTRAL VALUE IN CURV...ijistjournal
Synthetic Aperture Radar (SAR) images are inherently affected by multiplicative speckle noise, due to the coherent nature of scattering phenomena. In this paper, a novel algorithm capable of suppressing speckle noise using Particle Swarm Optimization (PSO) technique is presented. The algorithm initially identifies homogenous region from the corrupted image and uses PSO to optimize the Thresholding of curvelet coefficients to recover the original image. Average Power Spectrum Value (APSV) has been used as objective function of PSO. The Proposed algorithm removes Speckle noise effectively and the performance of the algorithm is tested and compared with Mean filter, Median filter, Lee filter, Statistic Lee filter, Kuan filter, frost filter and gamma filter., outperforming conventional filtering methods.
FITTED OPERATOR FINITE DIFFERENCE METHOD FOR SINGULARLY PERTURBED PARABOLIC C...ieijjournal
In this paper, we study the numerical solution of singularly perturbed parabolic convection-diffusion type
with boundary layers at the right side. To solve this problem, the backward-Euler with Richardson
extrapolation method is applied on the time direction and the fitted operator finite difference method on the
spatial direction is used, on the uniform grids. The stability and consistency of the method were established
very well to guarantee the convergence of the method. Numerical experimentation is carried out on model
examples, and the results are presented both in tables and graphs. Further, the present method gives a more
accurate solution than some existing methods reported in the literature.
FITTED OPERATOR FINITE DIFFERENCE METHOD FOR SINGULARLY PERTURBED PARABOLIC C...ieijjournal
In this paper, we study the numerical solution of singularly perturbed parabolic convection-diffusion type
with boundary layers at the right side. To solve this problem, the backward-Euler with Richardson
extrapolation method is applied on the time direction and the fitted operator finite difference method on the
spatial direction is used, on the uniform grids. The stability and consistency of the method were established
very well to guarantee the convergence of the method. Numerical experimentation is carried out on model
examples, and the results are presented both in tables and graphs. Further, the present method gives a more
accurate solution than some existing methods reported in the literature.
The document analyzes the use of the Tent map as a source of pseudorandom bits for generating binary codes. It evaluates the Tent map's period length, discrimination value, and merit factor. The Tent map is proposed as an alternative to traditional low-complexity pseudorandom bit generators. Different window functions are applied to the binary codes generated from the Tent map to reduce side lobes and improve performance. Results show discrimination increases with sequence length and some window functions perform better than others at different lengths.
One of the fundamental issues in computer science is ordering a list of items. Although there is a number of sorting algorithms, sorting problem has attracted a great deal of research, because efficient sorting is important to optimize the use of other algorithms. This paper presents a new sorting algorithm which sort the elements based on their average, which runs faster.. This algorithm was analyzed, implemented and tested and the results are promising for a random data
Parametric Sensitivity Analysis of a Mathematical Model of Two Interacting Po...IOSR Journals
Experts in the mathematical modeling for two interacting technologies have observed the different contributions between the intraspecific and the interspecific coefficients in conjunction with the starting population sizes and the trading period. In this complex multi-parameter system of competing technologies which evolve over time, we have used the numerical method of mathematical norms to measure the sensitivity values of the intraspecific coefficients b and e, the starting population sizes of the two interacting technologies and the duration of trading. We have observed that the two intraspecific coefficients can be considered as most sensitive parameter while the starting populations are called least sensitive. We will expect these contributions to provide useful insights in the determination of the important parameters which drive the dynamics of the technological substitution model in the context of one-at-a-timesensitivity analysis
Multivariate Analysis of Cauchy’s InequalityIRJET Journal
This paper investigates the multivariate generalization of Cauchy's inequality 1 + x ≤ ex, where x is any non-negative real number. Specifically, it aims to prove the inequality (1 + x1)(1 + x2)...(1 + xn) ≤ e(x1+x2+...+xn), where x1, x2, ..., xn are pairwise distinct non-negative real numbers. The proof is based on notions from empty product conventions and Beppo Levi's theorem of monotone convergence. This inequality is also extended to simultaneous inequalities and its relationship to ordinary differential equation Cauchy problems and population dynamics is explored. Direct approaches using definitions of monotone functions and mean value theore
Inversion of Magnetic Anomalies Due to 2-D Cylindrical Structures – By an Art...ijsc
Application of Artificial Neural Network Committee Machine (ANNCM) for the inversion of magnetic anomalies caused by a long-2D horizontal circular cylinder is presented. Although, the subsurface targets are of arbitrary shape, they are assumed to be regular geometrical shape for convenience of mathematical analysis. ANNCM inversion extract the parameters of the causative subsurface targets include depth to the centre of the cylinder (Z), the inclination of magnetic vector(Ɵ)and the constant term (A)comprising the radius(R)and the intensity of the magnetic field(I). The method of inversion is demonstrated over a theoretical model with and without random noise in order to study the effect of noise on the technique and then extended to real field data. It is noted that the method under discussion ensures fairly accurate results even in the presence of noise. ANNCM analysis of vertical magnetic anomaly near Karimnagar, Telangana, India, has shown satisfactory results in comparison with other inversion techniques that are in vogue.The statistics of the predicted parameters relative to the measured data, show lower sum error (<9.58%) and higher correlation coefficient (R>91%) indicating that good matching and correlation is achieved between the measured and predicted parameters.
INVERSIONOF MAGNETIC ANOMALIES DUE TO 2-D CYLINDRICAL STRUCTURES –BY AN ARTIF...ijsc
Application of Artificial Neural Network Committee Machine (ANNCM) for the inversion of magnetic
anomalies caused by a long-2D horizontal circular cylinder is presented. Although, the subsurface targets
are of arbitrary shape, they are assumed to be regular geometrical shape for convenience of mathematical
analysis. ANNCM inversion extract the parameters of the causative subsurface targets include depth to the
centre of the cylinder (Z), the inclination of magnetic vector(Ɵ)and the constant term (A)comprising the
radius(R)and the intensity of the magnetic field(I). The method of inversion is demonstrated over a
theoretical model with and without random noise in order to study the effect of noise on the technique and
then extended to real field data. It is noted that the method under discussion ensures fairly accurate results
even in the presence of noise. ANNCM analysis of vertical magnetic anomaly near Karimnagar, Telangana,
India, has shown satisfactory results in comparison with other inversion techniques that are in vogue.The
statistics of the predicted parameters relative to the measured data, show lower sum error (<9.58%) and
higher correlation coefficient (R>91%) indicating that good matching and correlation is achieved between
the measured and predicted parameters.
Similar to BINARY TREE SORT IS MORE ROBUST THAN QUICK SORT IN AVERAGE CASE (20)
Levelised Cost of Hydrogen (LCOH) Calculator ManualMassimo Talia
The aim of this manual is to explain the
methodology behind the Levelized Cost of
Hydrogen (LCOH) calculator. Moreover, this
manual also demonstrates how the calculator
can be used for estimating the expenses associated with hydrogen production in Europe
using low-temperature electrolysis considering different sources of electricity
Digital Twins Computer Networking Paper Presentation.pptxaryanpankaj78
A Digital Twin in computer networking is a virtual representation of a physical network, used to simulate, analyze, and optimize network performance and reliability. It leverages real-time data to enhance network management, predict issues, and improve decision-making processes.
Road construction is not as easy as it seems to be, it includes various steps and it starts with its designing and
structure including the traffic volume consideration. Then base layer is done by bulldozers and levelers and after
base surface coating has to be done. For giving road a smooth surface with flexibility, Asphalt concrete is used.
Asphalt requires an aggregate sub base material layer, and then a base layer to be put into first place. Asphalt road
construction is formulated to support the heavy traffic load and climatic conditions. It is 100% recyclable and
saving non renewable natural resources.
With the advancement of technology, Asphalt technology gives assurance about the good drainage system and with
skid resistance it can be used where safety is necessary such as outsidethe schools.
The largest use of Asphalt is for making asphalt concrete for road surfaces. It is widely used in airports around the
world due to the sturdiness and ability to be repaired quickly, it is widely used for runways dedicated to aircraft
landing and taking off. Asphalt is normally stored and transported at 150’C or 300’F temperature
Null Bangalore | Pentesters Approach to AWS IAMDivyanshu
#Abstract:
- Learn more about the real-world methods for auditing AWS IAM (Identity and Access Management) as a pentester. So let us proceed with a brief discussion of IAM as well as some typical misconfigurations and their potential exploits in order to reinforce the understanding of IAM security best practices.
- Gain actionable insights into AWS IAM policies and roles, using hands on approach.
#Prerequisites:
- Basic understanding of AWS services and architecture
- Familiarity with cloud security concepts
- Experience using the AWS Management Console or AWS CLI.
- For hands on lab create account on [killercoda.com](https://killercoda.com/cloudsecurity-scenario/)
# Scenario Covered:
- Basics of IAM in AWS
- Implementing IAM Policies with Least Privilege to Manage S3 Bucket
- Objective: Create an S3 bucket with least privilege IAM policy and validate access.
- Steps:
- Create S3 bucket.
- Attach least privilege policy to IAM user.
- Validate access.
- Exploiting IAM PassRole Misconfiguration
-Allows a user to pass a specific IAM role to an AWS service (ec2), typically used for service access delegation. Then exploit PassRole Misconfiguration granting unauthorized access to sensitive resources.
- Objective: Demonstrate how a PassRole misconfiguration can grant unauthorized access.
- Steps:
- Allow user to pass IAM role to EC2.
- Exploit misconfiguration for unauthorized access.
- Access sensitive resources.
- Exploiting IAM AssumeRole Misconfiguration with Overly Permissive Role
- An overly permissive IAM role configuration can lead to privilege escalation by creating a role with administrative privileges and allow a user to assume this role.
- Objective: Show how overly permissive IAM roles can lead to privilege escalation.
- Steps:
- Create role with administrative privileges.
- Allow user to assume the role.
- Perform administrative actions.
- Differentiation between PassRole vs AssumeRole
Try at [killercoda.com](https://killercoda.com/cloudsecurity-scenario/)
Prediction of Electrical Energy Efficiency Using Information on Consumer's Ac...PriyankaKilaniya
Energy efficiency has been important since the latter part of the last century. The main object of this survey is to determine the energy efficiency knowledge among consumers. Two separate districts in Bangladesh are selected to conduct the survey on households and showrooms about the energy and seller also. The survey uses the data to find some regression equations from which it is easy to predict energy efficiency knowledge. The data is analyzed and calculated based on five important criteria. The initial target was to find some factors that help predict a person's energy efficiency knowledge. From the survey, it is found that the energy efficiency awareness among the people of our country is very low. Relationships between household energy use behaviors are estimated using a unique dataset of about 40 households and 20 showrooms in Bangladesh's Chapainawabganj and Bagerhat districts. Knowledge of energy consumption and energy efficiency technology options is found to be associated with household use of energy conservation practices. Household characteristics also influence household energy use behavior. Younger household cohorts are more likely to adopt energy-efficient technologies and energy conservation practices and place primary importance on energy saving for environmental reasons. Education also influences attitudes toward energy conservation in Bangladesh. Low-education households indicate they primarily save electricity for the environment while high-education households indicate they are motivated by environmental concerns.
Generative AI Use cases applications solutions and implementation.pdfmahaffeycheryld
Generative AI solutions encompass a range of capabilities from content creation to complex problem-solving across industries. Implementing generative AI involves identifying specific business needs, developing tailored AI models using techniques like GANs and VAEs, and integrating these models into existing workflows. Data quality and continuous model refinement are crucial for effective implementation. Businesses must also consider ethical implications and ensure transparency in AI decision-making. Generative AI's implementation aims to enhance efficiency, creativity, and innovation by leveraging autonomous generation and sophisticated learning algorithms to meet diverse business challenges.
https://www.leewayhertz.com/generative-ai-use-cases-and-applications/
Blood finder application project report (1).pdfKamal Acharya
Blood Finder is an emergency time app where a user can search for the blood banks as
well as the registered blood donors around Mumbai. This application also provide an
opportunity for the user of this application to become a registered donor for this user have
to enroll for the donor request from the application itself. If the admin wish to make user
a registered donor, with some of the formalities with the organization it can be done.
Specialization of this application is that the user will not have to register on sign-in for
searching the blood banks and blood donors it can be just done by installing the
application to the mobile.
The purpose of making this application is to save the user’s time for searching blood of
needed blood group during the time of the emergency.
This is an android application developed in Java and XML with the connectivity of
SQLite database. This application will provide most of basic functionality required for an
emergency time application. All the details of Blood banks and Blood donors are stored
in the database i.e. SQLite.
This application allowed the user to get all the information regarding blood banks and
blood donors such as Name, Number, Address, Blood Group, rather than searching it on
the different websites and wasting the precious time. This application is effective and
user friendly.
Tools & Techniques for Commissioning and Maintaining PV Systems W-Animations ...Transcat
Join us for this solutions-based webinar on the tools and techniques for commissioning and maintaining PV Systems. In this session, we'll review the process of building and maintaining a solar array, starting with installation and commissioning, then reviewing operations and maintenance of the system. This course will review insulation resistance testing, I-V curve testing, earth-bond continuity, ground resistance testing, performance tests, visual inspections, ground and arc fault testing procedures, and power quality analysis.
Fluke Solar Application Specialist Will White is presenting on this engaging topic:
Will has worked in the renewable energy industry since 2005, first as an installer for a small east coast solar integrator before adding sales, design, and project management to his skillset. In 2022, Will joined Fluke as a solar application specialist, where he supports their renewable energy testing equipment like IV-curve tracers, electrical meters, and thermal imaging cameras. Experienced in wind power, solar thermal, energy storage, and all scales of PV, Will has primarily focused on residential and small commercial systems. He is passionate about implementing high-quality, code-compliant installation techniques.
Supermarket Management System Project Report.pdfKamal Acharya
Supermarket management is a stand-alone J2EE using Eclipse Juno program.
This project contains all the necessary required information about maintaining
the supermarket billing system.
The core idea of this project to minimize the paper work and centralize the
data. Here all the communication is taken in secure manner. That is, in this
application the information will be stored in client itself. For further security the
data base is stored in the back-end oracle and so no intruders can access it.
This study Examines the Effectiveness of Talent Procurement through the Imple...DharmaBanothu
In the world with high technology and fast
forward mindset recruiters are walking/showing interest
towards E-Recruitment. Present most of the HRs of
many companies are choosing E-Recruitment as the best
choice for recruitment. E-Recruitment is being done
through many online platforms like Linkedin, Naukri,
Instagram , Facebook etc. Now with high technology E-
Recruitment has gone through next level by using
Artificial Intelligence too.
Key Words : Talent Management, Talent Acquisition , E-
Recruitment , Artificial Intelligence Introduction
Effectiveness of Talent Acquisition through E-
Recruitment in this topic we will discuss about 4important
and interlinked topics which are
FULL STACK PROGRAMMING - Both Front End and Back End
BINARY TREE SORT IS MORE ROBUST THAN QUICK SORT IN AVERAGE CASE
1. International Journal of Computer Science, Engineering and Applications (IJCSEA) Vol.2, No.2, April 2012
DOI : 10.5121/ijcsea.2012.2209 115
BINARY TREE SORT IS MORE ROBUST
THAN QUICK SORT IN AVERAGE CASE
Soubhik Chakraborty*1
and Rajarshi Sarkar2
1
Department of Applied Mathematics, BIT Mesra, Ranchi-835215, India
2
Department of Computer Science & Engineering , BIT Mesra, Ranchi-835215, India
Email of the corresponding author: soubhikc@yahoo.co.in
ABSTRACT
Average case complexity, in order to be a useful and reliable measure, has to be robust. The probability
distribution, generally uniform, over which expectation is taken should be realistic over the problem
domain. But algorithm books do not certify that the uniform inputs are always realistic. Do the results hold
even for non uniform inputs? In this context we observe that Binary Tree sort is more robust than the fast
and popular Quick sort in the average case.
KEY WORDS
Binary Tree sort, robustness, expectation, probability distribution, average case complexity
1. Introduction
Average Case complexity is both interesting and intriguing part of algorithm analysis. It is
interesting as some algorithms with bad worst case complexity performs better on the average
like Quick sort and Binary Tree sort, the latter being investigated here. It is intriguing as for a
complex code it is difficult to identify the pivotal operation or the pivotal region in the code for
taking mathematical expectation for finding the average case complexity. Moreover, the
probability distribution over which expectation is taken (which is generally uniform) may not be
realistic over the domain of the problem. For comparison based sorting algorithms, the dominant
operation turns out to be comparison. But algorithm books do not certify that the uniform inputs
are always realistic. It is quite possible that the inputs come from non uniform distributions such
as Binomial. Do the results derived assuming uniform inputs hold even for non uniform inputs?
In other words, is the average case complexity having a robustness appeal? Quick sort
unfortunately is not very robust (Sourabh and Chakraborty [1]). But our study shows that Binary
Tree sort which has average and worst case complexity similar to those of Quick sort, namely
O(nlogn) and O(n2
) respectively, is quite robust . Using computer experiments, a series of runs of
a code for various inputs, we verify that the model y = a+bx+error where x = nlog2n and y is
average (mean) sorting time, does hold remarkably well even for non uniform inputs like
Binomial, Poisson and Standard Normal. For a comprehensive literature on sorting, we refer to
2. International Journal of Computer Science, Engineering and Applications (IJCSEA) Vol.2, No.2, April 2012
116
Knuth[2]. For a sound theoretical discussion on sorting with reference to the input probability
distribution, Mahmoud[3] should be consulted. The next section gives the algorithm of Binary
Tree sort. In section 3 we provide the statistical analysis. Section 4 gives a brief discussion on the
results. Section 5 is reserved for conclusion.
2. Algorithm for Binary Tree Sort[4][5]:
Fig.1.Binary Search Tree
a) In order to build the binary search tree, we begin with the 0th
element 16 which we take
as the root of the tree.
b) While inserting the 1st
element, i.e. 7, 7 is compared with its root element 16 and being
less than 16 it is made the left child of the root node 16.
c) While inserting the 2nd
element of the list, i.e. 18, 18 is compared with the root element
16.But 18 is greater than 16, so it is made the right child of the root node 16.
d) Proceeding in a similar fashion, all the elements are appropriately placed in the binary
search tree.
e) Now to get the elements in the sorted order, the tree is traversed in in-order. The in-order
traversal of the binary search tree lists the elements in ascending order.
In-order traverse:
1) Traverse the left subtree of the root R in in-order.
2) Process the root R.
3) Traverse the right subtree of R in in-order.
16
7 18
14
95
22
626
8 30
3. International Journal of Computer Science, Engineering and Applications (IJCSEA) Vol.2, No.2, April 2012
117
3. Statistical Analysis
Our experimental work deals with certain empirical results in Binary tree sort algorithm. Here we
make a statistical case study on the robustness of average complexity measures, which are
derived assuming uniform distribution, for non-uniform inputs(both discrete and continuous).For
simulating different variates, we consulted Kennedy and Gentle[6].
We have done a robustness study of Binary tree sort algorithm to verify whether the O(nlogn)
complexity holds in the average case if the input comes from distributions other than uniform. We
primarily include Monte Carlo Simulation (where inputs from different probability distributions
will be simulated) and statistical modeling using simple linear regression with the size of input as
the predictor and average sorting time as the response variable. We are taking 100 trials for
calculating the average time of sorting different discrete and continuous distribution inputs of
sample size n. The observation have been taken on fixed parameters for different distribution, but
with varying sample size n in the range [5000, 50000]. The standard non uniform probability
distributions investigated are Binomial, Poisson and Normal, the first two being discrete while the
last one is continuous. We will of course be investigating discrete uniform and continuous
uniform distributions also for the sake of completeness.
System Specification:
Processor : Intel(R) Celeron(R) CPU 2.13GHz
Hard Disk : 80GB
RAM : 256MB
Operating System : Windows XP
3.1 Average Case Complexity of Binary Tree Sort for Binomial Distribution Inputs:
The Binomial distribution Binomial (m, p) inputs are taken with parameters m and p, where
m=100 and p=0.5 are fixed. The empirical results are shown in table 1 and fig. 2.
Table 1: Table of mean sorting time and standard deviation (both in seconds) of Binomial distribution
inputs for Binary tree sort
N n logn Mean time SD
5000 18494.85 0.0086 0.0078
10000 40000.00 0.0140 0.0139
15000 62641.37 0.0273 0.0146
20000 86020.60 0.0389 0.0234
25000 109948.50 0.0483 0.0279
30000 134313.64 0.0542 0.0326
35000 159042.38 0.0602 0.0385
40000 184082.40 0.0691 0.0402
45000 209394.56 0.0742 0.0475
50000 234948.50 0.0825 0.0502
4. International Journal of Computer Science, Engineering and Applications (IJCSEA) Vol.2, No.2, April 2012
118
Fig. 2 Plot showing empirical O(nlogn) complexity for binomial inputs
Experimental results as shown in fig.2 are supporting O(n logn) complexity. We write
yavg(n)=Oemp(nlogn).
3.2. Average Case Complexity for Poisson Distribution Inputs:
The Poisson distribution Poisson (µ) inputs are taken with parameter µ where µ=1 is fixed. The
empirical results are shown in table 2 and fig. 3.
Table 2: Table of mean sorting time and standard deviation (both in sec.) of Poisson
distribution inputs for Binary tree sort
n n logn Mean time SD
5000 18494.85 0.0032 0.0064
10000 40000.00 0.0048 0.0073
15000 62641.37 0.0158 0.0121
20000 86020.60 0.0220 0.0159
25000 109948.50 0.0314 0.0139
30000 134313.64 0.0346 0.0251
35000 159042.38 0.0391 0.0199
40000 184082.40 0.0436 0.0250
45000 209394.56 0.0543 0.0291
50000 234948.50 0.0591 0.0302
5. International Journal of Computer Science, Engineering and Applications (IJCSEA) Vol.2, No.2, April 2012
119
Fig.3 Plot showing empirical O(nlogn) complexity for Poisson inputs
Experimental results as shown in above fig. 3 are supporting O(n logn) complexity .We write yavg
(n)=Oemp(nlogn).
3.3. Average Case Complexity for Discrete Uniform Distribution Inputs:
The Discrete Uniform [1, 2, 3….., k] inputs are taken with parameter k where k=50 is fixed. The
empirical results are shown in table 3 and fig.4.
Table 3: Table of mean sorting time and standard deviation (both in sec.) of Discrete
Uniform distribution inputs for Binary tree sort
N n logn Mean time SD
5000 18494.85 0.0031 0.0062
10000 40000.00 0.0110 0.0072
15000 62641.37 0.0124 0.0116
20000 86020.60 0.0190 0.0154
25000 109948.50 0.0217 0.0173
30000 134313.64 0.0297 0.0176
35000 159042.38 0.0296 0.0202
40000 184082.40 0.0390 0.0234
45000 209394.56 0.0564 0.0418
50000 234948.50 0.0602 0.0378
6. International Journal of Computer Science, Engineering and Applications (IJCSEA) Vol.2, No.2, April 2012
120
Fig. 4 Plot showing empirical O(nlogn) complexity for Discrete Uniform inputs
Experimental results as shown in above fig.4 are again supporting O(n log n) complexity. We
write yavg (n)=Oemp(nlogn).
3.4. Average Case Complexity for Continuous Uniform Distribution Inputs:
The Continuous Uniform distribution Continuous Uniform [0, θ] inputs are taken with parameter
mean θ,where θ =1 is fixed. The empirical results are shown in table 4 and fig.5.
Table 4: Table of mean sorting time and standard deviation (both in sec.) of Continuous Uniform
distribution inputs for Binary tree sort
N n logn Mean time SD
5000 18494.85 0.0031 0.0062
10000 40000.00 0.0080 0.0080
15000 62641.37 0.0173 0.0109
20000 86020.60 0.0201 0.0099
25000 109948.50 0.0266 0.0172
30000 134313.64 0.0313 0.0157
35000 159042.38 0.0375 0.0224
40000 184082.40 0.0453 0.0256
45000 209394.56 0.0484 0.0284
50000 234948.50 0.0545 0.0296
7. International Journal of Computer Science, Engineering and Applications (IJCSEA) Vol.2, No.2, April 2012
121
Fig. 5. Plot showing empirical O(nlogn) complexity for Continuous Uniform inputs
Experimental results as shown in above fig.5 are supporting O(n log n) complexity. We write yavg
(n)=Oemp(nlogn).
3.5. Average Case Complexity for Standard Normal Distribution Inputs:
For the Standard Normal distribution inputs, the empirical results are shown in table 5 and fig.6.
Table 5: Table of mean sorting time and standard deviation (both in sec.) of Standard Normal distribution
inputs for Binary tree sort
N n logn Mean time SD
5000 18494.85 0.0000 0.0000
10000 40000.00 0.0062 0.0076
15000 62641.37 0.0064 0.0078
20000 86020.60 0.0064 0.0078
25000 109948.50 0.0092 0.0075
30000 134313.64 0.0115 0.0067
35000 159042.38 0.0150 0.0000
40000 184082.40 0.0156 0.0004
45000 209394.56 0.0175 0.0065
50000 234948.50 0.0185 0.0056
8. International Journal of Computer Science, Engineering and Applications (IJCSEA) Vol.2, No.2, April 2012
122
Fig.6 Plot showing empirical O(nlogn) complexity for Standard Normal inputs.
Experimental results as shown in above fig.6 are supporting O(n log n) complexity. We write yavg
(n)=Oemp(nlogn).
Remark: Box Muller transformation was used for simulating standard normal variates (Kennedy
and Gentle [6]).
4. Discussion
There are a few points to be discussed. There is an important result which says average case
complexity under the universal distribution equals worst case complexity [7]. So for those
algorithms whose worst case complexity is different (higher or worse) than average case
complexity, it is risky to say “this algorithm has a bad worst case but performs better on the
average” without verifying the robustness. Since Binary tree sort is robust while Quick sort is not,
we can use Binary tree sort in case there is a possibility that the sorting elements need not be
uniform. The other important question is: why did we work on time which is system dependent?
The answer to this question is that time of a computing operation is actually the weight (and not
the count) of the operation and time of the entire code is a weighted sum of the computing
operations, the weights being the corresponding times. A mathematical complexity bound (like
big oh) is count based and operation specific. A bound that is weight based and consequently
allows for mixing of different types of operations (like mixing comparisons with interchanges in
sorting) is essentially a statistical bound. Empirical O, written as O with a subscript emp, is only
an empirical estimate of the statistical bound obtained over a finite range by supplying numerical
values to the weights. And these numerical values are obtained by running computer experiments.
So design and analysis of computer experiments become crucial especially when the response is a
complexity such as time. For a general literature on computer experiments we refer to the seminal
paper by Sacks et. al. [8] and the book by Fang, Li and Sudjianto [9]. A recent book which gives
a computer experiment outlook to algorithmic complexity is by Chakraborty and Sourabh [10].
This is the first published book on statistical complexity bounds.
9. International Journal of Computer Science, Engineering and Applications (IJCSEA) Vol.2, No.2, April 2012
123
5. Conclusion
The robustness of average case measure O(nlogn) for Binary tree sort has be verified by
observing mean time. This may be crossed checked against the mean comparisons experimentally
counted rather than working on time directly. Only for algorithms such as Heap sort where worst
and average complexities are equal, it is safe to rely on an average complexity measure. In case
average case complexity seems to be better than the same for worst case, the robustness must be
confirmed first as otherwise it could be a dangerous practice to rely on it. The recent rejection of
some proofs (see [11] in addition to [1] for example) is testimony to this fact. And this is
precisely where Binary Tree sorts beats the fast and popular Quick sort.
References
[1] S. K. Sourabh and S. Chakraborty, How robust is quicksort average complexity? arXiv:0811.4376v1
[cs.DS]
[2] D. E. Knuth, The Art of Programming, Vol. 3: Sorting and Searching, Pearson Edu. reprint, 2000
[3] H. Mahmoud, Sorting: A Distribution Theory, John Wiley, 2000
[4] Y. Kanetkar, Data Structures through C, BPB publications, New Delhi, 2nd Ed.,2009
[5] S. Lipschutz (Adapted by G. A. V. Pai), Data Structures, Schaum’s Outlines Series, Tata McGraw-
Hill Publishing Company Ltd. 2006.
[6] W. Kennedy and J. Gentle, Statistical Computing, Marcel Dekker Inc., 1980
[7] .M. Li and P.M.B Vitanyi, Average case complexity under the universal distribution equals worst
case complexity, Inf . Proc. Lett., 42,no.3,1992,145-149
[8] Sacks, J., W. Welch, T. Mitchell and H. Wynn, Design and Analysis of Computer Experiments,
Statistical Science, vol. 4(4), 1989
[9] Fang, K. T., Li, R. and Sudjianto, A., Design and Modeling of Computer Experiments, Chapman and
Hall, 2006
[10] Chakraborty, S. and Sourabh, S. K., A Computer Experiment Oriented Approach to
Algorithmic Complexity, Lambert Academic Publishing, 2010
[11] S. Chakraborty, S. K. Sourabh, How Robust are Average Complexity Measures? A Statistical Case
Study, Applied Math. and Compu., vol. 189(2), 2007, 1787-1797