Logic coverage criteria are central aspect of programs and specifications in the testing of software. A
Boolean expression with n variables in expression 2n distinct test cases is required for exhaustive testing
.This is expensive even when n is modestly large. The possible solution is to select a small subset of all
possible test cases which can effectively detect most of the common faults. Test case prioritization is one of
the key techniques for making testing cost effective. In present study performance index of test suite is
calculated for two Boolean specification testing techniques MUMCUT and Minimal-MUMCUT.
Performance index helps to measure the efficiency and determine when testing can be stopped in case of
limited resources. This paper evaluates the testability of generated single faults according to the number of
test cases used to detect them. Test cases are generated from logic expressions in irredundant normal form
(IDNF) derived from specifications or source code. The efficiency of prioritization techniques has been
validated by an empirical study done on bench mark expressions using Performance Index (PI) metric.
This document presents an adaptive model switching technique for variable fidelity optimization using population-based algorithms. The technique aims to provide reliable high-fidelity optimum designs with reasonable computational expense by leveraging multiple models of varying fidelity. It switches models by comparing the error distribution of the current model to the distribution of recent fitness function improvements over the population. The method was tested on airfoil and cantilever beam design problems, showing substantially better balance of optimum quality and efficiency than purely low- or high-fidelity optimizations.
Taguchi design of experiments nov 24 2013Charlton Inao
This document provides an overview of Taguchi design of experiments. It defines Taguchi methods, which use orthogonal arrays to conduct a minimal number of experiments that can provide full information on factors that affect performance. The assumptions of Taguchi methods include additivity of main effects. The key steps in an experiment are selecting variables and their levels, choosing an orthogonal array, assigning variables to columns, conducting experiments, and analyzing data through sensitivity analysis and ANOVA.
Surrogate-based design is an effective approach for modeling computationally expensive system behavior. In such application, it is often challenging to characterize the expected accuracy of the surrogate. In addition to global and local error measures, regional error measures can be used to understand and interpret the surrogate accuracy in the regions of interest. This paper develops the Regional Error Estimation of Surrogate (REES) method to quantify the level of the error in any given subspace (or region) of the entire domain, when all the available training points have been invested to build the surrogate. In this approach, the accuracy of the surrogate in each subspace is estimated by modeling the variations of the mean and the maximum error in that subspace with increasing number of training points (in an iterative process). A regression model is used for this purpose. At each iteration, the intermediate surrogate is constructed using a subset of the entire training data, and tested over the remaining points. The evaluated errors at the intermediate test points at each iteration are used for training the regression model that represents the error variation with sample points. The effectiveness of the proposed method is illustrated using standard test problems. To this end, the predicted regional errors of the surrogate constructed using all the training points are compared with the regional errors estimated over a large set of test points.
Model based test case prioritization using neural network classificationcseij
Model-based testing for real-life software systems often require a large number of tests, all of which cannot
exhaustively be run due to time and cost constraints. Thus, it is necessary to prioritize the test cases in
accordance with their importance the tester perceives. In this paper, this problem is solved by improving
our given previous study, namely, applying classification approach to the results of our previous study
functional relationship between the test case prioritization group membership and the two attributes:
important index and frequency for all events belonging to given group are established. A for classification
purpose, neural network (NN) that is the most advances is preferred and a data set obtained from our study
for all test cases is classified using multilayer perceptron (MLP) NN. The classification results for
commercial test prioritization application show the high classification accuracies about 96% and the
acceptable test prioritization performances are achieved.
This paper advances the Domain Segmentation based on Uncertainty in the Surrogate (DSUS) framework which is a novel approach to characterize the uncertainty in surrogates. The leave-one-out cross-validation technique is adopted in the DSUS framework to measure local errors of a surrogate. A method is proposed in this paper to evaluate the performance of the leave-out-out cross-validation errors as local error measures. This method evaluates local errors by comparing: (i) the leave-one-out cross-validation error with (ii) the actual local error estimated within a local hypercube for each training point. The comparison results show that the leave-one-out cross-validation strategy can capture the local errors of a surrogate. The DSUS framework is then applied to key aspects of wind resource as- sessment and wind farm cost modeling. The uncertainties in the wind farm cost and the wind power potential are successfully characterized, which provides designers/users more confidence when using these models
This document presents a new 3-level approach to simultaneously select the best surrogate model type, kernel function, and hyper-parameters for approximation models. The approach uses Regional Error Estimation of Surrogates (REES) to evaluate error and select models. It compares a cascaded technique that performs sequential optimization versus a one-step technique. Numerical examples on benchmark problems show the one-step technique reduces maximum and median errors by at least 60% with lower computational cost compared to the cascaded approach. Future work involves applying the one-step method to more complex problems and developing an online platform for collaborative surrogate model selection.
The analysis of complex system behavior often demands expensive experiments or computational simulations. Surrogates modeling techniques are often used to provide a tractable and inexpensive approximation of such complex system behavior. Owing to the lack of knowledge regarding the suitability of particular surrogate modeling techniques, model selection approach can be helpful to choose the best surrogate technique. Popular model selection approaches include: (i) split sample, (ii) cross-validation, (iii) bootstrapping, and (iv) Akaike's information criterion (AIC) (Queipo et al. 2005; Bozdogan et al. 2000). However, the effectiveness of these model selection methods is limited by the lack of accurate measures of local and global errors in surrogates.
This paper develops a novel and model-independent concept to quantify the local/global reliability of surrogates, to assist in model selection (in surrogate applications). This method is called the Generalized-Regional Error Estimation of Surrogate (G-REES). In this method, intermediate surrogates are iteratively constructed over heuristic subsets of the available sample points (i.e., intermediate training points), and tested over the remaining available sample points (i.e., intermediate test points). The fraction of sample points used as intermediate training points is fixed at each iteration, with the total number of iterations being pre-specified. The estimated median and maximum relative errors for the heuristic subsets at each iteration are used to fit a distribution of the median and maximum error, respectively. The statistical mode of the median and the maximum error distributions are then determined. These mode values are then represented as functions of the density of training points (at the corresponding iteration). Regression methods, called Variation of Error with Sample Density (VESD), are used for this purpose. The VESD models are then used to predict the expected median and maximum errors, when all the sample points are used as training points.
The effectiveness of the proposed model selection criterion is explored to find the best surrogate between candidates including: (i) Kriging, (ii) Radial Basis Functions (RBF), (iii) Extended Radial Basis Functions (ERBF), and (iv) Quadratic Response Surface (QRS), for standard test functions and a wind farm capacity factor function. The results will be compared with the relative accuracy of the surrogates evaluated on additional test points, and also with the prediction sum of square (PRESS) error given by leave-one-out cross-validation.
The application of G-REES to a standard test problem with two design variables (Branin-hoo function) show that the proposed method predicts the median and the maximum value of the global error with a higher level of confidence compared to PRESS. It also shows that model selection based on G-REES method is significantly more reliable than that currently performed using error measures such as PRESS. The
BOOLEAN SPECIFICATION BASED TESTING TECHNIQUES: A SURVEYcscpconf
Boolean expressions are major focus of specifications and they are very much prone to
introduction of faults, this survey presents various Boolean specification based testing
techniques, and covers more than 30 papers for the same. The various Boolean specification
based testing techniques like Cause effect graph, fosters strategy, meaningful impact strategy,
Branch Operator Strategy (BOR), Modified Condition/ Decision Coverage (MCDC) compared
on the basis of their fault detection effectiveness and the size of test suite. This collection
represents most of the existing work performed on Boolean specification based testing
techniques. This survey describes the basic algorithms used by these strategies and it also
includes operator and operand fault categories for evaluating the performance of above mentioned testing techniques. Finally, this survey contains short summaries of all the papers that use Boolean specification based testing techniques. These techniques have been empirically evaluated by various researchers on a simplified safety related real time control system.
This document presents an adaptive model switching technique for variable fidelity optimization using population-based algorithms. The technique aims to provide reliable high-fidelity optimum designs with reasonable computational expense by leveraging multiple models of varying fidelity. It switches models by comparing the error distribution of the current model to the distribution of recent fitness function improvements over the population. The method was tested on airfoil and cantilever beam design problems, showing substantially better balance of optimum quality and efficiency than purely low- or high-fidelity optimizations.
Taguchi design of experiments nov 24 2013Charlton Inao
This document provides an overview of Taguchi design of experiments. It defines Taguchi methods, which use orthogonal arrays to conduct a minimal number of experiments that can provide full information on factors that affect performance. The assumptions of Taguchi methods include additivity of main effects. The key steps in an experiment are selecting variables and their levels, choosing an orthogonal array, assigning variables to columns, conducting experiments, and analyzing data through sensitivity analysis and ANOVA.
Surrogate-based design is an effective approach for modeling computationally expensive system behavior. In such application, it is often challenging to characterize the expected accuracy of the surrogate. In addition to global and local error measures, regional error measures can be used to understand and interpret the surrogate accuracy in the regions of interest. This paper develops the Regional Error Estimation of Surrogate (REES) method to quantify the level of the error in any given subspace (or region) of the entire domain, when all the available training points have been invested to build the surrogate. In this approach, the accuracy of the surrogate in each subspace is estimated by modeling the variations of the mean and the maximum error in that subspace with increasing number of training points (in an iterative process). A regression model is used for this purpose. At each iteration, the intermediate surrogate is constructed using a subset of the entire training data, and tested over the remaining points. The evaluated errors at the intermediate test points at each iteration are used for training the regression model that represents the error variation with sample points. The effectiveness of the proposed method is illustrated using standard test problems. To this end, the predicted regional errors of the surrogate constructed using all the training points are compared with the regional errors estimated over a large set of test points.
Model based test case prioritization using neural network classificationcseij
Model-based testing for real-life software systems often require a large number of tests, all of which cannot
exhaustively be run due to time and cost constraints. Thus, it is necessary to prioritize the test cases in
accordance with their importance the tester perceives. In this paper, this problem is solved by improving
our given previous study, namely, applying classification approach to the results of our previous study
functional relationship between the test case prioritization group membership and the two attributes:
important index and frequency for all events belonging to given group are established. A for classification
purpose, neural network (NN) that is the most advances is preferred and a data set obtained from our study
for all test cases is classified using multilayer perceptron (MLP) NN. The classification results for
commercial test prioritization application show the high classification accuracies about 96% and the
acceptable test prioritization performances are achieved.
This paper advances the Domain Segmentation based on Uncertainty in the Surrogate (DSUS) framework which is a novel approach to characterize the uncertainty in surrogates. The leave-one-out cross-validation technique is adopted in the DSUS framework to measure local errors of a surrogate. A method is proposed in this paper to evaluate the performance of the leave-out-out cross-validation errors as local error measures. This method evaluates local errors by comparing: (i) the leave-one-out cross-validation error with (ii) the actual local error estimated within a local hypercube for each training point. The comparison results show that the leave-one-out cross-validation strategy can capture the local errors of a surrogate. The DSUS framework is then applied to key aspects of wind resource as- sessment and wind farm cost modeling. The uncertainties in the wind farm cost and the wind power potential are successfully characterized, which provides designers/users more confidence when using these models
This document presents a new 3-level approach to simultaneously select the best surrogate model type, kernel function, and hyper-parameters for approximation models. The approach uses Regional Error Estimation of Surrogates (REES) to evaluate error and select models. It compares a cascaded technique that performs sequential optimization versus a one-step technique. Numerical examples on benchmark problems show the one-step technique reduces maximum and median errors by at least 60% with lower computational cost compared to the cascaded approach. Future work involves applying the one-step method to more complex problems and developing an online platform for collaborative surrogate model selection.
The analysis of complex system behavior often demands expensive experiments or computational simulations. Surrogates modeling techniques are often used to provide a tractable and inexpensive approximation of such complex system behavior. Owing to the lack of knowledge regarding the suitability of particular surrogate modeling techniques, model selection approach can be helpful to choose the best surrogate technique. Popular model selection approaches include: (i) split sample, (ii) cross-validation, (iii) bootstrapping, and (iv) Akaike's information criterion (AIC) (Queipo et al. 2005; Bozdogan et al. 2000). However, the effectiveness of these model selection methods is limited by the lack of accurate measures of local and global errors in surrogates.
This paper develops a novel and model-independent concept to quantify the local/global reliability of surrogates, to assist in model selection (in surrogate applications). This method is called the Generalized-Regional Error Estimation of Surrogate (G-REES). In this method, intermediate surrogates are iteratively constructed over heuristic subsets of the available sample points (i.e., intermediate training points), and tested over the remaining available sample points (i.e., intermediate test points). The fraction of sample points used as intermediate training points is fixed at each iteration, with the total number of iterations being pre-specified. The estimated median and maximum relative errors for the heuristic subsets at each iteration are used to fit a distribution of the median and maximum error, respectively. The statistical mode of the median and the maximum error distributions are then determined. These mode values are then represented as functions of the density of training points (at the corresponding iteration). Regression methods, called Variation of Error with Sample Density (VESD), are used for this purpose. The VESD models are then used to predict the expected median and maximum errors, when all the sample points are used as training points.
The effectiveness of the proposed model selection criterion is explored to find the best surrogate between candidates including: (i) Kriging, (ii) Radial Basis Functions (RBF), (iii) Extended Radial Basis Functions (ERBF), and (iv) Quadratic Response Surface (QRS), for standard test functions and a wind farm capacity factor function. The results will be compared with the relative accuracy of the surrogates evaluated on additional test points, and also with the prediction sum of square (PRESS) error given by leave-one-out cross-validation.
The application of G-REES to a standard test problem with two design variables (Branin-hoo function) show that the proposed method predicts the median and the maximum value of the global error with a higher level of confidence compared to PRESS. It also shows that model selection based on G-REES method is significantly more reliable than that currently performed using error measures such as PRESS. The
BOOLEAN SPECIFICATION BASED TESTING TECHNIQUES: A SURVEYcscpconf
Boolean expressions are major focus of specifications and they are very much prone to
introduction of faults, this survey presents various Boolean specification based testing
techniques, and covers more than 30 papers for the same. The various Boolean specification
based testing techniques like Cause effect graph, fosters strategy, meaningful impact strategy,
Branch Operator Strategy (BOR), Modified Condition/ Decision Coverage (MCDC) compared
on the basis of their fault detection effectiveness and the size of test suite. This collection
represents most of the existing work performed on Boolean specification based testing
techniques. This survey describes the basic algorithms used by these strategies and it also
includes operator and operand fault categories for evaluating the performance of above mentioned testing techniques. Finally, this survey contains short summaries of all the papers that use Boolean specification based testing techniques. These techniques have been empirically evaluated by various researchers on a simplified safety related real time control system.
The document discusses editing techniques for writing technical documents in English, including editing for clarity, consistency in sentence structure, and examples of rewriting sentences for improved clarity and consistency. It also provides a brief biography of Ted Knoy, who has over 20 years experience teaching technical English writing and editing in Taiwan. He is the author of 14 books on technical and professional writing in English.
Approximation models (or surrogate models) provide an efficient substitute to expen- sive physical simulations and an efficient solution to the lack of physical models of system behavior. However, it is challenging to quantify the accuracy and reliability of such ap- proximation models in a region of interest or the overall domain without additional system evaluations. Standard error measures, such as the mean squared error, the cross-validation error, and the Akaikes information criterion, provide limited (often inadequate) informa- tion regarding the accuracy of the final surrogate. This paper introduces a novel and model independent concept to quantify the level of errors in the function value estimated by the final surrogate in any given region of the design domain. This method is called the Re- gional Error Estimation of Surrogate (REES). Assuming the full set of available sample points to be fixed, intermediate surrogates are iteratively constructed over a sample set comprising all samples outside the region of interest and heuristic subsets of samples inside the region of interest (i.e., intermediate training points). The intermediate surrogate is tested over the remaining sample points inside the region of interest (i.e., intermediate test points). The fraction of sample points inside region of interest, which are used as interme- diate training points, is fixed at each iteration, with the total number of iterations being pre-specified. The estimated median and maximum relative errors within the region of in- terest for the heuristic subsets at each iteration are used to fit a distribution of the median and maximum error, respectively. The estimated statistical mode of the median and the maximum error, and the absolute maximum error are then represented as functions of the density of intermediate training points, using regression models. The regression models are then used to predict the expected median and maximum regional errors when all the sample points are used as training points. Standard test functions and a wind farm power generation problem are used to illustrate the effectiveness and the utility of such a regional error quantification method.
This document presents three techniques to accelerate the estimation of a sequence of discrete choice models (DCMs): standardization, warm start, and early stopping. Standardization rescales independent variables to a common scale. Warm start initializes subsequent models with the optimized parameters from previous models. Early stopping monitors the log likelihood improvement over iterations and stops estimation early if improvement plateaus. These techniques are tested on two sequences of 100 DCMs estimated using the BFGS algorithm. Results show that each technique individually and combined provide significant speedups for estimating sequences of DCMs.
This document analyzes and compares three different drying systems - pneumatic dryers, spiral dryers, and rotation dryers with a drum - based on five criteria: heat transfer coefficient, price, drying energy, thermal usefulness, and specific energy use. The Promethee-Gaia methodology is applied to rank the alternatives. Based on the Promethee I, II, and Gaia analyses in the Decision Lab software, the pneumatic dryer is ranked as the best system as it provides the most benefits in terms of cost savings on investment and energy. The spiral dryer is next best, followed by the rotation dryer. Sensitivity analyses show the pneumatic dryer remains the top ranked system even when the criteria
International Journal of Engineering Research and Applications (IJERA) is a team of researchers not publication services or private publications running the journals for monetary benefits, we are association of scientists and academia who focus only on supporting authors who want to publish their work. The articles published in our journal can be accessed online, all the articles will be archived for real time access.
Our journal system primarily aims to bring out the research talent and the works done by sciaentists, academia, engineers, practitioners, scholars, post graduate students of engineering and science. This journal aims to cover the scientific research in a broader sense and not publishing a niche area of research facilitating researchers from various verticals to publish their papers. It is also aimed to provide a platform for the researchers to publish in a shorter of time, enabling them to continue further All articles published are freely available to scientific researchers in the Government agencies,educators and the general public. We are taking serious efforts to promote our journal across the globe in various ways, we are sure that our journal will act as a scientific platform for all researchers to publish their works online.
One of the primary drawbacks plaguing wider acceptance of surrogate models is their low fidelity in general. This issue can be in a large part attributed to the lack of automated model selection techniques, particularly ones that do not make limiting assumptions regarding the choice of model types and kernel types. A novel model selection technique was recently developed to perform optimal model search concurrently at three levels: (i) optimal model type (e.g., RBF), (ii) optimal kernel type (e.g., multiquadric), and (iii) optimal values of hyper-parameters (e.g., shape parameter) that are conventionally kept constant. The error measures to be minimized in this optimal model selection process are determined by the Predictive Estimation of Model Fidelity (PEMF) method, which has been shown to be significantly more accurate than typical cross-validation-based error metrics. In this paper, we make the following important advancements to the PEMF-based model selection framework, now called the Concurrent Surrogate Model Selection or COS- MOS framework: (i) The optimization formulation is modified through binary coding to allow surrogates with differing num- bers of candidate kernels and kernels with differing numbers of hyper-parameters (which was previously not allowed). (ii) A robustness criterion, based on the variance of errors, is added to the existing criteria for model selection. (iii) A larger candidate pool of 16 surrogate-kernel combinations is considered for selection − possibly making COSMOS one of the most comprehensive surrogate model selection framework (in theory and implementation) currently available. The effectiveness of the COSMOS framework is demonstrated by successfully applying it to four benchmark problems (with 2-30 variables) and an airfoil design problem. The optimal model selection results illustrate how diverse models provide important tradeoffs for different problems.
A Review on Feature Selection Methods For Classification TasksEditor IJCATR
In recent years, application of feature selection methods in medical datasets has greatly increased. The challenging task in
feature selection is how to obtain an optimal subset of relevant and non redundant features which will give an optimal solution without
increasing the complexity of the modeling task. Thus, there is a need to make practitioners aware of feature selection methods that have
been successfully applied in medical data sets and highlight future trends in this area. The findings indicate that most existing feature
selection methods depend on univariate ranking that does not take into account interactions between variables, overlook stability of the
selection algorithms and the methods that produce good accuracy employ more number of features. However, developing a universal
method that achieves the best classification accuracy with fewer features is still an open research area.
This paper proposes a novel model management technique to be applied in population- based heuristic optimization. This technique adaptively selects different computational models (both physics-based and statistical models) to be used during optimization, with the overall goal to end with high fidelity solutions in a reasonable time period. For example, in optimizing an aircraft wing to obtain maximum lift-to-drag ratio, one can use low-fidelity models such as given by the vortex lattice method, or a high-fidelity finite volume model (that solves the full Navier-Stokes equations), or a surrogate model that substitutes the high-fidelity model.The information from models with different levels of fidelity is inte- grated into the heuristic optimization process using a novel model-switching metric. In this context, models could be surrogate models, low-fidelity physics-based analytical mod- els, and medium-to-high fidelity computational models (based on grid density). The model switching technique replaces the current model with the next higher fidelity model, when a stochastic switching criterion is met at a given iteration during the optimization process. The switching criteria is based on whether the uncertainty associated with the current model output dominates the latest improvement of the fitness function. In the case of the physics-based models, the uncertainty in their output is quantified through an inverse assessment process by comparing with high-fidelity model responses or experimental data (if available). To determine the fidelity of surrogate models, the Predictive Estimation of Model Fidelity (PEMF) method is applied. The effectiveness of the proposed method is demonstrated by applying it to airfoil optimization with the objective to maximize the lift to drag ratio of the wing under different flow regimes. It was found that the tuned low fidelity model dominates the optimization process in terms of computational time and function calls.
This document discusses feature selection concepts and methods. It defines features as attributes that determine which class an instance belongs to. Feature selection aims to select a relevant subset of features by removing irrelevant, redundant and unnecessary data. This improves learning accuracy, model performance and interpretability. The document categorizes feature selection algorithms as filter, wrapper or embedded methods based on how they evaluate feature subsets. It also discusses concepts like feature relevance, search strategies, successor generation and evaluation measures used in feature selection algorithms.
The document discusses finding the optimal parameter K for the emerging pattern-based classifier PCL. PCL selects the top-K emerging patterns from each class that match a test instance and uses these patterns to predict the class label. The authors develop an algorithm to determine the best value of K for PCL on different datasets, as K's optimal value varies. They sample subsets of training data to find the K that performs best on average, maximizing the likelihood of good performance on the whole dataset. Experimental results show that finding the best K improves PCL's accuracy and that using incremental frequent pattern maintenance techniques speeds up the algorithm significantly.
PROGRAM TEST DATA GENERATION FOR BRANCH COVERAGE WITH GENETIC ALGORITHM: COMP...cscpconf
In search based test data generation, the problem of test data generation is reduced to that of
function minimization or maximization.Traditionally, for branch testing, the problem of test data
generation has been formulated as a minimization problem. In this paper we define an alternate
maximization formulation and experimentally compare it with the minimization formulation. We
use a genetic algorithm as the search technique and in addition to the usual genetic algorithm
operators we also employ the path prefix strategy as a branch ordering strategy and memory and elitism. Results indicate that there is no significant difference in the performance or the coverage obtained through the two approaches and either could be used in test data generation when coupled with the path prefix strategy, memory and elitism.
An SPRT Procedure for an Ungrouped Data using MMLE ApproachIOSR Journals
This document describes a sequential probability ratio test (SPRT) procedure for analyzing ungrouped software failure data using a modified maximum likelihood estimation (MMLE) approach. The SPRT procedure can help quickly detect unreliable software by making decisions with fewer observed failures than traditional hypothesis testing methods. Parameters are estimated using MMLE, which approximates functions in the maximum likelihood equation with linear functions to simplify calculations compared to other estimation methods. The document provides details on how to apply the SPRT procedure and MMLE parameter estimation to a software reliability growth model to analyze software failure data sequentially and detect unreliable software components earlier.
Software reliability models (SRMs) are very important for estimating and predicting software
reliability in the testing/debugging phase. The contributions of this paper are as follows. First, a
historical review of the Gompertz SRM is given. Based on several software failure data, the
parameters of the Gompertz software reliability model are estimated using two estimation
methods, the traditional maximum likelihood and the least square. The methods of estimation are
evaluated using the MSE and R-squared criteria. The results show that the least square
estimation is an attractive method in term of predictive performance and can be used when the
maximum likelihood method fails to give good prediction results.
An introduction to variable and feature selectionMarco Meoni
Presentation of a great paper from Isabelle Guyon (Clopinet) and André Elisseeff (Max Planck Institute) back in 2003, which outlines the main techniques for feature selection and model validation in machine learning systems
Harmony Search for Multi-objective Optimization - SBRN 2012lucasmpavelski
Slides used in the presentation of the article "Harmony Search for Multi-objective
Optimization" in the 2012 Brazilian Symposium on Neural Networks (SBRN). Link to the article: http://ieeexplore.ieee.org/xpl/articleDetails.jsp?reload=true&arnumber=6374852
Threshold benchmarking for feature ranking techniquesjournalBEEI
In prediction modeling, the choice of features chosen from the original feature set is crucial for accuracy and model interpretability. Feature ranking techniques rank the features by its importance but there is no consensus on the number of features to be cut-off. Thus, it becomes important to identify a threshold value or range, so as to remove the redundant features. In this work, an empirical study is conducted for identification of the threshold benchmark for feature ranking algorithms. Experiments are conducted on Apache Click dataset with six popularly used ranker techniques and six machine learning techniques, to deduce a relationship between the total number of input features (N) to the threshold range. The area under the curve analysis shows that ≃ 33-50% of the features are necessary and sufficient to yield a reasonable performance measure, with a variance of 2%, in defect prediction models. Further, we also find that the log2(N) as the ranker threshold value represents the lower limit of the range.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
MOVEMENT ASSISTED COMPONENT BASED SCALABLE FRAMEWORK FOR DISTRIBUTED WIRELESS...ijcsa
Intelligent networks are becoming more enveloping and dwelling a new generation of applications are
deployed over the peer-to-peer networks. Intelligent networks are very attractive because of their role in
improving the scalability and enhancing performance by enabling direct and real-time communication
among the participating network stations. A suitable solution for resource management in distributed wireless systems is required which should support fault-tolerant operations, requested resources (at shortest path), minimize overhead generation during network management, balancing the load distribution between the participating stations and high probability of lookup success and many more. This article
presents a Movement Assisted Component Based Scalable Framework (MAC-SF) for the distributed
network which manages the distributed wireless resources and applications; monitors the behavior of the
distributed wireless applications transparently and attains accurate resource projections, manages the
connections between the participating network stations and distributes the active objects in response to the
user requests and changing processing and network conditions. This system is also compared with some
exiting systems. Results shows that MAC-SF is a better system and can be used in any wireless network.
Quality estimation of machine translation outputs through stemmingijcsa
Machine Translation is the challenging problem for Indian languages. Every day we can see some machine
translators being developed , but getting a high quality automatic translation is still a very distant dream .
The correct translated sentence for Hindi language is rarely found. In this paper, we are emphasizing on
English-Hindi language pair, so in order to preserve the correct MT output we present a ranking system,
which employs some machine learning techniques and morphological features. In ranking no human
intervention is required. We have also validated our results by comparing it with human ranking.
The proposed technique is capable of segmenting the lung’s air and its soft tissues followed by estimating
the lung’s air volume and its variations throughout the image sequence. The work presents a methodology
that consists of three steps. In the first step, CT image sequence are to be given as input where histograms
of all images within the sequence are calculated and they are overlaid in order to form the sequence’s
combined histogram to extract lung air area. In the second step, segment the lung air area by using
optimum threshold based method. In the third step, voxels in the rendered volume will be counted and the
percentage of air consumption will be estimated. The result will indicate a very good ability of the
proposed method for estimating the lung’s air volume and its variations in a respiratory image sequence.
Algeria engages with determination on the path renewable energies to bring global and long-lasting
solutions to the environmental challenges and to the problems of conservation of the energy resources of
fossil origin. Our study is interested on the wind spinneret which seems one of the most promising with a
very high global growth rate. The object of this article is to estimate the wind deposit of the region of Oran
(Es Senia), important stage in any planning and realization of wind plant. In our work, we began with the
processing of schedules data relative to the wind collected over a period of more than 50 years, to evaluate
the wind potential while determining its frequencies. Then, we calculated the total electrical energy
produced at various heights with three types of wind turbines.The analysis of the results shows that the
wind turbines of major powers allow producing important quantities of energy when we increase the height
of their hubs to take advantage of stronger speeds of wind.
Grid computing or network computing is developed to make the available electric power in the similar way
as it is available for the grid. For that we just plug in the power and whoever needs power, may use it. In
grid computing if a system needs more power than available it can share the computing with other
machines connected in a grid. In this way we can use the power of a super computer without a huge cost
and the CPU cycles that were wasted previously can also be utilized. For performing grid computation in
joined computers through the internet, the software must be installed which supports grid computation on
each computer inside the VO. The software handles information queries, storage management, processing
scheduling, authentication and data encryption to ensure information security.
The document discusses editing techniques for writing technical documents in English, including editing for clarity, consistency in sentence structure, and examples of rewriting sentences for improved clarity and consistency. It also provides a brief biography of Ted Knoy, who has over 20 years experience teaching technical English writing and editing in Taiwan. He is the author of 14 books on technical and professional writing in English.
Approximation models (or surrogate models) provide an efficient substitute to expen- sive physical simulations and an efficient solution to the lack of physical models of system behavior. However, it is challenging to quantify the accuracy and reliability of such ap- proximation models in a region of interest or the overall domain without additional system evaluations. Standard error measures, such as the mean squared error, the cross-validation error, and the Akaikes information criterion, provide limited (often inadequate) informa- tion regarding the accuracy of the final surrogate. This paper introduces a novel and model independent concept to quantify the level of errors in the function value estimated by the final surrogate in any given region of the design domain. This method is called the Re- gional Error Estimation of Surrogate (REES). Assuming the full set of available sample points to be fixed, intermediate surrogates are iteratively constructed over a sample set comprising all samples outside the region of interest and heuristic subsets of samples inside the region of interest (i.e., intermediate training points). The intermediate surrogate is tested over the remaining sample points inside the region of interest (i.e., intermediate test points). The fraction of sample points inside region of interest, which are used as interme- diate training points, is fixed at each iteration, with the total number of iterations being pre-specified. The estimated median and maximum relative errors within the region of in- terest for the heuristic subsets at each iteration are used to fit a distribution of the median and maximum error, respectively. The estimated statistical mode of the median and the maximum error, and the absolute maximum error are then represented as functions of the density of intermediate training points, using regression models. The regression models are then used to predict the expected median and maximum regional errors when all the sample points are used as training points. Standard test functions and a wind farm power generation problem are used to illustrate the effectiveness and the utility of such a regional error quantification method.
This document presents three techniques to accelerate the estimation of a sequence of discrete choice models (DCMs): standardization, warm start, and early stopping. Standardization rescales independent variables to a common scale. Warm start initializes subsequent models with the optimized parameters from previous models. Early stopping monitors the log likelihood improvement over iterations and stops estimation early if improvement plateaus. These techniques are tested on two sequences of 100 DCMs estimated using the BFGS algorithm. Results show that each technique individually and combined provide significant speedups for estimating sequences of DCMs.
This document analyzes and compares three different drying systems - pneumatic dryers, spiral dryers, and rotation dryers with a drum - based on five criteria: heat transfer coefficient, price, drying energy, thermal usefulness, and specific energy use. The Promethee-Gaia methodology is applied to rank the alternatives. Based on the Promethee I, II, and Gaia analyses in the Decision Lab software, the pneumatic dryer is ranked as the best system as it provides the most benefits in terms of cost savings on investment and energy. The spiral dryer is next best, followed by the rotation dryer. Sensitivity analyses show the pneumatic dryer remains the top ranked system even when the criteria
International Journal of Engineering Research and Applications (IJERA) is a team of researchers not publication services or private publications running the journals for monetary benefits, we are association of scientists and academia who focus only on supporting authors who want to publish their work. The articles published in our journal can be accessed online, all the articles will be archived for real time access.
Our journal system primarily aims to bring out the research talent and the works done by sciaentists, academia, engineers, practitioners, scholars, post graduate students of engineering and science. This journal aims to cover the scientific research in a broader sense and not publishing a niche area of research facilitating researchers from various verticals to publish their papers. It is also aimed to provide a platform for the researchers to publish in a shorter of time, enabling them to continue further All articles published are freely available to scientific researchers in the Government agencies,educators and the general public. We are taking serious efforts to promote our journal across the globe in various ways, we are sure that our journal will act as a scientific platform for all researchers to publish their works online.
One of the primary drawbacks plaguing wider acceptance of surrogate models is their low fidelity in general. This issue can be in a large part attributed to the lack of automated model selection techniques, particularly ones that do not make limiting assumptions regarding the choice of model types and kernel types. A novel model selection technique was recently developed to perform optimal model search concurrently at three levels: (i) optimal model type (e.g., RBF), (ii) optimal kernel type (e.g., multiquadric), and (iii) optimal values of hyper-parameters (e.g., shape parameter) that are conventionally kept constant. The error measures to be minimized in this optimal model selection process are determined by the Predictive Estimation of Model Fidelity (PEMF) method, which has been shown to be significantly more accurate than typical cross-validation-based error metrics. In this paper, we make the following important advancements to the PEMF-based model selection framework, now called the Concurrent Surrogate Model Selection or COS- MOS framework: (i) The optimization formulation is modified through binary coding to allow surrogates with differing num- bers of candidate kernels and kernels with differing numbers of hyper-parameters (which was previously not allowed). (ii) A robustness criterion, based on the variance of errors, is added to the existing criteria for model selection. (iii) A larger candidate pool of 16 surrogate-kernel combinations is considered for selection − possibly making COSMOS one of the most comprehensive surrogate model selection framework (in theory and implementation) currently available. The effectiveness of the COSMOS framework is demonstrated by successfully applying it to four benchmark problems (with 2-30 variables) and an airfoil design problem. The optimal model selection results illustrate how diverse models provide important tradeoffs for different problems.
A Review on Feature Selection Methods For Classification TasksEditor IJCATR
In recent years, application of feature selection methods in medical datasets has greatly increased. The challenging task in
feature selection is how to obtain an optimal subset of relevant and non redundant features which will give an optimal solution without
increasing the complexity of the modeling task. Thus, there is a need to make practitioners aware of feature selection methods that have
been successfully applied in medical data sets and highlight future trends in this area. The findings indicate that most existing feature
selection methods depend on univariate ranking that does not take into account interactions between variables, overlook stability of the
selection algorithms and the methods that produce good accuracy employ more number of features. However, developing a universal
method that achieves the best classification accuracy with fewer features is still an open research area.
This paper proposes a novel model management technique to be applied in population- based heuristic optimization. This technique adaptively selects different computational models (both physics-based and statistical models) to be used during optimization, with the overall goal to end with high fidelity solutions in a reasonable time period. For example, in optimizing an aircraft wing to obtain maximum lift-to-drag ratio, one can use low-fidelity models such as given by the vortex lattice method, or a high-fidelity finite volume model (that solves the full Navier-Stokes equations), or a surrogate model that substitutes the high-fidelity model.The information from models with different levels of fidelity is inte- grated into the heuristic optimization process using a novel model-switching metric. In this context, models could be surrogate models, low-fidelity physics-based analytical mod- els, and medium-to-high fidelity computational models (based on grid density). The model switching technique replaces the current model with the next higher fidelity model, when a stochastic switching criterion is met at a given iteration during the optimization process. The switching criteria is based on whether the uncertainty associated with the current model output dominates the latest improvement of the fitness function. In the case of the physics-based models, the uncertainty in their output is quantified through an inverse assessment process by comparing with high-fidelity model responses or experimental data (if available). To determine the fidelity of surrogate models, the Predictive Estimation of Model Fidelity (PEMF) method is applied. The effectiveness of the proposed method is demonstrated by applying it to airfoil optimization with the objective to maximize the lift to drag ratio of the wing under different flow regimes. It was found that the tuned low fidelity model dominates the optimization process in terms of computational time and function calls.
This document discusses feature selection concepts and methods. It defines features as attributes that determine which class an instance belongs to. Feature selection aims to select a relevant subset of features by removing irrelevant, redundant and unnecessary data. This improves learning accuracy, model performance and interpretability. The document categorizes feature selection algorithms as filter, wrapper or embedded methods based on how they evaluate feature subsets. It also discusses concepts like feature relevance, search strategies, successor generation and evaluation measures used in feature selection algorithms.
The document discusses finding the optimal parameter K for the emerging pattern-based classifier PCL. PCL selects the top-K emerging patterns from each class that match a test instance and uses these patterns to predict the class label. The authors develop an algorithm to determine the best value of K for PCL on different datasets, as K's optimal value varies. They sample subsets of training data to find the K that performs best on average, maximizing the likelihood of good performance on the whole dataset. Experimental results show that finding the best K improves PCL's accuracy and that using incremental frequent pattern maintenance techniques speeds up the algorithm significantly.
PROGRAM TEST DATA GENERATION FOR BRANCH COVERAGE WITH GENETIC ALGORITHM: COMP...cscpconf
In search based test data generation, the problem of test data generation is reduced to that of
function minimization or maximization.Traditionally, for branch testing, the problem of test data
generation has been formulated as a minimization problem. In this paper we define an alternate
maximization formulation and experimentally compare it with the minimization formulation. We
use a genetic algorithm as the search technique and in addition to the usual genetic algorithm
operators we also employ the path prefix strategy as a branch ordering strategy and memory and elitism. Results indicate that there is no significant difference in the performance or the coverage obtained through the two approaches and either could be used in test data generation when coupled with the path prefix strategy, memory and elitism.
An SPRT Procedure for an Ungrouped Data using MMLE ApproachIOSR Journals
This document describes a sequential probability ratio test (SPRT) procedure for analyzing ungrouped software failure data using a modified maximum likelihood estimation (MMLE) approach. The SPRT procedure can help quickly detect unreliable software by making decisions with fewer observed failures than traditional hypothesis testing methods. Parameters are estimated using MMLE, which approximates functions in the maximum likelihood equation with linear functions to simplify calculations compared to other estimation methods. The document provides details on how to apply the SPRT procedure and MMLE parameter estimation to a software reliability growth model to analyze software failure data sequentially and detect unreliable software components earlier.
Software reliability models (SRMs) are very important for estimating and predicting software
reliability in the testing/debugging phase. The contributions of this paper are as follows. First, a
historical review of the Gompertz SRM is given. Based on several software failure data, the
parameters of the Gompertz software reliability model are estimated using two estimation
methods, the traditional maximum likelihood and the least square. The methods of estimation are
evaluated using the MSE and R-squared criteria. The results show that the least square
estimation is an attractive method in term of predictive performance and can be used when the
maximum likelihood method fails to give good prediction results.
An introduction to variable and feature selectionMarco Meoni
Presentation of a great paper from Isabelle Guyon (Clopinet) and André Elisseeff (Max Planck Institute) back in 2003, which outlines the main techniques for feature selection and model validation in machine learning systems
Harmony Search for Multi-objective Optimization - SBRN 2012lucasmpavelski
Slides used in the presentation of the article "Harmony Search for Multi-objective
Optimization" in the 2012 Brazilian Symposium on Neural Networks (SBRN). Link to the article: http://ieeexplore.ieee.org/xpl/articleDetails.jsp?reload=true&arnumber=6374852
Threshold benchmarking for feature ranking techniquesjournalBEEI
In prediction modeling, the choice of features chosen from the original feature set is crucial for accuracy and model interpretability. Feature ranking techniques rank the features by its importance but there is no consensus on the number of features to be cut-off. Thus, it becomes important to identify a threshold value or range, so as to remove the redundant features. In this work, an empirical study is conducted for identification of the threshold benchmark for feature ranking algorithms. Experiments are conducted on Apache Click dataset with six popularly used ranker techniques and six machine learning techniques, to deduce a relationship between the total number of input features (N) to the threshold range. The area under the curve analysis shows that ≃ 33-50% of the features are necessary and sufficient to yield a reasonable performance measure, with a variance of 2%, in defect prediction models. Further, we also find that the log2(N) as the ranker threshold value represents the lower limit of the range.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
MOVEMENT ASSISTED COMPONENT BASED SCALABLE FRAMEWORK FOR DISTRIBUTED WIRELESS...ijcsa
Intelligent networks are becoming more enveloping and dwelling a new generation of applications are
deployed over the peer-to-peer networks. Intelligent networks are very attractive because of their role in
improving the scalability and enhancing performance by enabling direct and real-time communication
among the participating network stations. A suitable solution for resource management in distributed wireless systems is required which should support fault-tolerant operations, requested resources (at shortest path), minimize overhead generation during network management, balancing the load distribution between the participating stations and high probability of lookup success and many more. This article
presents a Movement Assisted Component Based Scalable Framework (MAC-SF) for the distributed
network which manages the distributed wireless resources and applications; monitors the behavior of the
distributed wireless applications transparently and attains accurate resource projections, manages the
connections between the participating network stations and distributes the active objects in response to the
user requests and changing processing and network conditions. This system is also compared with some
exiting systems. Results shows that MAC-SF is a better system and can be used in any wireless network.
Quality estimation of machine translation outputs through stemmingijcsa
Machine Translation is the challenging problem for Indian languages. Every day we can see some machine
translators being developed , but getting a high quality automatic translation is still a very distant dream .
The correct translated sentence for Hindi language is rarely found. In this paper, we are emphasizing on
English-Hindi language pair, so in order to preserve the correct MT output we present a ranking system,
which employs some machine learning techniques and morphological features. In ranking no human
intervention is required. We have also validated our results by comparing it with human ranking.
The proposed technique is capable of segmenting the lung’s air and its soft tissues followed by estimating
the lung’s air volume and its variations throughout the image sequence. The work presents a methodology
that consists of three steps. In the first step, CT image sequence are to be given as input where histograms
of all images within the sequence are calculated and they are overlaid in order to form the sequence’s
combined histogram to extract lung air area. In the second step, segment the lung air area by using
optimum threshold based method. In the third step, voxels in the rendered volume will be counted and the
percentage of air consumption will be estimated. The result will indicate a very good ability of the
proposed method for estimating the lung’s air volume and its variations in a respiratory image sequence.
Algeria engages with determination on the path renewable energies to bring global and long-lasting
solutions to the environmental challenges and to the problems of conservation of the energy resources of
fossil origin. Our study is interested on the wind spinneret which seems one of the most promising with a
very high global growth rate. The object of this article is to estimate the wind deposit of the region of Oran
(Es Senia), important stage in any planning and realization of wind plant. In our work, we began with the
processing of schedules data relative to the wind collected over a period of more than 50 years, to evaluate
the wind potential while determining its frequencies. Then, we calculated the total electrical energy
produced at various heights with three types of wind turbines.The analysis of the results shows that the
wind turbines of major powers allow producing important quantities of energy when we increase the height
of their hubs to take advantage of stronger speeds of wind.
Grid computing or network computing is developed to make the available electric power in the similar way
as it is available for the grid. For that we just plug in the power and whoever needs power, may use it. In
grid computing if a system needs more power than available it can share the computing with other
machines connected in a grid. In this way we can use the power of a super computer without a huge cost
and the CPU cycles that were wasted previously can also be utilized. For performing grid computation in
joined computers through the internet, the software must be installed which supports grid computation on
each computer inside the VO. The software handles information queries, storage management, processing
scheduling, authentication and data encryption to ensure information security.
Effect of mesh grid structure in reducing hot carrier effect of nmos device s...ijcsa
This paper presents the critical effect of mesh grid that should be considered during process and device
simulation using modern TCAD tools in order to develop and optimize their accurate electrical
characteristics. Here, the computational modelling process of developing the NMOS device structure is
performed in Athena and Atlas. The effect of Mesh grid on net doping profile, n++, and LDD sheet
resistance that could link to unwanted “Hot Carrier Effect” were investigated by varying the device grid
resolution in both directions. It is found that y-grid give more profound effect in the doping concentration,
the junction depth formation and the value of threshold voltage during simulation. Optimized mesh grid is
obtained and tested for more accurate and faster simulation. Process parameter (such as oxide thicknesses
and Sheet resistance) as well as Device Parameter (such as linear gain “beta” and SPICE level 3 mobility
roll-off parameter “ Theta”) are extracted and investigated for further different applications.
Hybrid hmmdtw based speech recognition with kernel adaptive filtering methodijcsa
This document summarizes a research paper that proposes a new approach for speech recognition using kernel adaptive filtering for speech enhancement and a hybrid HMM/DTW method for recognition. It first discusses adaptive filters and the LMS algorithm, then introduces kernel adaptive filters using the KLMS algorithm to transform input data into a high-dimensional feature space. Finally, it describes using HMM to train speech features and DTW for classification and recognition. The experimental results showed an improvement in recognition rates compared to traditional methods.
REAL TIME SPECIAL EFFECTS GENERATION AND NOISE FILTRATION OF AUDIO SIGNAL USI...ijcsa
Digital signal processing is being increasingly used for audio processing applications. Digital audio effects
refer to all those algorithms that are used for enhancing sound in any of the steps of a processing chain of
music production. Real time audio effects generation is a highly challenging task in the field of signal
processing. Now a day, almost every high end multimedia audio device does digital signal processing in
one form or another. For years musicians have used different techniques to give their music a unique
sound. Earlier, these techniques were implemented after a lot of work and experimentation. However, now
with the emergence of digital signal processing this task is simplified to a great extent. In this article, the
generations of special effects like echo, flanging, reverberation, stereo, karaoke, noise filtering etc are
successfully implemented using MATLAB and an attractive GUI has been designed for the same.
Graph coloring is the assignment of colors to the graph vertices and edges in the graph theory. We can
divide the graph coloring in two types. The first is vertex coloring and the second is edge coloring. The
condition which we follow in graph coloring is that the incident vertices/edges have not the same color.
There are some algorithms which solve the problem of graph coloring. Some are offline algorithm and
others are online algorithm. Where offline means the graph is known in advance and the online means that
the edges of the graph are arrive one by one as an input, and We need to color each edge as soon as it is
added to the graph and the main issue is that we want to minimize the number of colors. We cannot change
the color of an edge after colored in an online algorithm. In this paper, we improve the online algorithm
for edge coloring. There is also a theorem which proves that if the maximum degree of a graph is Δ, then it
is possible to color its edges, in polynomial time, using at most Δ+ 1 color. The algorithm provided by
Vizing is offline, i.e., it assumes the whole graph is known in advance. In online algorithm edges arrive one
by one in a random permutation. This online algorithm is inspired by a distributed offline algorithm of
Panconesi and Srinivasan, referred as PS algorithm, works on 2-rounds which we extend by reusing colors
online in multiple rounds.
In this paper a new evolutionary algorithm, for continuous nonlinear optimization problems, is surveyed.
This method is inspired by the life of a bird, called Cuckoo.
The Cuckoo Optimization Algorithm (COA) is evaluated by using the Rastrigin function. The problem is a
non-linear continuous function which is used for evaluating optimization algorithms. The efficiency of the
COA has been studied by obtaining optimal solution of various dimensions Rastrigin function in this paper.
The mentioned function also was solved by FA and ABC algorithms. Comparing the results shows the COA
has better performance than other algorithms.
Application of algorithm to test function has proven its capability to deal with difficult optimization
Security & privacy issues of cloud & grid computing networksijcsa
Cloud computing is a new field in Internet computing that provides novel perspectives in internetworking
technologies. Cloud computing has become a significant technology in field of information technology.
Security of confidential data is a very important area of concern as it can make way for very big problems
if unauthorized users get access to it. Cloud computing should have proper techniques where data is
segregated properly for data security and confidentiality. This paper strives to compare and contrast cloud
computing with grid computing, along with the Tools and simulation environment & Tips to store data and
files safely in Cloud.
Augmented split –protocol; an ultimate d do s defenderijcsa
Distributed Denials of Service (DDoS) attacks have become the daunting problem for businesses, state
administrator and computer system users. Prevention and detection of a DDoS attack is a major research
topic for researchers throughout the world. As new remedies are developed to prevent or mitigate DDoS
attacks, invaders are continually evolving new methods to circumvent these new procedures. In this paper,
we describe various DDoS attack mechanisms, categories, scope of DDoS attacks and their existing
countermeasures. In response, we propose to introduce DDoS resistant Augmented Split-protocol (ASp).
The migratory nature and role changeover ability of servers in Split-protocol architecture will avoid
bottleneck at the server side. It also offers the unique ability to avoid server saturation and compromise
from DDoS attacks. The goal of this paper is to present the concept and performance of (ASp) as a
defensive tool against DDoS attacks.
AN ENHANCED EDGE ADAPTIVE STEGANOGRAPHY APPROACH USING THRESHOLD VALUE FOR RE...ijcsa
The document summarizes an enhanced edge adaptive steganography approach using a threshold value for region selection. It aims to improve the quality and modification rate of a stego image compared to Sobel and Canny edge detection techniques. The proposed approach uses a threshold value to select high frequency pixels from the cover image for data embedding using LSBMR. Experimental results on 100 images show about a 0.2-0.6% improvement in image quality measured by PSNR and a 4-10% improvement in modification rate measured by MSE compared to Sobel and Canny edge detection.
Mining sequential patterns for interval basedijcsa
Sequential pattern mining finds the frequent subsequences or patterns from the given sequences.
TPrefixSpan algorithm finds the relevant frequent patterns from the given sequential patterns formed using
interval based events. In our proposed work, we add multiple constraints like item, length and aggregate to
the interval based TPrefixSpan algorithm. By adding these constraints the efficiency and effectiveness of
the algorithm improves. The proposed constraint based algorithm CTPrefixSpan has been applied to
synthetic medical dataset. The algorithm can be applied for stock market analysis, DNA sequences analysis
etc.
KEYWORDS
Sequential patterns, temporal patterns, Constraints, Interval based events.
The singer expresses how they used to long to meet Jesus during important events in his life like his birth and second coming, but now sees Jesus manifest through their romantic partner. They find Jesus's love, care, and presence through seeing him in their partner's eyes, feeling him in their touch, and hearing him in their voice which allows the singer to fully trust and love their partner.
Alila Hotels and Resorts will be opening their first property in Thailand, Alila Cha-Am, in February 2008. The luxury beachside resort in Cha-Am was designed by a leading Thai architect and will feature 79 rooms and villas, two signature restaurants, a spa, and pools for relaxation. The design aims to connect guests with their emotions and promote well-being through the spaces and experiences offered.
The song "Someone to Watch Over Me" is a jazz standard written in 1926 with music by George Gershwin and lyrics by Ira Gershwin. It expresses the idea of wanting someone to care for and protect you through life's uncertainties. The lyrics describe longing for a romantic partner to watch over, protect, and guide you through whatever may come.
The Thiptara Thai restaurant in the Peninsula Hotel delivers an authentic Thai dining experience. Located on the banks of the Chao Phraya River, it has decor and location that transport diners to Thailand. The food is meticulously prepared in traditional Thai styles but uses high quality ingredients, resulting in flavors that are authentically Thai but also five-star. Dining at Thiptara combines all the right elements to provide an authentic Thai dining experience in an elegant setting.
Code coverage based test case selection and prioritizationijseajournal
This document describes an approach for code coverage based test case selection and prioritization. It presents algorithms for test case selection (TCS) and test case prioritization (TCP) that aim to select a minimal set of test cases that achieve maximum code coverage. The TCS algorithm analyzes test cases and statements to cluster them into outdated, required and surplus test cases. The TCP algorithm then prioritizes the required test cases based on their individual statement coverage to achieve full code coverage with as few test cases as possible. An example application of the algorithms on a sample program is provided, demonstrating a reduction in test suite size.
A NOVEL APPROACH FOR TEST CASEPRIORITIZATIONIJCSEA Journal
This paper proposes a novel approach to test case prioritization that calculates the product of a test case's statement coverage and number of function calls to determine priority. The test cases are ordered based on this product metric, with the highest product value first. An algorithm is presented and evaluations show the approach improves fault detection rates over non-prioritized test cases, as measured by the APFD metric. The approach addresses potential ambiguities when multiple test cases have the same product value by further prioritizing based on number of function calls or execution order. The results demonstrate the effectiveness of the proposed prioritization formula and algorithm.
Enhanced technique for regression testingeSAT Journals
Abstract
Regression testing perform where to retest modified version of software which already modified by developers on the basis of
Feedback and testing of the software. Regression testing performs where no internal software knowledge and code known by the
tester. Here all test cases not possible to retest this is a problem. So here we will use test case priority technique. that will
provide priority to retest test case first. this test case priority technique uses for increase efficiency of regression testing. Because
high priority test case execute first. and in our proposed technique we are set priority on the basis of [average time for
execution(figure.1).which find out average time for execution on the basis of fault/time. and we will also ignore test cases which
contain same average time for execution and we will select only one test cases in this situation.. After arrange test cases in
descending order by average time of test cases execute test suite. This method gives more prioritization of test cases is compare to
previous techniques. Finally we will get more prioritized test suite. Then we have also checked efficiency of this new method by
APFD Metric [1]. Our primary purpose of this research paper to introduce new prioritization technique which provides better
result is compare to previous techniques. which improves efficiency of Regression testing.
Keywords: Regression Testing, Prioritization, Test Case, Test Suite, APFD Metric[1].
Abstract—Combinatorial testing (also called interaction testing) is an effective specification-based test input generation technique. By now most of research work in combinatorial testing aims to propose novel approaches trying to generate test suites with minimum size that still cover all the pairwise, triple, or n-way combinations of factors. Since the difficulty of solving this problem is demonstrated to be NP-hard, existing approaches have been designed to generate optimal or near optimal combinatorial test suites in polynomial time. In this paper, we try to apply particle swarm optimization (PSO), a kind of meta-heuristic search technique, to pairwise testing (i.e. a special case of combinatorial testing aiming to cover all the pairwise combinations). To systematically build pairwise test suites, we propose two different PSO based algorithms. One algorithm is based on one-test-at-a-time strategy and the other is based on IPO-like strategy. In these two different algorithms, we use PSO to complete the construction of a single test. To successfully apply PSO to cover more uncovered pairwise combinations in this construction process, we provide a detailed description on how to formulate the search space, define the fitness function and set some heuristic settings. To verify the effectiveness of our approach, we implement these algorithms and choose some typical inputs. In our empirical study, we analyze the impact factors of our approach and compare our approach to other well-known approaches. Final empirical results show the effectiveness and efficiency of our approach.
Test design techniques: Structured and Experienced-based techniquesKhuong Nguyen
This document discusses different types of software testing techniques, including structured-based techniques like cyclomatic complexity and statement/decision coverage, as well as experience-based techniques like error guessing and exploratory testing. It explains how to calculate cyclomatic complexity and coverage percentages. Choosing the appropriate testing technique depends on factors like system type, standards, requirements, risk level, documentation, tester knowledge, time and budget. Testing usually involves combining different techniques.
Guidelines to Understanding Design of Experiment and Reliability Predictionijsrd.com
This paper will focus on how to plan experiments effectively and how to analyse data correctly. Practical and correct methods for analysing data from life testing will also be provided. This paper gives an extensive overview of reliability issues, definitions and prediction methods currently used in the industry. It defines different methods and correlations between these methods in order to make reliability comparison statements from different manufacturers' in easy way that may use different prediction methods and databases for failure rates. The paper finds however such comparison very difficult and risky unless the conditions for the reliability statements are scrutinized and analysed in detail.
Regression testing concentrates on finding defects after a major code change has occurred. Specifically, it
exposes software regressions or old bugs that have reappeared. It is an expensive testing process that has
been estimated to account for almost half of the cost of software maintenance. To improve the regression
testing process, test case prioritization techniques organizes the execution level of test cases. Further, it
gives an improved rate of fault identification, when test suites cannot run to completion.
A study on the efficiency of a test analysis method utilizing test-categories...Tsuyoshi Yumoto
This document describes a study on improving the efficiency of test analysis through utilizing test categories based on application under test (AUT) knowledge and known faults. The study proposes a method for defining test categories based on logical structures of features to guide test condition determination. A verification experiment was conducted and showed measurable improvement in test coverage when using the proposed method. The method aims to minimize variability in test analysis results by providing a standardized process for testers to follow.
Special Double Sampling Plan for truncated life tests based on the Marshall-O...ijceronline
International Journal of Computational Engineering Research (IJCER) is dedicated to protecting personal information and will make every reasonable effort to handle collected information appropriately. All information collected, as well as related requests, will be handled as carefully and efficiently as possible in accordance with IJCER standards for integrity and objectivity.
EXTRACTING THE MINIMIZED TEST SUITE FOR REVISED SIMULINK/STATEFLOW MODELijaia
Test case generation techniques are successfully employed to generate test cases from a formal model. A problem is that as the model evolves, test suites tend to grow in size, making it too costly to execute entire test suites. This paper aims to propose a practical approach to reduce the size of test suites for modified Simulink/Stateflow (SL/SF) model, which is popularly used for modeling software behavior in many industries like automobile manufacturers. The model for describing a system is frequently modified until it is fixed. The proposed technique is capable of extracting the minimized sized test suite in terms of test coverage, by taking into account both the modified and the affected portion of revised SL/SF model. Two real models for the ECUs deployed in a commercial car are used for an empirical study.
TEST CASE PRIORITIZATION FOR OPTIMIZING A REGRESSION TESTijfcstjournal
Regression testing makes sure that upgradation of software in terms of adding new features or for bug
fixing purposes should not hamper previously working functionalities. Whenever a software is upgraded or
modified, a set of test cases are run on each of its functions to assure that the change to that function is not
affecting other parts of the software that were previously running flawlessly. For achieving this, all existing
test cases need to run as well as new test cases might be required to be created. It is not feasible to reexecute
every test case for all the functions of a given software, because if there is a large number of test
cases to be run, then a lot of time and effort would be required. This problem can be addressed by
prioritizing test cases. Test case prioritization technique reorders the priority in which test cases are
implemented, in an attempt to ensure that maximum faults are uncovered early on by the high priority test
cases implemented first. In this paper we propose an optimized test case prioritization technique using Ant
Colony Optimization (ACO) to reduce the cost, effort and time taken to perform regression testing and also
uncover maximum faults. Comparison of different techniques such as Retest All, Test Case Minimization,
Test Case Prioritization, Random Test Case Selection and Test Case Prioritization using ACO is also
depicted.
TEST CASE PRIORITIZATION FOR OPTIMIZING A REGRESSION TESTijfcstjournal
Regression testing makes sure that upgradation of software in terms of adding new features or for bug fixing purposes should not hamper previously working functionalities. Whenever a software is upgraded or modified, a set of test cases are run on each of its functions to assure that the change to that function is not affecting other parts of the software that were previously running flawlessly. For achieving this, all existing test cases need to run as well as new test cases might be required to be created. It is not feasible to re- execute every test case for all the functions of a given software, because if there is a large number of test cases to be run, then a lot of time and effort would be required. This problem can be addressed by prioritizing test cases. Test case prioritization technique reorders the priority in which test cases are implemented, in an attempt to ensure that maximum faults are uncovered early on by the high priority test cases implemented first. In this paper we propose an optimized test case prioritization technique using Ant Colony Optimization (ACO) to reduce the cost, effort and time taken to perform regression testing and also uncover maximum faults. Comparison of different techniques such as Retest All, Test Case Minimization, Test Case Prioritization, Random Test Case Selection and Test Case Prioritization using ACO is also depicted.
TEST CASE PRIORITIZATION FOR OPTIMIZING A REGRESSION TESTijfcstjournal
Regression testing makes sure that upgradation of software in terms of adding new features or for bug
fixing purposes should not hamper previously working functionalities. Whenever a software is upgraded or
modified, a set of test cases are run on each of its functions to assure that the change to that function is not
affecting other parts of the software that were previously running flawlessly. For achieving this, all existing
test cases need to run as well as new test cases might be required to be created. It is not feasible to reexecute every test case for all the functions of a given software, because if there is a large number of test
cases to be run, then a lot of time and effort would be required. This problem can be addressed by
prioritizing test cases. Test case prioritization technique reorders the priority in which test cases are
implemented, in an attempt to ensure that maximum faults are uncovered early on by the high priority test
cases implemented first. In this paper we propose an optimized test case prioritization technique using Ant
Colony Optimization (ACO) to reduce the cost, effort and time taken to perform regression testing and also
uncover maximum faults. Comparison of different techniques such as Retest All, Test Case Minimization,
Test Case Prioritization, Random Test Case Selection and Test Case Prioritization using ACO is also
depicted.
TEST CASE PRIORITIZATION FOR OPTIMIZING A REGRESSION TESTijfcstjournal
Regression testing makes sure that upgradation of software in terms of adding new features or for bug
fixing purposes should not hamper previously working functionalities. Whenever a software is upgraded or
modified, a set of test cases are run on each of its functions to assure that the change to that function is not
affecting other parts of the software that were previously running flawlessly. For achieving this, all existing
test cases need to run as well as new test cases might be required to be created. It is not feasible to reexecute
every test case for all the functions of a given software, because if there is a large number of test
cases to be run, then a lot of time and effort would be required. This problem can be addressed by
prioritizing test cases. Test case prioritization technique reorders the priority in which test cases are
implemented, in an attempt to ensure that maximum faults are uncovered early on by the high priority test
cases implemented first. In this paper we propose an optimized test case prioritization technique using Ant
Colony Optimization (ACO) to reduce the cost, effort and time taken to perform regression testing and also
uncover maximum faults. Comparison of different techniques such as Retest All, Test Case Minimization,
Test Case Prioritization, Random Test Case Selection and Test Case Prioritization using ACO is also
depicted.
International Journal of Business and Management Invention (IJBMI) is an international journal intended for professionals and researchers in all fields of Business and Management. IJBMI publishes research articles and reviews within the whole field Business and Management, new teaching methods, assessment, validation and the impact of new technologies and it will continue to provide information on the latest trends and developments in this ever-expanding subject. The publications of papers are selected through double peer reviewed to ensure originality, relevance, and readability. The articles published in our journal can be accessed online.
Configuration Navigation Analysis Model for Regression Test Case Prioritizationijsrd.com
Regression testing has been receiving increasing attention nowadays. Numerous regression testing strategies have been proposed. Most of them take into account various metrics like cost as well as the ability to find faults quickly thereby saving overall testing time. In this paper, a new model called the Configuration Navigation Analysis Model is proposed which tries to consider all stakeholders and various testing aspects while prioritizing regression test cases.
The document discusses software testing objectives, principles, techniques and processes. It covers black-box and white-box testing, unit and integration testing, and challenges of object-oriented testing. Testing aims to find bugs but can never prove their absence. Exhaustive testing is impossible so testing must be planned and systematic. Frameworks like xUnit can help automate unit testing.
Test case prioritization usinf regression testing.pptxmaheshwari581940
This document discusses regression testing and test case prioritization techniques. It proposes prioritizing test cases based on six factors: customer priority, changes to requirements, code complexity, reusability, application flow, and fault impact. An algorithm and genetic algorithm are presented to assign weights and prioritize test cases. The approach aims to improve software quality and increase the fault detection rate. Metrics like APFD and ATEI are discussed to analyze the number of faults detected and test cases used.
Test case prioritization using firefly algorithm for software testingJournal Papers
Firefly Algorithm is applied to optimize the ordering of test cases for software testing. Test cases are represented as fireflies, with their similarity distance calculated using string metrics determining the firefly brightness. The Firefly Algorithm prioritizes test cases by moving brighter fireflies, representing more dissimilar test cases, to the front of the test sequence. Experiments on benchmark programs show the Firefly Algorithm approach achieves better or equal average percentage of faults detected and time performance compared to existing works.
Principles of design of experiments (doe)20 5-2014Awad Albalwi
This document discusses experimental design and optimization. It defines key terms like factors, responses, and residuals. It explains that experimental design is used to systematically examine problems in research, development and production. Factorial design is introduced as a method to study the effects of all factors and interactions on responses. The document provides an example experimental design to investigate if playing violent video games causes violent behavior. It outlines defining the population, randomly selecting a sample, using control and experimental conditions, measuring dependent variables, and comparing results to draw conclusions.
Principles of design of experiments (doe)20 5-2014
Performance analysis of logic
1. International Journal on Computational Sciences & Applications (IJCSA) Vol.4, No.4, August 2014
PERFORMANCE ANALYSIS OF LOGIC
COVERAGE CRITERIA USING MUMCUT
AND MINIMAL MUMCUT
Usha Badhera and Annu Maheshwari
Banasthali Vidyapith, Jaipur
ABSTRACT
Logic coverage criteria are central aspect of programs and specifications in the testing of software. A
Boolean expression with n variables in expression 2n distinct test cases is required for exhaustive testing
.This is expensive even when n is modestly large. The possible solution is to select a small subset of all
possible test cases which can effectively detect most of the common faults. Test case prioritization is one of
the key techniques for making testing cost effective. In present study performance index of test suite is
calculated for two Boolean specification testing techniques MUMCUT and Minimal-MUMCUT.
Performance index helps to measure the efficiency and determine when testing can be stopped in case of
limited resources. This paper evaluates the testability of generated single faults according to the number of
test cases used to detect them. Test cases are generated from logic expressions in irredundant normal form
(IDNF) derived from specifications or source code. The efficiency of prioritization techniques has been
validated by an empirical study done on bench mark expressions using Performance Index (PI) metric.
KEYWORDS
MUMCUT, Minimal MUMCUT, Performance Index, Prioritization, Fault
1.Introduction
Logic coverage criteria have been studied intensively both in theory and practice. Testing with a
high coverage of logic expressions always implies a high probability of detecting faults. In the
past decade, may testing criteria have been proposed for software characterized by complex
logical decisions, such as those in safety-critical software [1], [2], [3]. In such cases, it is
imperative that the logical decisions be adequately tested for the occurrence of plausible faults. In
recent years, more sophisticated coverage criteria have been advocated, like BOR (Boolean
Operator Testing Strategy),BMIS(Basic Meaningful Impact Strategy),modified
condition/decision coverage (MC/DC) [1],[2],[4] and the MUMCUT criteria[5].Compliance of
the MC/DC criterion has been mandated by Federal Aviation Administration for the approval of
airborne software. MC/DC was ‘‘developed to address the concerns of testing Boolean
expressions and can be used to guide the selection of test cases at all levels of specification, from
initial requirements to source code’’ [1]. While MC/DC is effective in fault detection, it also
consumes a great deal of testing resources [2], [6]. MUMCUT strategy is to generate test cases
that can guarantee detection of seven types of single faults provided that the original expression is
in irredundant disjunctive normal form (IDNF)[6]. In this strategy, there is no restriction on the
number and occurrence of variables in the given Boolean expressions.Minimal-MUMCUT [7]
that improves the MUMCUT strategy by considering the feasibility problem of the three testing
constituents of the MUMCUT strategy. It reduces the test suite size as compared to MUMCUT
DOI:10.5121/ijcsa.2014.4404 39
2. International Journal on Computational Sciences & Applications (IJCSA) Vol.4, No.4, August 2014
without compromising any fault detection capability. Thus, the extra tests required by the
MUMCUT criterion are of little, if any, value based on the theoretical and empirical studies
conducted [7].In this study test cases generated by Minimal-MUMCUT and MUMCUT strategies
have been used. The efficiency is measured by Performance Index (PI) using test cases generated
by both strategies and comparison has been done between them.
40
2. Preliminary and Previous Work
This section covers the basics of previous work, to introduce the notation and terminologies to be
used in this paper, as well as the main ideas that motivated this work. Section 2.1 presents
MUMCUT strategy and Section 2.2 presents Minimal MUMCUT strategy for generating test
cases from logical expressions. A well-known metric Performance Index (PI) for evaluating the
efficiency of prioritized test cases is outlined in Section 2.3. In Section 2.4, seven different types
of single faults generated from irredundant disjunctive normal form (IDNF) of original expression
and hierarchy of fault classes supported by Minimal-MUMCUT test cases for specification-based
testing is presented.
2.1 MUMCUT Strategy
Chen [8] proposed three test case selection strategies for generating test cases from Boolean
expressions.
2.1.1 Multiple Unique True Point (MUTP): The MUTP strategy selects enough test points
from UTP(S) so that all possible truth values (0 and 1) of every literal not occurring in the
term are covered for every.
Example:
3. = + . A UTP for the first term must have = 1, = 1.Tests for and to
each equal 0 and 1 are 1101 and 1110. A UTP for the second term must have = 1 and = 1.
Tests for to each equal 0 and 1 are 0111 and 1011. A test set is {1101, 1110, 0111, and
1011}.
2.1.2 Multiple Near False Point (MNFP): The MNFP strategy selects enough test points from
(
4. ) so that all possible truth values of every literal not occurring in the
term are
covered for every and every.
Example
6. )for and so that and each equal 0 and 1 are 0101, 0110,
1001, and 1010. (
7. ) for and so that and each equal 0 and 1 are 0101, 1001, 0110,
and 1010. A test set is {0101, 0110, 1001, and 1010}.
2.1.3 Corresponding Unique True Point and Near False Point pair (CUTPNFP) : Requires
the selection of a unique true point in the set of all unique true points of
term of
8. and a near
false point from the set of all near false points for the
literal in the
term of
9. such that
and only differ in the corresponding value of the
literal of the
term in
10. , for every
possible combination of and .
Example: Consider = + . A for the first term must have = 1, = 1. If = 0
and = 1, tests for the literals in are 1101, 0101, and 1001. A for the second term must
have = 1, = 1. If = 1 and = 0, tests for the literals in are 1011, 1001, and 1010. A
test set is {1101, 0101, 1001, 1011, and 1010}. The MUMCUT strategy, which integrates the
MUTP, MNFP and CUTPNFP strategies, was formally defined by Chen [1, 2].A test set T is said
11. International Journal on Computational Sciences Applications (IJCSA) Vol.4, No.4, August 2014
to satisfy the MUMCUT strategy if T satisfies all of the MUTP, MNFP and CUTPNFP strategies.
Such a test set is called a MUMCUT test set.
41
2.2 Minimal MUMCUT Strategy
The amalgamation of MUTP, MNFP and CUTPNFP strategies is referred to as the MUMCUT
strategy [6]. According to Minimal-MUMCUT, if MUTP criterion is infeasible, then prioritized
test set will be MUTP (U) test cases followed by overlapping NFPs. Overlapping NFPs is a set
covering combinatorial optimize test cases. For example: (ab)|(b!c)|(!bc) MUTPs for this
Boolean expression are : 111,010,001,101. In this expression (ab) and (b! c) terms are MUTP
infeasible and only(!bc) term is MUTP feasible which covers both values 0 and 1 for missing
literal a. The NFPs for the expression (ab)|(b!c)|(!bc) 011,101,000,010,100 but the
overlapping NFPs are 100(b of ab and c of !bc),000(b of b!c),011(a of ab and b of !bc
and c of b!c).
MUTP criterion is infeasible: Test Set = MUTP + NFP
If CUTPNFP criterion is infeasible then prioritized test set will be MUTP test cases followed by
CUTPNFP. For example: (ab)|(b!c)|(!bc) CUTPNFP criterion is infeasible for this Boolean
expression. CUTPNFPs test cases for above Boolean expression are 111,011,100,010,000,001.
CUTPNFP criterion is infeasible: Test Set = MUTP+CUTPNFP
If a MUTP criterion is feasible then prioritized test set will be MUTP (U) test cases followed by
CUTPNFP(C) followed by MNFP (N) test cases.
MUTP feasible: Test Set =MUTP+PCUTPNFP+MNFP
2.3 Performance Index
The performance of the prioritization technique used is known as effectiveness. It is necessary to
assess effectiveness of the ordering of the test suite. Since APFD (Average Percentage of Faults
Detected) denotes the global optimum of fault detection, that is, fault detection rate. This section
measures the local optimum of fault detection, which means the cost to find more faults as more
test cases are executed. [8], [9]
To demonstrate the efficiency of a coverage criterion, we combine the effectiveness with the effort
required in testing. To quantify this, we use a metric called performance index (PIT) of a minimal
test suite T which is defined as:
PIT = Number of faults revealed by test suite T/ Size of the minimal test suite T
Here PIT essentially reflects the average cost of finding a defect (i.e., the efficiency) using a test
criterion in terms of number of test cases required. A high PI score implies a low cost for finding
faults and high efficiency. A threshold can be set for the performance index. For example
threshold is 1. If the number of faults detected and the number of faults increase at the same rate,
PI will always be 1. If CI is greater than 1, means the technique is efficient at detecting faults,
because detecting more faults requires fewer tests.
2.4 Faults in Logical Expressions
A faulty implementation is referred to as single-fault expression if (1) it differs from the original
expression by one syntactic change; and (2) it is not equivalent to the original expression. Fault
12. International Journal on Computational Sciences Applications (IJCSA) Vol.4, No.4, August 2014
generation is handled by a series of complex string manipulations on JAVA Eclipse using JAVA
Collection Frameworks. The general methodology of generating faults starts with the expression,
which is just an infix string at this point, being passed through a tokenizer. The tokens then are
searched for the one that will have the fault inserted before or after.This study considers the
following classes of simple faults for logical decisions. A decision S in n variables can always be
written in disjunctive normal form (DNF) as a sum of product [8]. S= # + $% +
42
Table1. Types of Faults
Fault Description Example
Expression Negation Fault
(ENF)
The expression or sub-expression is
negated
+ + e
Term Negation Fault
(TNF)
A term is negated + + '
Term Omission Fault
(TOF)
A term is omitted + '
Operator Reference Fault
(ORF)
An OR operator(+) is implemented
as the AND operator or vice versa
. + ' or + + +
'
Literal Negation Fault
(LNF)
A literal is negated + + '
Literal Omission Fault
(LOF)
A literal is omitted + + '
Literal Insertion Fault
(LIF)
A literal is inserted + + '
Literal Reference Fault
(LRF)
A literal is implemented as another
literal
+ + '
3. Proposed Work
This section describes calculation of performance index metric for benchmark Boolean
expressions. For example consider Boolean expression (abc)| (de). Table shows the test
cases have been generated by MUMCUT and Minimal MUMCUT strategies for Boolean
expression (abc)| (de) on JAVA Eclipse using JAVA Collection Frameworks. The total
number test cases generated by MUMCUT strategy is 14 whereas total number of test cases
generated by Minimal MUMCUT is 7.
15. International Journal on Computational Sciences Applications (IJCSA) Vol.4, No.4, August 2014
45
31 (adc)|(de) 1/0 1/1 1/1 1/1 0/0 0/0 0/1
× ×
32 (aec)|(de) 1/1 1/0 1/1 1/1 0/0 0/0 0/0
×
Table3 (c): Gaining expected values for Boolean Expression (abc) | (de)
S.N. Faulty Expressions Test Cases Generated By Minimal MUMCUT strategy
11101 11110 00111 11011 01101 11010 10110
33 (aba)|(de) 1/1 1/1 1/1 1/1 0/0 0/1 0/0
×
34 (abb)|(de) 1/1 1/1 1/1 1/1 0/0 0/1 0/0
×
35 (abd)|(de) 1/0 1/1 1/1 1/1 0/0 0/1 0/0
× ×
36 (abe)|(de) 1/1 1/0 1/1 1/1 0/0 0/0 0/0
×
37 (abc)|(ae) 1/1 1/1 1/0 1/1 0/0 0/0 0/0
×
38 (abc)|(be) 1/1 1/1 1/0 1/1 0/1 0/0 0/0
× ×
39 (abc)|(ce) 1/1 1/1 1/1 1/0 0/1 0/0 0/0
× ×
40 (abc)|(ee) 1/1 1/1 1/1 1/1 0/1 0/0 0/0
×
41 (abc)|(da) 1/1 1/1 1/0 1/1 0/0 0/1 0/1
× × ×
42 (abc)|(db) 1/1 1/1 1/0 1/1 0/0 0/1 0/0
× ×
43 (abc)|(dc) 1/1 1/1 1/1 1/0 0/0 0/0 0/1
× ×
44 (abc)|(dd) 1/1 1/1 1/1 1/1 0/0 0/1 0/1
× ×
Some of the calculations for gaining expected values are given in Table 4 for Boolean expression
(abc) | (de). Table 4 shows the values of original expression when all the 7 test cases are
applied on the original expression and on the faulty expressions which are generated by Operator
Negation Fault. If the value of faulty expression is as same as the value of original expression that
means fault is not detected and if the value of faulty expression is not same as the value of
original expression that means fault is detected.
16. International Journal on Computational Sciences Applications (IJCSA) Vol.4, No.4, August 2014
46
4. Experiments and Results
In Section 4 Boolean expression (abc)| (de) generated total 44 faulty expressions and all
faults have been detected using 7 test cases generated using Minimal MUMCUT strategy. For
Boolean expression (abc)| (de) value of performance index of MUMCUT test suite (T1) is:
Number of faults revealed by MUMCUT test suite (T1) =44
Size of the MUCUT test suite =14
PIT1 = Number of faults revealed by test suite T/ Size of the minimal test suite T
PIT1 = 44/14
PIT1 = 3.14
For Boolean expression (abc)| (de) value of performance index of Minimal MUMCUT test
suite (T2) is:
Number of faults revealed by Minimal MUMCUT test suite (T2) =44
Size of the MUCUT test suite =7
PIT1 = Number of faults revealed by test suite T/ Size of the minimal test suite T
PIT1 = 44/7
PIT1 = 6.28
Table 4 shows the comparison of performance index metric values calculated using test suite
generated by MUMCUT Strategy and test suite generated by Minimal MUMCUT Strategy. This
comparison concludes that calculated Performance Index (PI) yields high value for Boolean
expressions which are not feasible for Multiple Unique True Point (MUTP) criterion.
Prioritization of test cases generated by Minimal MUMCUT yields higher Performance index
value which was introduced to evaluate the efficiency.
Table 4: Performance Index using test cases of MUMCUT Minimal MUMCUT strategies
S.N. Predicate Feasibility
Criteria
Performance Index
MUMCUT
Strategy(T1)
Minimal
MUMCUT
Strategy (T2)
1 (abc)|(de) MUTP feasible 3.14 6.28
2 (ab)|(bc) MUTP infeasible 4.33 5.2
3 (a!bd)|(a!cd)|(e) MUTP infeasible 5.16 6.88
4 (!ab)|(cd) MUTP feasible 4.4 7.33
5 (ab)|(ac)|(bc) MUTP infeasible 8.83 8.83
6 (ab)|(b!c)|(!bc) U,C,N infeasible 5.57 5.57
7 (abc)|(abd)|(!b!d)|(!de) U,C,N infeasible 7.64 10
Below graph in Figure1 shows the comparison of performance index value for all Boolean
expressions listed in Table 4 using test cases generated by MUMCUT and Minimal MUMCUT
strategies.
17. International Journal on Computational Sciences Applications (IJCSA) Vol.4, No.4, August 2014
10
9
8
7
6
5
4
3
2
1
0
6.28
3.14
5.2
4.33
E1 E2
5.16
Performance Index Value
E3 E4 E5 E6 E7
Boolean Expressions
Figure1. Comparison of Performance Index (PI) using MUMCUT
5. CONCLUSION
The Minimal-MUMCUT criterion provides better logic fault detection
analysis of test suite generated from Minimal MUMCUT and MUMCUT strategies have been
done. The Performance Index (PI) of
and compared with MUMCUT prioritized test suites.
techniques for Boolean specification for which various 2
variables can be formed. Test cases generated by Minimal MUMCUT
than the test cases generated by MUMCUT
Performance Index (PI) yields high value
Multiple Unique True Point (MUTP)
MUMCUT yields higher Performance index value which was introduced to evaluate the
efficiency. However, it is guaranteed that if even a single
MUMCUT test set, fault detection will be sac
REFERENCES
with n
trategy mostly lesser
[1] Chilenski, J.J., Miller, S.P., (1994), “Applicability of modified condition/decision coverage to software
testing” Software Engineering Journal 9 (5), 193
[2] Dupuy, A., Leveson, N., (2000), “An empirical evaluation of the MC/DC coverage criterion on the
HETE-2 satellite software” In: Proceedings of Digital Aviation Systems Conference (DASC 2000).
[3] Chilenski, J.J., (2001),“An investigation of three forms of the modified condition decision coverage
(MCDC) criterion” Tech. Rep. DOT/ FAA/AR
Department of Transportation, Washington, DC
[4] Jones, J.A., Harrold, M.J., (2003), “Test
condition/decision coverage” IEEE Transactions on Software Engineering 29 (3), 195
[5] Yu Y.T.,Lau M.F., Chen T.Y.,(2005), “Automatic generation of test cases from Boolean
specifications using the MUMCUT strategy” Journal of Systems and Software 79(6), 820
[6] Lau M.F., Chen T.Y, (2001), “Test Case Selection strategies based on Boolean Specifications”
Software Testing, Verification and Reliability, 11(3), 165
Minimal MUMCUT test suites
.In this paper performance
. test suite generated from Minimal MUMCUT is computed
Minimal MUMCUT identified testing
n distinct Boolean functions wi
strategy are
strategy. This paper concludes that calculated
for Boolean expressions which are not feasible for
criterion. Prioritization of test cases generated by Minimal
test is removed from a Minimal
sacrificed for the fault types. [6]
193–229
01),“AR-01/18, Federal Aviation Administration, US
Test-suite reduction and prioritization for modified
195–209.
CUT 165-180
4.4
8.83
5.57
7.64
6.88
7.33
8.83
5.57
10
MUMCUT
Minimal MUMCUT
47
e Minimal-
820–840.
18. International Journal on Computational Sciences Applications (IJCSA) Vol.4, No.4, August 2014
[7] Kaminski, G., Ammann, P., (2009), “Using a fault hierarchy to improve the efficiency of DNF logic
mutation testing” In Software Testing Verification and Validation, ICST’09. International Conference
on (pp. 386-395). IEEE
[8] Chen T.Y. and Lau M.F., (2001),”Test Suite reduction and fault detecting effectiveness : an empirical
evaluation”. In Proceeding of 6th International Conference on Reliable Software Technologies-Ada-
Europe 2001, number 2043, in LCNS
[9] Gupta, A., Jalote, P., (2008) “An approach for experimentally evaluating effectiveness and
efficiency of coverage criteria for software testing” International Journal on Software Tools for
Technology Transfer, 10(2), 145-160.
[10] Chen Z.Y., Fang C.R., XU B.W., (2012), “Comparing logic Coverage Criteria on test Case
48
Prioritization”.
[11] Yu Y.T., Lau M.F., (2012), “Fault based Test Suite Prioritization for Specification based Testing”,
Information and Software technology 54, 179-202