This document presents a study that develops a software effort estimation model using an Adaptive Neuro Fuzzy Inference System (ANFIS). The study evaluates the proposed ANFIS model using COCOMO81 datasets and compares its performance to an Artificial Neural Network (ANN) model and the intermediate COCOMO model. The results show that the ANFIS model provides better estimates than the ANN and COCOMO models, with lower values for metrics like the Root Mean Square Error and Magnitude of Relative Error.
Fault localization is time-consuming and difficult,
which makes it the bottleneck of the
debugging progress. To help facilitate this task, t
here exist many fault localization techniques
that help narrow down the region of the suspicious
code in a program. Better accuracy in fault
localization is achieved from heavy computation cos
t. Fault localization techniques that can
effectively locate faults also manifest slow respon
se rate. In this paper, we promote the use of
pre-computing to distribute the time-intensive comp
utations to the idle period of coding phase,
in order to speed up such techniques and achieve bo
th low-cost and high accuracy. We raise the
research problems of finding suitable techniques th
at can be pre-computed and adapt it to the
pre-computing paradigm in a continuous integration
environment. Further, we use an existing
fault localization technique to demonstrate our res
earch exploration, and shows visions and
challenges of the related methodologies.
Feature selection using modified particle swarm optimisation for face recogni...eSAT Journals
Abstract
One of the major influential factors which affects the accuracy of classification rate is the selection of right features. Not all features have vital role in classification. Many of the features in the dataset may be redundant and irrelevant, which increase the computational cost and may reduce classification rate. In this paper, we used DCT(Discrete cosine transform) coefficients as features for face recognition application. The coefficients are optimally selected based on a modified PSO algorithm. In this, the choice of coefficients is done by incorporating the average of the mean normalized standard deviations of various classes and giving more weightage to the lower indexed DCT coefficients. The algorithm is tested on ORL database. A recognition rate of 97% is obtained. Average number of features selected is about 40 percent for a 10 × 10 input. The modified PSO took about 50 iterations for convergence. These performance figures are found to be better than some of the work reported in literature.
Keywords: Particle swarm optimization, Discrete cosine transform, feature extraction, feature selection, face recognition, classification rate.
A report on designing a model for improving CPU Scheduling by using Machine L...MuskanRath1
Disclaimer: Please let me know in case some of the portions of the article match your research. I would include the link to your research in the description section of my article.
Description:
The main concern of our paper describes that we are proposing a model for a uniprocessor system for improving CPU scheduling. Our model is implemented at low-level language or assembly language and LINUX is used for the implementation of the model as it is an open-source environment and its kernel is editable.
There are several methods to predict the length of the CPU bursts, such as the exponential averaging method, however, these methods may not give accurate or reliable predicted values. In this paper, we will propose a Machine Learning (ML) based on the best approach to estimate the length of the CPU bursts for processes. We will make use of Bayesian Theory for our model as a classifier tool that will decide which process will execute first in the ready queue. The proposed approach aims to select the most significant attributes of the process using feature selection techniques and then predicts the CPU-burst for the process in the grid. Furthermore, applying attribute selection techniques improves the performance in terms of space, time, and estimation.
Software cost estimation is a key open issue for the software industry, which
suffers from cost overruns frequently. As the most popular technique for object-oriented
software cost estimation is Use Case Points (UCP) method, however, it has two major
drawbacks: the uncertainty of the cost factors and the abrupt classification. To address
these two issues, refined the use case complexity classification using fuzzy logic theory which
mitigate the uncertainty of cost factors and improve the accuracy of classification.
Software estimation is a crucial task in software engineering. Software estimation
encompasses cost, effort, schedule, and size. The importance of software estimation becomes
critical in the early stages of the software life cycle when the details of software have not
been revealed yet. Several commercial and non-commercial tools exist to estimate software
in the early stages. Most software effort estimation methods require software size as one of
the important metric inputs and consequently, software size estimation in the early stages
becomes essential.
The proposed method presents a techniques using fuzzy logic theory to improve the
accuracy of the use case points method by refining the use case classification.
Fault localization is time-consuming and difficult,
which makes it the bottleneck of the
debugging progress. To help facilitate this task, t
here exist many fault localization techniques
that help narrow down the region of the suspicious
code in a program. Better accuracy in fault
localization is achieved from heavy computation cos
t. Fault localization techniques that can
effectively locate faults also manifest slow respon
se rate. In this paper, we promote the use of
pre-computing to distribute the time-intensive comp
utations to the idle period of coding phase,
in order to speed up such techniques and achieve bo
th low-cost and high accuracy. We raise the
research problems of finding suitable techniques th
at can be pre-computed and adapt it to the
pre-computing paradigm in a continuous integration
environment. Further, we use an existing
fault localization technique to demonstrate our res
earch exploration, and shows visions and
challenges of the related methodologies.
Feature selection using modified particle swarm optimisation for face recogni...eSAT Journals
Abstract
One of the major influential factors which affects the accuracy of classification rate is the selection of right features. Not all features have vital role in classification. Many of the features in the dataset may be redundant and irrelevant, which increase the computational cost and may reduce classification rate. In this paper, we used DCT(Discrete cosine transform) coefficients as features for face recognition application. The coefficients are optimally selected based on a modified PSO algorithm. In this, the choice of coefficients is done by incorporating the average of the mean normalized standard deviations of various classes and giving more weightage to the lower indexed DCT coefficients. The algorithm is tested on ORL database. A recognition rate of 97% is obtained. Average number of features selected is about 40 percent for a 10 × 10 input. The modified PSO took about 50 iterations for convergence. These performance figures are found to be better than some of the work reported in literature.
Keywords: Particle swarm optimization, Discrete cosine transform, feature extraction, feature selection, face recognition, classification rate.
A report on designing a model for improving CPU Scheduling by using Machine L...MuskanRath1
Disclaimer: Please let me know in case some of the portions of the article match your research. I would include the link to your research in the description section of my article.
Description:
The main concern of our paper describes that we are proposing a model for a uniprocessor system for improving CPU scheduling. Our model is implemented at low-level language or assembly language and LINUX is used for the implementation of the model as it is an open-source environment and its kernel is editable.
There are several methods to predict the length of the CPU bursts, such as the exponential averaging method, however, these methods may not give accurate or reliable predicted values. In this paper, we will propose a Machine Learning (ML) based on the best approach to estimate the length of the CPU bursts for processes. We will make use of Bayesian Theory for our model as a classifier tool that will decide which process will execute first in the ready queue. The proposed approach aims to select the most significant attributes of the process using feature selection techniques and then predicts the CPU-burst for the process in the grid. Furthermore, applying attribute selection techniques improves the performance in terms of space, time, and estimation.
Software cost estimation is a key open issue for the software industry, which
suffers from cost overruns frequently. As the most popular technique for object-oriented
software cost estimation is Use Case Points (UCP) method, however, it has two major
drawbacks: the uncertainty of the cost factors and the abrupt classification. To address
these two issues, refined the use case complexity classification using fuzzy logic theory which
mitigate the uncertainty of cost factors and improve the accuracy of classification.
Software estimation is a crucial task in software engineering. Software estimation
encompasses cost, effort, schedule, and size. The importance of software estimation becomes
critical in the early stages of the software life cycle when the details of software have not
been revealed yet. Several commercial and non-commercial tools exist to estimate software
in the early stages. Most software effort estimation methods require software size as one of
the important metric inputs and consequently, software size estimation in the early stages
becomes essential.
The proposed method presents a techniques using fuzzy logic theory to improve the
accuracy of the use case points method by refining the use case classification.
A COMBINATION OF PALMER ALGORITHM AND GUPTA ALGORITHM FOR SCHEDULING PROBLEM ...ijfls
The apparel industry is a class of textile industry. Generally, the production scheduling problem in the apparel industry belongs to Flow Shop Scheduling Problems (FSSP). There are many algorithms/techniques/heuristics for solving FSSP. Two of them are the Palmer Algorithm and the Gupta Algorithm. Hyper-heuristic is a class of heuristics that enables to combine of some heuristics to produce a new heuristic. GPHH is a hyper-heuristic that is based on genetic programming that is proposed to solve FSSP [1]. This paper presents the development of a computer program that implements the GPHH. Some experiments have been conducted for measuring the performance of GPHH. From the experimental results, GPHH has shown a better performance than the Palmer Algorithm and Gupta Algorithm.
BINARY SINE COSINE ALGORITHMS FOR FEATURE SELECTION FROM MEDICAL DATAacijjournal
A well-constructed classification model highly depends on input feature subsets from a dataset, which may contain redundant, irrelevant, or noisy features. This challenge can be worse while dealing with medical datasets. The main aim of feature selection as a pre-processing task is to eliminate these features and select the most effective ones. In the literature, metaheuristic algorithms show a successful performance to find optimal feature subsets. In this paper, two binary metaheuristic algorithms named S-shaped binary Sine Cosine Algorithm (SBSCA) and V-shaped binary Sine Cosine Algorithm (VBSCA) are proposed for feature selection from the medical data. In these algorithms, the search space remains continuous, while a binary position vector is generated by two transfer functions S-shaped and V-shaped for each solution. The proposed algorithms are compared with four latest binary optimization algorithms over five medical datasets from the UCI repository. The experimental results confirm that using both bSCA variants enhance the accuracy of classification on these medical datasets compared to four other algorithms.
Software reliability models (SRMs) are very important for estimating and predicting software
reliability in the testing/debugging phase. The contributions of this paper are as follows. First, a
historical review of the Gompertz SRM is given. Based on several software failure data, the
parameters of the Gompertz software reliability model are estimated using two estimation
methods, the traditional maximum likelihood and the least square. The methods of estimation are
evaluated using the MSE and R-squared criteria. The results show that the least square
estimation is an attractive method in term of predictive performance and can be used when the
maximum likelihood method fails to give good prediction results.
ESTIMATING HANDLING TIME OF SOFTWARE DEFECTScsandit
The problem of accurately predicting handling time for software defects is of great practical
importance. However, it is difficult to suggest a practical generic algorithm for such estimates,
due in part to the limited information available when opening a defect and the lack of a uniform
standard for defect structure. We suggest an algorithm to address these challenges that is
implementable over different defect management tools. Our algorithm uses machine learning
regression techniques to predict the handling time of defects based on past behaviour of similar
defects. The algorithm relies only on a minimal set of assumptions about the structure of the
input data. We show how an implementation of this algorithm predicts defect handling time with
promising accuracy results
Surrogate-based design is an effective approach for modeling computationally expensive system behavior. In such application, it is often challenging to characterize the expected accuracy of the surrogate. In addition to global and local error measures, regional error measures can be used to understand and interpret the surrogate accuracy in the regions of interest. This paper develops the Regional Error Estimation of Surrogate (REES) method to quantify the level of the error in any given subspace (or region) of the entire domain, when all the available training points have been invested to build the surrogate. In this approach, the accuracy of the surrogate in each subspace is estimated by modeling the variations of the mean and the maximum error in that subspace with increasing number of training points (in an iterative process). A regression model is used for this purpose. At each iteration, the intermediate surrogate is constructed using a subset of the entire training data, and tested over the remaining points. The evaluated errors at the intermediate test points at each iteration are used for training the regression model that represents the error variation with sample points. The effectiveness of the proposed method is illustrated using standard test problems. To this end, the predicted regional errors of the surrogate constructed using all the training points are compared with the regional errors estimated over a large set of test points.
Review on Algorithmic and Non Algorithmic Software Cost Estimation Techniquesijtsrd
Effective software cost estimation is the most challenging and important activities in software development. Developers want a simple and accurate method of efforts estimation. Estimation of the cost before starting of work is a prediction and prediction always not accurate. Software effort estimation is a very critical task in the software engineering and to control quality and efficiency a suitable estimation technique is crucial. This paper gives a review of various available software effort estimation methods, mainly focus on the algorithmic model and non algorithmic model. These existing methods for software cost estimation are illustrated and their aspect will be discussed. No single technique is best for all situations, and thus a careful comparison of the results of several approaches is most likely to produce realistic estimation. This paper provides a detailed overview of existing software cost estimation models and techniques. This paper presents the strength and weakness of various cost estimation methods. This paper focuses on some of the relevant reasons that cause inaccurate estimation. Pa Pa Win | War War Myint | Hlaing Phyu Phyu Mon | Seint Wint Thu "Review on Algorithmic and Non-Algorithmic Software Cost Estimation Techniques" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-3 | Issue-5 , August 2019, URL: https://www.ijtsrd.com/papers/ijtsrd26511.pdfPaper URL: https://www.ijtsrd.com/engineering/-/26511/review-on-algorithmic-and-non-algorithmic-software-cost-estimation-techniques/pa-pa-win
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
This paper proposes a novel model management technique to be applied in population- based heuristic optimization. This technique adaptively selects different computational models (both physics-based and statistical models) to be used during optimization, with the overall goal to end with high fidelity solutions in a reasonable time period. For example, in optimizing an aircraft wing to obtain maximum lift-to-drag ratio, one can use low-fidelity models such as given by the vortex lattice method, or a high-fidelity finite volume model (that solves the full Navier-Stokes equations), or a surrogate model that substitutes the high-fidelity model.The information from models with different levels of fidelity is inte- grated into the heuristic optimization process using a novel model-switching metric. In this context, models could be surrogate models, low-fidelity physics-based analytical mod- els, and medium-to-high fidelity computational models (based on grid density). The model switching technique replaces the current model with the next higher fidelity model, when a stochastic switching criterion is met at a given iteration during the optimization process. The switching criteria is based on whether the uncertainty associated with the current model output dominates the latest improvement of the fitness function. In the case of the physics-based models, the uncertainty in their output is quantified through an inverse assessment process by comparing with high-fidelity model responses or experimental data (if available). To determine the fidelity of surrogate models, the Predictive Estimation of Model Fidelity (PEMF) method is applied. The effectiveness of the proposed method is demonstrated by applying it to airfoil optimization with the objective to maximize the lift to drag ratio of the wing under different flow regimes. It was found that the tuned low fidelity model dominates the optimization process in terms of computational time and function calls.
Analysis of selection schemes for solving job shop scheduling problem using g...eSAT Journals
Abstract Scheduling problems have the standard consideration in the field of manufacturing. Among the various types of scheduling problems, the job shop scheduling problem is one of the most interesting NP-hard problems. As the job shop scheduling is an optimization problem, Genetic algorithm was selected to solve it In this study. Selection scheme is one of the important operators of Genetic algorithm. The choice of selection method to be applied for solving problems has a wide role in the Genetic algorithm process. The speed of convergence towards the optimum solution for the chosen problem is largely determined by the selection mechanism used in the Genetic algorithm. Depending upon the selection scheme applied, the population fitness over the successive generations could be improved. There are various type of selection schemes in genetic algorithm are available, where each selection scheme has its own feasibility for solving a particular problem. In this study, the selection schemes namely Stochastic Universal Sampling (SUS), Roulette Wheel Selection (RWS), Rank Based Roulette Wheel Selection (RRWS) and Binary Tournament Selection (BTS) were chosen for implementation. The characteristics of chosen selection mechanisms of Genetic algorithm for solving the job shop scheduling problem were analyzed. The Genetic algorithm with four different selection schemes is tested on instances of 7 benchmark problems of different size. The result shows that the each of the four selection schemes of Genetic algorithm have been successfully applied to the job shop scheduling problems efficiently and the performance of Stochastic Universal Sampling selection method is better than all other four selection schemes. Keywords: Genetic Algorithm, Makespan, Selection schemes
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology.
This paper advances the Domain Segmentation based on Uncertainty in the Surrogate (DSUS) framework which is a novel approach to characterize the uncertainty in surrogates. The leave-one-out cross-validation technique is adopted in the DSUS framework to measure local errors of a surrogate. A method is proposed in this paper to evaluate the performance of the leave-out-out cross-validation errors as local error measures. This method evaluates local errors by comparing: (i) the leave-one-out cross-validation error with (ii) the actual local error estimated within a local hypercube for each training point. The comparison results show that the leave-one-out cross-validation strategy can capture the local errors of a surrogate. The DSUS framework is then applied to key aspects of wind resource as- sessment and wind farm cost modeling. The uncertainties in the wind farm cost and the wind power potential are successfully characterized, which provides designers/users more confidence when using these models
Owing to the multitude of surrogate modeling techniques, developed in the recent years and the diverse characteristics offered by them, automated adaptive model selection ap- proaches could be helpful in selecting the most suitable surrogate for a given problem. Surrogate selection could be performed at three different levels: (i) model type selection, (ii) basis (or kernel) function selection, and (iii) hyper-parameter selection where hyper- parameters are those kernel parameters that are generally given by the users. Unlike the majority of existing model selection techniques, this paper explores the development of a method that performs selection coherently at all the three levels. In this context, the REES method is used to provide measures of the median and maximum errors of a candi- date surrogate model. Two approaches are used for the 3-level selection; (i) A Cascaded approach performs each level in a nested loop in the order going from model-kernel-hyper- parameters; (ii) A more advanced One-Step approach solves a MINLP to simultaneously optimize the model, kernel, and hyper-parameters. In both approaches, multiobjective optimization is performed to yield the best trade-offs between the estimated median and maximum errors. Candidate surrogates that are considered include (i) Kriging, (ii) Radial Basis Function (RBF), and (iii) Support Vector Regression (SVR), and multiple candidate kernels are allowed within these surrogate models. The 3-level REES-based model selec- tion is compared with model selection based on error estimated on a large set of additional test points, for validation purposes. Numerical experiments on a 2-variable, 6-variable, and 18-variable test problems, and wind farm power generation problem, show that the proposed approach provides unique flexibility in model selection and is also reasonably ac- curate when compared with selection based on errors estimated on additional test points.
Computer aided classification of Bascal cell carcinoma using adaptive Neuro-f...Editor IJMTER
For skin lesion detection pathologists examine biopsies to make diagnostic
assessment largely based on cell anatomy and tissue distribution. However in many
instances it is subjective and often leads to considerable variability. Whereas computer
diagnostic tools enable objective judgments by making use of quantitative measures.
Paper presents a diagnosis system based on an adaptive Neuro-fuzzy inference system
for effective classification of Bascal cell carcinoma images from the given set of all types
of skin lesions. System divide in three parts. Image Processing, Feature Extraction, and
classification. First part deals with the noise reduction and artifacts removing from the
set of images. Second part deals with extracting variety of features of Bascal Cell
Carcinoma using the Greedy feature flip algorithm (G-flip), and classification method
using ANFIS algorithm and finally Part three deals with the results that is classification
of BCC images from the variety of pre-cancerous stage images that is Actinic Keratosis
and also other images called psoriasis which looks as cancer images at a first look . The
results confirmed that the proposed ANFIS model has potential in classifying the skin
cancer diagnosis.
Approximation models (or surrogate models) provide an efficient substitute to expen- sive physical simulations and an efficient solution to the lack of physical models of system behavior. However, it is challenging to quantify the accuracy and reliability of such ap- proximation models in a region of interest or the overall domain without additional system evaluations. Standard error measures, such as the mean squared error, the cross-validation error, and the Akaikes information criterion, provide limited (often inadequate) informa- tion regarding the accuracy of the final surrogate. This paper introduces a novel and model independent concept to quantify the level of errors in the function value estimated by the final surrogate in any given region of the design domain. This method is called the Re- gional Error Estimation of Surrogate (REES). Assuming the full set of available sample points to be fixed, intermediate surrogates are iteratively constructed over a sample set comprising all samples outside the region of interest and heuristic subsets of samples inside the region of interest (i.e., intermediate training points). The intermediate surrogate is tested over the remaining sample points inside the region of interest (i.e., intermediate test points). The fraction of sample points inside region of interest, which are used as interme- diate training points, is fixed at each iteration, with the total number of iterations being pre-specified. The estimated median and maximum relative errors within the region of in- terest for the heuristic subsets at each iteration are used to fit a distribution of the median and maximum error, respectively. The estimated statistical mode of the median and the maximum error, and the absolute maximum error are then represented as functions of the density of intermediate training points, using regression models. The regression models are then used to predict the expected median and maximum regional errors when all the sample points are used as training points. Standard test functions and a wind farm power generation problem are used to illustrate the effectiveness and the utility of such a regional error quantification method.
A comparative study on remote tracking of parkinson’s disease progression usi...ijfcstjournal
In recent years, applications of data mining method
s are become more popular in many fields of medical
diagnosis and evaluations. The data mining methods
are appropriate tools for discovering and extractin
g
of available knowledge in medical databases. In thi
s study, we divided 11 data mining algorithms into
five
groups which are applied to a dataset of patient’s
clinical variables data with Parkinson’s Disease (P
D) to
study the disease progression. The dataset includes
22 properties of 42 people that all of our algorit
hms
are applied to this dataset. The Decision Table wit
h 0.9985 correlation coefficients has the best accu
racy
and Decision Stump with 0.7919 correlation coeffici
ents has the lowest accuracy.
Distribution of maximal clique size underijfcstjournal
In this paper, we analyze the evolution of a small-world network and its subsequent transformation to a
random network using the idea of link rewiring under the well-known Watts-Strogatz model for complex
networks. Every link u-v in the regular network is considered for rewiring with a certain probability and if
chosen for rewiring, the link u-v is removed from the network and the node u is connected to a randomly
chosen node w (other than nodes u and v). Our objective in this paper is to analyze the distribution of the
maximal clique size per node by varying the probability of link rewiring and the degree per node (number
of links incident on a node) in the initial regular network. For a given probability of rewiring and initial
number of links per node, we observe the distribution of the maximal clique per node to follow a Poisson
distribution. We also observe the maximal clique size per node in the small-world network to be very close
to that of the average value and close to that of the maximal clique size in a regular network. There is no
appreciable decrease in the maximal clique size per node when the network transforms from a regular
network to a small-world network. On the other hand, when the network transforms from a small-world
network to a random network, the average maximal clique size value decreases significantly.
An interactive approach to requirements prioritization using quality factorsijfcstjournal
As the prevalence of software increases, so does the complexity and the number of requirements assoc
iated
to the software project. This presents a dilemma for the developers to clearly identify and prioriti
ze the
most important requirements in order to del
iver the project in given amount of resources and time.
A
number of prioritization methods have been proposed which provide consistent results, but they are v
ery
difficult and complex to implement in practical scenarios as well as lack proper structure to
analyze the
requirements properly. In this study, the users can provide their requirements in two forms: text ba
sed
story form and use case form.
Moreover, the existing prioritization techniques have a very little or no
interaction with the users. So, in t
his paper an attempt has been made to make the prioritization process
user interactive by adding a second level of prioritization where after the developer has properly a
nalyzed
and ranked the requirements on the basis of quality attributes in the first le
vel, takes the opinion of distinct
user’s about the requirements priority sequence. The developer then calculates the disagreement valu
e
associated with each user sequence in order to find out the final priority sequence.
A COMBINATION OF PALMER ALGORITHM AND GUPTA ALGORITHM FOR SCHEDULING PROBLEM ...ijfls
The apparel industry is a class of textile industry. Generally, the production scheduling problem in the apparel industry belongs to Flow Shop Scheduling Problems (FSSP). There are many algorithms/techniques/heuristics for solving FSSP. Two of them are the Palmer Algorithm and the Gupta Algorithm. Hyper-heuristic is a class of heuristics that enables to combine of some heuristics to produce a new heuristic. GPHH is a hyper-heuristic that is based on genetic programming that is proposed to solve FSSP [1]. This paper presents the development of a computer program that implements the GPHH. Some experiments have been conducted for measuring the performance of GPHH. From the experimental results, GPHH has shown a better performance than the Palmer Algorithm and Gupta Algorithm.
BINARY SINE COSINE ALGORITHMS FOR FEATURE SELECTION FROM MEDICAL DATAacijjournal
A well-constructed classification model highly depends on input feature subsets from a dataset, which may contain redundant, irrelevant, or noisy features. This challenge can be worse while dealing with medical datasets. The main aim of feature selection as a pre-processing task is to eliminate these features and select the most effective ones. In the literature, metaheuristic algorithms show a successful performance to find optimal feature subsets. In this paper, two binary metaheuristic algorithms named S-shaped binary Sine Cosine Algorithm (SBSCA) and V-shaped binary Sine Cosine Algorithm (VBSCA) are proposed for feature selection from the medical data. In these algorithms, the search space remains continuous, while a binary position vector is generated by two transfer functions S-shaped and V-shaped for each solution. The proposed algorithms are compared with four latest binary optimization algorithms over five medical datasets from the UCI repository. The experimental results confirm that using both bSCA variants enhance the accuracy of classification on these medical datasets compared to four other algorithms.
Software reliability models (SRMs) are very important for estimating and predicting software
reliability in the testing/debugging phase. The contributions of this paper are as follows. First, a
historical review of the Gompertz SRM is given. Based on several software failure data, the
parameters of the Gompertz software reliability model are estimated using two estimation
methods, the traditional maximum likelihood and the least square. The methods of estimation are
evaluated using the MSE and R-squared criteria. The results show that the least square
estimation is an attractive method in term of predictive performance and can be used when the
maximum likelihood method fails to give good prediction results.
ESTIMATING HANDLING TIME OF SOFTWARE DEFECTScsandit
The problem of accurately predicting handling time for software defects is of great practical
importance. However, it is difficult to suggest a practical generic algorithm for such estimates,
due in part to the limited information available when opening a defect and the lack of a uniform
standard for defect structure. We suggest an algorithm to address these challenges that is
implementable over different defect management tools. Our algorithm uses machine learning
regression techniques to predict the handling time of defects based on past behaviour of similar
defects. The algorithm relies only on a minimal set of assumptions about the structure of the
input data. We show how an implementation of this algorithm predicts defect handling time with
promising accuracy results
Surrogate-based design is an effective approach for modeling computationally expensive system behavior. In such application, it is often challenging to characterize the expected accuracy of the surrogate. In addition to global and local error measures, regional error measures can be used to understand and interpret the surrogate accuracy in the regions of interest. This paper develops the Regional Error Estimation of Surrogate (REES) method to quantify the level of the error in any given subspace (or region) of the entire domain, when all the available training points have been invested to build the surrogate. In this approach, the accuracy of the surrogate in each subspace is estimated by modeling the variations of the mean and the maximum error in that subspace with increasing number of training points (in an iterative process). A regression model is used for this purpose. At each iteration, the intermediate surrogate is constructed using a subset of the entire training data, and tested over the remaining points. The evaluated errors at the intermediate test points at each iteration are used for training the regression model that represents the error variation with sample points. The effectiveness of the proposed method is illustrated using standard test problems. To this end, the predicted regional errors of the surrogate constructed using all the training points are compared with the regional errors estimated over a large set of test points.
Review on Algorithmic and Non Algorithmic Software Cost Estimation Techniquesijtsrd
Effective software cost estimation is the most challenging and important activities in software development. Developers want a simple and accurate method of efforts estimation. Estimation of the cost before starting of work is a prediction and prediction always not accurate. Software effort estimation is a very critical task in the software engineering and to control quality and efficiency a suitable estimation technique is crucial. This paper gives a review of various available software effort estimation methods, mainly focus on the algorithmic model and non algorithmic model. These existing methods for software cost estimation are illustrated and their aspect will be discussed. No single technique is best for all situations, and thus a careful comparison of the results of several approaches is most likely to produce realistic estimation. This paper provides a detailed overview of existing software cost estimation models and techniques. This paper presents the strength and weakness of various cost estimation methods. This paper focuses on some of the relevant reasons that cause inaccurate estimation. Pa Pa Win | War War Myint | Hlaing Phyu Phyu Mon | Seint Wint Thu "Review on Algorithmic and Non-Algorithmic Software Cost Estimation Techniques" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-3 | Issue-5 , August 2019, URL: https://www.ijtsrd.com/papers/ijtsrd26511.pdfPaper URL: https://www.ijtsrd.com/engineering/-/26511/review-on-algorithmic-and-non-algorithmic-software-cost-estimation-techniques/pa-pa-win
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
This paper proposes a novel model management technique to be applied in population- based heuristic optimization. This technique adaptively selects different computational models (both physics-based and statistical models) to be used during optimization, with the overall goal to end with high fidelity solutions in a reasonable time period. For example, in optimizing an aircraft wing to obtain maximum lift-to-drag ratio, one can use low-fidelity models such as given by the vortex lattice method, or a high-fidelity finite volume model (that solves the full Navier-Stokes equations), or a surrogate model that substitutes the high-fidelity model.The information from models with different levels of fidelity is inte- grated into the heuristic optimization process using a novel model-switching metric. In this context, models could be surrogate models, low-fidelity physics-based analytical mod- els, and medium-to-high fidelity computational models (based on grid density). The model switching technique replaces the current model with the next higher fidelity model, when a stochastic switching criterion is met at a given iteration during the optimization process. The switching criteria is based on whether the uncertainty associated with the current model output dominates the latest improvement of the fitness function. In the case of the physics-based models, the uncertainty in their output is quantified through an inverse assessment process by comparing with high-fidelity model responses or experimental data (if available). To determine the fidelity of surrogate models, the Predictive Estimation of Model Fidelity (PEMF) method is applied. The effectiveness of the proposed method is demonstrated by applying it to airfoil optimization with the objective to maximize the lift to drag ratio of the wing under different flow regimes. It was found that the tuned low fidelity model dominates the optimization process in terms of computational time and function calls.
Analysis of selection schemes for solving job shop scheduling problem using g...eSAT Journals
Abstract Scheduling problems have the standard consideration in the field of manufacturing. Among the various types of scheduling problems, the job shop scheduling problem is one of the most interesting NP-hard problems. As the job shop scheduling is an optimization problem, Genetic algorithm was selected to solve it In this study. Selection scheme is one of the important operators of Genetic algorithm. The choice of selection method to be applied for solving problems has a wide role in the Genetic algorithm process. The speed of convergence towards the optimum solution for the chosen problem is largely determined by the selection mechanism used in the Genetic algorithm. Depending upon the selection scheme applied, the population fitness over the successive generations could be improved. There are various type of selection schemes in genetic algorithm are available, where each selection scheme has its own feasibility for solving a particular problem. In this study, the selection schemes namely Stochastic Universal Sampling (SUS), Roulette Wheel Selection (RWS), Rank Based Roulette Wheel Selection (RRWS) and Binary Tournament Selection (BTS) were chosen for implementation. The characteristics of chosen selection mechanisms of Genetic algorithm for solving the job shop scheduling problem were analyzed. The Genetic algorithm with four different selection schemes is tested on instances of 7 benchmark problems of different size. The result shows that the each of the four selection schemes of Genetic algorithm have been successfully applied to the job shop scheduling problems efficiently and the performance of Stochastic Universal Sampling selection method is better than all other four selection schemes. Keywords: Genetic Algorithm, Makespan, Selection schemes
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology.
This paper advances the Domain Segmentation based on Uncertainty in the Surrogate (DSUS) framework which is a novel approach to characterize the uncertainty in surrogates. The leave-one-out cross-validation technique is adopted in the DSUS framework to measure local errors of a surrogate. A method is proposed in this paper to evaluate the performance of the leave-out-out cross-validation errors as local error measures. This method evaluates local errors by comparing: (i) the leave-one-out cross-validation error with (ii) the actual local error estimated within a local hypercube for each training point. The comparison results show that the leave-one-out cross-validation strategy can capture the local errors of a surrogate. The DSUS framework is then applied to key aspects of wind resource as- sessment and wind farm cost modeling. The uncertainties in the wind farm cost and the wind power potential are successfully characterized, which provides designers/users more confidence when using these models
Owing to the multitude of surrogate modeling techniques, developed in the recent years and the diverse characteristics offered by them, automated adaptive model selection ap- proaches could be helpful in selecting the most suitable surrogate for a given problem. Surrogate selection could be performed at three different levels: (i) model type selection, (ii) basis (or kernel) function selection, and (iii) hyper-parameter selection where hyper- parameters are those kernel parameters that are generally given by the users. Unlike the majority of existing model selection techniques, this paper explores the development of a method that performs selection coherently at all the three levels. In this context, the REES method is used to provide measures of the median and maximum errors of a candi- date surrogate model. Two approaches are used for the 3-level selection; (i) A Cascaded approach performs each level in a nested loop in the order going from model-kernel-hyper- parameters; (ii) A more advanced One-Step approach solves a MINLP to simultaneously optimize the model, kernel, and hyper-parameters. In both approaches, multiobjective optimization is performed to yield the best trade-offs between the estimated median and maximum errors. Candidate surrogates that are considered include (i) Kriging, (ii) Radial Basis Function (RBF), and (iii) Support Vector Regression (SVR), and multiple candidate kernels are allowed within these surrogate models. The 3-level REES-based model selec- tion is compared with model selection based on error estimated on a large set of additional test points, for validation purposes. Numerical experiments on a 2-variable, 6-variable, and 18-variable test problems, and wind farm power generation problem, show that the proposed approach provides unique flexibility in model selection and is also reasonably ac- curate when compared with selection based on errors estimated on additional test points.
Computer aided classification of Bascal cell carcinoma using adaptive Neuro-f...Editor IJMTER
For skin lesion detection pathologists examine biopsies to make diagnostic
assessment largely based on cell anatomy and tissue distribution. However in many
instances it is subjective and often leads to considerable variability. Whereas computer
diagnostic tools enable objective judgments by making use of quantitative measures.
Paper presents a diagnosis system based on an adaptive Neuro-fuzzy inference system
for effective classification of Bascal cell carcinoma images from the given set of all types
of skin lesions. System divide in three parts. Image Processing, Feature Extraction, and
classification. First part deals with the noise reduction and artifacts removing from the
set of images. Second part deals with extracting variety of features of Bascal Cell
Carcinoma using the Greedy feature flip algorithm (G-flip), and classification method
using ANFIS algorithm and finally Part three deals with the results that is classification
of BCC images from the variety of pre-cancerous stage images that is Actinic Keratosis
and also other images called psoriasis which looks as cancer images at a first look . The
results confirmed that the proposed ANFIS model has potential in classifying the skin
cancer diagnosis.
Approximation models (or surrogate models) provide an efficient substitute to expen- sive physical simulations and an efficient solution to the lack of physical models of system behavior. However, it is challenging to quantify the accuracy and reliability of such ap- proximation models in a region of interest or the overall domain without additional system evaluations. Standard error measures, such as the mean squared error, the cross-validation error, and the Akaikes information criterion, provide limited (often inadequate) informa- tion regarding the accuracy of the final surrogate. This paper introduces a novel and model independent concept to quantify the level of errors in the function value estimated by the final surrogate in any given region of the design domain. This method is called the Re- gional Error Estimation of Surrogate (REES). Assuming the full set of available sample points to be fixed, intermediate surrogates are iteratively constructed over a sample set comprising all samples outside the region of interest and heuristic subsets of samples inside the region of interest (i.e., intermediate training points). The intermediate surrogate is tested over the remaining sample points inside the region of interest (i.e., intermediate test points). The fraction of sample points inside region of interest, which are used as interme- diate training points, is fixed at each iteration, with the total number of iterations being pre-specified. The estimated median and maximum relative errors within the region of in- terest for the heuristic subsets at each iteration are used to fit a distribution of the median and maximum error, respectively. The estimated statistical mode of the median and the maximum error, and the absolute maximum error are then represented as functions of the density of intermediate training points, using regression models. The regression models are then used to predict the expected median and maximum regional errors when all the sample points are used as training points. Standard test functions and a wind farm power generation problem are used to illustrate the effectiveness and the utility of such a regional error quantification method.
A comparative study on remote tracking of parkinson’s disease progression usi...ijfcstjournal
In recent years, applications of data mining method
s are become more popular in many fields of medical
diagnosis and evaluations. The data mining methods
are appropriate tools for discovering and extractin
g
of available knowledge in medical databases. In thi
s study, we divided 11 data mining algorithms into
five
groups which are applied to a dataset of patient’s
clinical variables data with Parkinson’s Disease (P
D) to
study the disease progression. The dataset includes
22 properties of 42 people that all of our algorit
hms
are applied to this dataset. The Decision Table wit
h 0.9985 correlation coefficients has the best accu
racy
and Decision Stump with 0.7919 correlation coeffici
ents has the lowest accuracy.
Distribution of maximal clique size underijfcstjournal
In this paper, we analyze the evolution of a small-world network and its subsequent transformation to a
random network using the idea of link rewiring under the well-known Watts-Strogatz model for complex
networks. Every link u-v in the regular network is considered for rewiring with a certain probability and if
chosen for rewiring, the link u-v is removed from the network and the node u is connected to a randomly
chosen node w (other than nodes u and v). Our objective in this paper is to analyze the distribution of the
maximal clique size per node by varying the probability of link rewiring and the degree per node (number
of links incident on a node) in the initial regular network. For a given probability of rewiring and initial
number of links per node, we observe the distribution of the maximal clique per node to follow a Poisson
distribution. We also observe the maximal clique size per node in the small-world network to be very close
to that of the average value and close to that of the maximal clique size in a regular network. There is no
appreciable decrease in the maximal clique size per node when the network transforms from a regular
network to a small-world network. On the other hand, when the network transforms from a small-world
network to a random network, the average maximal clique size value decreases significantly.
An interactive approach to requirements prioritization using quality factorsijfcstjournal
As the prevalence of software increases, so does the complexity and the number of requirements assoc
iated
to the software project. This presents a dilemma for the developers to clearly identify and prioriti
ze the
most important requirements in order to del
iver the project in given amount of resources and time.
A
number of prioritization methods have been proposed which provide consistent results, but they are v
ery
difficult and complex to implement in practical scenarios as well as lack proper structure to
analyze the
requirements properly. In this study, the users can provide their requirements in two forms: text ba
sed
story form and use case form.
Moreover, the existing prioritization techniques have a very little or no
interaction with the users. So, in t
his paper an attempt has been made to make the prioritization process
user interactive by adding a second level of prioritization where after the developer has properly a
nalyzed
and ranked the requirements on the basis of quality attributes in the first le
vel, takes the opinion of distinct
user’s about the requirements priority sequence. The developer then calculates the disagreement valu
e
associated with each user sequence in order to find out the final priority sequence.
LOGMIN: A MODEL FOR CALL LOG MINING IN MOBILE DEVICESijfcstjournal
In today’s instant communication era, mobile phones play an important role in the efficient communication with respect to both individual and official communication strata. With the drastic explosion in the quantity of calls received and made, there is a need for analyses of patterns in these call logs to assist the user of the
mobile device in the optimal utilization. This paper proposes a model termed “LogMin” (Log Mining of Calls in Mobile devices) which is aimed towards mining of call log in mobile phones to discover patterns and keep the user informed about the trends in the log. The logging of calls would facilitate the user to get an insight into patterns based on the six different parameters identified by the proposed LogMin model. The proposed model is validated with a prototype implementation in the Android platform and various experiments were conducted on it. The results of the experiments in the LogMin Android implementation validate the efficiency of the proposed model with respect to user’s relevancy metric which is computed as 96.52%.
The Major Occupation in India is the Agriculture; the people involved in the Agriculture belong to the poor
class and category. The people of the farming community are unaware of the new techniques and Agromachines,
which would direct the world to greater heights in the field of agriculture. Though the farmers
work hard, they are cheated by agents in today’s market. This serves as a opportunity to solve
all the problems that farmers face in the current world. The eAgro crop marketing will serve as a better
way for the farmers to sell their products within the country with some mediocre knowledge about using
the website. This would provide information to the farmers about current market rate of agro-products,
their sale history and profits earned in a sale. This site will also help the farmers to know about the market
information and to view agricultural schemes of the Government provided to farmers.
Migration strategies for object oriented system to component based systemijfcstjournal
Migration of object oriented system to component based System is not an easy task, not only technically a
lot of changes needs to be done but also numerous other issues needs to be kept in mind. However Component based Software development has been gaining its popularity from the past few years and has higher reusability scope. Programs built using CBSE approach are confirmed to be suitable to new environments. These days it’s a universal practice to reuse components in project to achieve better quality and to save time. So moving to CBSE from object oriented seems wise decision. Number of approaches has been introduced to implement this and each one of them has its own pros and cons. The paper focuses on the brief review on works of different authors in this area from the year 2000 to 2014.
Defragmentation of indian legal cases withijfcstjournal
The main aim of this research paper is to develop a rule based knowledge database for legal expert system
for consumer protection act, a domain within the Indian legal system which is often in demand. A
knowledge database developed here will further help the legal expert system to determine type of case with
respect to the Indian Judicial System. In this paper a rule based knowledge database will be developed to
determine the type of the case. The main aim of the study is to build a prototype which will rule based in
nature. The rule based knowledge database development is the first phase in the development of
comprehensive rule based legal expert system for consumer protection act which will be of great help in
process of solving consumer related cases.
Nowadays the technological progress allows us to have highly flexible solutions, easily accessible with
lower levels of investment, which leads to many companies adopting SaaS (Software-as-a-Service) to
support their business processes. Associated with this movement and considering the advantages of SaaS, it
is important to understand whether work is being developed that is underutilized because companies are
not taking advantage of it, and in this case it is necessary to understand the reasons thereof. This
knowledge is important even for people who do not use or do not develop/provide SaaS, since sooner or
later it will be unavoidable due to current trends. In the near future, nearly all decision-makers of IT
strategies will be forced to consider adopting SaaS as an IT solution for the convenience benefits
associated with technology or market competition. At that time they will have to know how to evaluate
impacts and decide. What are the real needs in the Portuguese market? What fears and what is being done
to mitigate them? What are the implications of the adoption of SaaS? Where should we focus attention on
SaaS offerings in order to create greater value? These are questions we must answer to actually be able to
assess and decide. Often, decision-makers of business strategies consider only the attractive incentives of
using SaaS ignoring the impacts associated with new technologies. The need for tools and processes to
assess these impacts before adopting a SaaS solution is crucial to ensure the sustainability of the
information system, reduce uncertainty and facilitate decision making. This article presents a framework
for evaluating impacts of SaaS called SIE (SaaS Impact Evaluation) which in addition to guidance for the
present research, aims to provide guidelines for the collection, data analysis, impact assessment and
decision making about including SaaS on the organizations strategic plans.
Comparative performance analysis of two anaphora resolution systemsijfcstjournal
Anaphora Resolution is the process of finding referents in the given discourse. Anaphora Resolution is one
of the complex tasks of linguistics. This paper presents the performance analysis of two computational
models that uses Gazetteer method for resolving anaphora in Hindi Language. In Gazetteer method
different classes (Gazettes) of elements are created. These Gazettes are used to provide external knowledge
to the system. The two models use Recency and Animistic factor for resolving anaphors. For Recency factor
first model uses the concept of centering approach and second uses the concept of Lappin Leass approach.
Gazetteers are used to provide Animistic knowledge. This paper presents the experimental results of both
the models. These experiments are conducted on short Hindi stories, news articles and biography content
from Wikipedia. The respective accuracy for both the model is analyzed and finally the conclusion is drawn
for the best suitable model for Hindi Language.
Mining of product reviews at aspect levelijfcstjournal
Today’s world is a world of Internet, almost all work can be done with the help of it, from simple mobile
phone recharge to biggest business deals can be done with the help of this technology. People spent their
most of the times on surfing on the Web; it becomes a new source of entertainment, education,
communication, shopping etc. Users not only use these websites but also give their feedback and
suggestions that will be useful for other users. In this way a large amount of reviews of users are collected
on the Web that needs to be explored, analyse and organized for better decision making. Opinion Mining or
Sentiment Analysis is a Natural Language Processing and Information Extraction task that identifies the
user’s views or opinions explained in the form of positive, negative or neutral comments and quotes
underlying the text. Aspect based opinion mining is one of the level of Opinion mining that determines the
aspect of the given reviews and classify the review for each feature. In this paper an aspect based opinion
mining system is proposed to classify the reviews as positive, negative and neutral for each feature.
Negation is also handled in the proposed system. Experimental results using reviews of products show the
effectiveness of the system.
α Nearness ant colony system with adaptive strategies for the traveling sales...ijfcstjournal
On account of ant colony algorithm easy to fall into local optimum, this paper presents an improved ant
colony optimization called α-AACS and reports its performance. At first, we provide an concise description
of the original ant colony system(ACS) and introduce α-nearness based on the minimum 1-tree for ACS’s
disadvantage, which better reflects the chances of a given link being a member of an optimal tour. Then, we
improve α-nearness by computing a lower bound and propose other adaptations for ACS. Finally, we
conduct a fair competition between our algorithm and others. The results clearly show that α-AACS has a
better global searching ability in finding the best solutions, which indicates that α-AACS is an effective
approach for solving the traveling salesman problem.
Web personalization using clustering of web usage dataijfcstjournal
The exponential growth in the number and the complexity of information resources and services on the Web
has made log data an indispensable resource to characterize the users for Web-based environment. It
creates information of related web data in the form of hierarchy structure through approximation. This
hierarchy structure can be used as the input for a variety of data mining tasks such as clustering,
association rule mining, sequence mining etc.
In this paper, we present an approach for personalizing web user environment dynamically when he
interacting with web by clustering of web usage data using concept hierarchy. The system is inferred from
the web server’s access logs by means of data and web usage mining techniques to extract the information
about users. The extracted knowledge is used for the purpose of offering a personalized view of the
services to users.
GRAPH COLOURING PROBLEM BASED ON DISCRETE IMPERIALIST COMPETITIVE ALGORITHMijfcstjournal
In graph theory, Graph Colouring Problem (GCP) is an assignment of colours to vertices of any given graph such that the colours on adjacent vertices are different. The GCP is known to be an optimization and NP-hard problem.Imperialist Competitive Algorithm (ICA) is a meta-heuristic optimization and stochastic search strategy which is inspired from socio-political phenomenon of imperialistic competition. The ICA contains two main operators: the assimilation and the imperialistic competition. The ICA has excellent capabilities such as high convergence rate and better global optimum achievement. In this research, a discrete version of ICA is proposed to deal with the solution of GCP. We call this algorithm as the DICA.The performance of the proposed method is compared with Genetic Algorithm (GA) on seven well-known graph colouring benchmarks. Experimental results demonstrate the superiority of the DICA for the benchmarks. This means DICA can produce optimal and valid solutions for different GCP instances.
SHORT LISTING LIKELY IMAGES USING PROPOSED MODIFIED-SIFT TOGETHER WITH CONVEN...ijfcstjournal
The paper proposes the modified-SIFT algorithm which will be a modified form of the scale invariant feature transform. The modification consists of considering successive groups of 8 rows of pixel, along the height of the image. These are used to construct 8 bin histograms for magnitude as well as orientation individually. As a result the number of feature descriptors is significantly less (95%) than the standard SIFT approach. Fewer feature descriptor leads to reduced accuracy. This reduction in accuracy is quite drastic when searching for a single (RANK1) image match; however accuracy improves if a band of likely (say tolerance of 10%) images is to be returned. The paper therefore proposes a two-stage-approach where
First Modified-SIFT is used to obtain a shortlisted band of likely images subsequently SIFT is applied within this band to find a perfect match. It may appear that this process is tedious however it provides a significant reduction in search time as compared to applying SIFT on the entire database. The minor reduction in accuracy can be offset by the considerable time gained while searching a large database. The
modified-SIFT algorithm when used in conjunction with a face cropping algorithm can also be used to find a match against disguised images.
Formal method techniques provides a suitable platform for the software development in software systems.
Formal methods and formal verification is necessary to prove the correctness and improve performance of
software systems in various levels of design and implementation, too. Security Discussion is an important
issue in computer systems. Since the antivirus applications have very important role in computer systems
security, verifying these applications is very essential and necessary. In this paper, we present four new
approaches for antivirus system behavior and a behavioral model of protection services in the antivirus
system is proposed. We divided the behavioral model in to preventive behavior and control behavior and
then we formal these behaviors. Finally by using some definitions we explain the way these behaviors are
mapped on each other by using our new approaches.
There are several kinds of errors for which error detecting and error correcting codes have been
constructed. Solid burst errors are common in many communications. In general communication due to the
long messages, the strings of solid bursts of small length may repeat in a vector itself. The concept of
repeated bursts is introduced by Beraradi, Dass and Verma[3] which has opened a new area of study. They
defined 2-repeated bursts and obtained results for detection and correction of such type of errors.
This paper considers a new similar kind of error which will be termed as '2-repeated solid burst error of
length b'. Lower and upper bounds on the number of parity checks required for the existence of codes that
detect 2-repeated solid burst error of length b or less are obtained. This is followed by an example of such
a code. Further, codes capable of detecting and simultaneously correcting such errors have also been dealt
with.
Analysis of software cost estimation usingijfcstjournal
The growing application of software and resource constraints in software projects development need a
more accurate estimate of the cost and effort because of the importance in program planning, coordinated
scheduling and resource management including the number of programming's and software design using
tools and modern methods of modeling. Effectively control of investment for software development is
achieved by accurate cost estimation.The accurate Software Cost Estimation (SCE) is very difficult in the
early stages of software development because many of input parameters that are effective in software's
effort are very vague and uncertain in the early stages. SCE that is the basis of software projects
development planning is considered to be of high accuracy, because if the estimate is less than actual
values, confidence factor is reduce and this is means the possibility of failure in project. Conversely, if the
project is estimated at more than the actual value it would be the concept of unhelpful investment and
waste of resources. In the evaluation of software projects is commonly used deterministic method. But
software world is totally different from the linear variables and nowadays for performance and estimation
should be used nonlinear and non-probabilistic methods. In this paper, we have studied the SCE Using
Fuzzy Logic (FL) and we have compared it with COCOMO model. Results of investigations show that FL is
a performance model for SCE.
Applications of artificial immune system a reviewijfcstjournal
The Biological Immune System is a remarkable information processing and self-learning system that offers
stimulation to build Artificial Immune System (AIS).During the last two decades, the field of AIS is
progressing slowly and steadily as a branch of Computational Intelligence (CI). At present the AIS
algorithms such as Negative Selection Theory, Clonal Selection Theory, Immune Networks Theory, Danger
theory and Dendritic Cell Algorithm are widely used to solve many real world problems in a vast range of
domain areas such as Network Intrusion Detection (NID), Anomaly Detection, Clustering and
classification and Pattern recognition. This review paper critically discusses the theoretical foundation,
research methodologies and applications of the AIS.
Path finding algorithm addresses problem of finding shortest path from source to destination avoiding
obstacles. There exist various search algorithms namely A*, Dijkstra's and ant colony optimization. Unlike
most path finding algorithms which require destination co-ordinates to compute path, the proposed
algorithm comprises of a new method which finds path using backtracking without requiring destination
co-ordinates. Moreover, in existing path finding algorithm, the number of iterations required to find path is
large. Hence, to overcome this, an algorithm is proposed which reduces number of iterations required to
traverse the path. The proposed algorithm is hybrid of backtracking and a new technique(modified 8-
neighbor approach). The proposed algorithm can become essential part in location based, network, gaming
applications. grid traversal, navigation, gaming applications, mobile robot and Artificial Intelligence.
Comparison of Cost Estimation Methods using Hybrid Artificial Intelligence on...IJERA Editor
Cost estimating at schematic design stage as the basis of project evaluation, engineering design, and cost
management, plays an important role in project decision under a limited definition of scope and constraints in
available information and time, and the presence of uncertainties. The purpose of this study is to compare the
performance of cost estimation models of two different hybrid artificial intelligence approaches: regression
analysis-adaptive neuro fuzzy inference system (RANFIS) and case based reasoning-genetic algorithm (CBRGA)
techniques. The models were developed based on the same 50 low-cost apartment project datasets in
Indonesia. Tested on another five testing data, the models were proven to perform very well in term of accuracy.
A CBR-GA model was found to be the best performer but suffered from disadvantage of needing 15 cost drivers
if compared to only 4 cost drivers required by RANFIS for on-par performance.
EMPIRICAL APPLICATION OF SIMULATED ANNEALING USING OBJECT-ORIENTED METRICS TO...ijcsa
The work is about using Simulated Annealing Algorithm for the effort estimation model parameter
optimization which can lead to the reduction in the difference in actual and estimated effort used in model
development.
The model has been tested using OOP’s dataset, obtained from NASA for research purpose.The data set
based model equation parameters have been found that consists of two independent variables, viz. Lines of
Code (LOC) along with one more attribute as a dependent variable related to software development effort
(DE). The results have been compared with the earlier work done by the author on Artificial Neural
Network (ANN) and Adaptive Neuro Fuzzy Inference System (ANFIS) and it has been observed that the
developed SA based model is more capable to provide better estimation of software development effort than
ANN and ANFIS
A Defect Prediction Model for Software Product based on ANFISIJSRD
Artificial intelligence techniques are day by day getting involvement in all the classification and prediction based process like environmental monitoring, stock exchange conditions, biomedical diagnosis, software engineering etc. However still there are yet to be simplify the challenges of selecting training criteria for design of artificial intelligence models used for prediction of results. This work focus on the defect prediction mechanism development using software metric data of KC1.We have taken subtractive clustering approach for generation of fuzzy inference system (FIS).The FIS rules are generated at different radius of influence of input attribute vectors and the developed rules are further modified by ANFIS technique to obtain the prediction of number of defects in software project using fuzzy logic system.
A Defect Prediction Model for Software Product based on ANFISIJSRD
Artificial intelligence techniques are day by day getting involvement in all the classification and prediction based process like environmental monitoring, stock exchange conditions, biomedical diagnosis, software engineering etc. However still there are yet to be simplify the challenges of selecting training criteria for design of artificial intelligence models used for prediction of results. This work focus on the defect prediction mechanism development using software metric data of KC1.We have taken subtractive clustering approach for generation of fuzzy inference system (FIS).The FIS rules are generated at different radius of influence of input attribute vectors and the developed rules are further modified by ANFIS technique to obtain the prediction of number of defects in software project using fuzzy logic system.
Artificial intelligence based pattern recognition is
one of the most important tools in process control to identify
process problems. The objective of this study was to
evaluate the relative performance of a feature-based
Recognizer compared with the raw data-based recognizer.
The study focused on recognition of seven commonly
researched patterns plotted on the quality chart. The
artificial intelligence based pattern recognizer trained using
the three selected statistical features resulted in significantly
better performance compared with the raw data-based
recognizer.
Using Data Mining to Identify COSMIC Function Point Measurement Competence IJECEIAES
Cosmic Function Point (CFP) measurement errors leads budget, schedule and quality problems in software projects. Therefore, it’s important to identify and plan requirements engineers’ CFP training need quickly and correctly. The purpose of this paper is to identify software requirements engineers’ COSMIC Function Point measurement competence development need by using machine learning algorithms and requirements artifacts created by engineers. Used artifacts have been provided by a large service and technology company ecosystem in Telco. First, feature set has been extracted from the requirements model at hand. To do the data preparation for educational data mining, requirements and COSMIC Function Point (CFP) audit documents have been converted into CFP data set based on the designed feature set. This data set has been used to train and test the machine learning models by designing two different experiment settings to reach statistically significant results. Ten different machine learning algorithms have been used. Finally, algorithm performances have been compared with a baseline and each other to find the best performing models on this data set. In conclusion, REPTree, OneR, and Support Vector Machines (SVM) with Sequential Minimal Optimization (SMO) algorithms achieved top performance in forecasting requirements engineers’ CFP training need.
ENSEMBLE REGRESSION MODELS FOR SOFTWARE DEVELOPMENT EFFORT ESTIMATION: A COMP...ijseajournal
As demand for computer software continually increases, software scope and complexity become higher than ever. The software industry is in real need of accurate estimates of the project under development. Software development effort estimation is one of the main processes in software project management. However, overestimation and underestimation may cause the software industry loses. This study determines which technique has better effort prediction accuracy and propose combined techniques that could provide better estimates. Eight different ensemble models to estimate effort with Ensemble Models were compared with each other base on the predictive accuracy on the Mean Absolute Residual (MAR) criterion and statistical tests. The results have indicated that the proposed ensemble models, besides delivering high efficiency in contrast to its counterparts, and produces the best responses for software project effort estimation. Therefore, the proposed ensemble models in this study will help the project managers working with development quality software.
Function Point Software Cost Estimates using Neuro-Fuzzy techniqueijceronline
Software estimation accuracy is among the greatest challenges for software developers. As Neurofuzzy based system is able to approximate the non-linear function with more precision so it is used as a soft computing approach to generate model by formulating the relationship based on its training. The approach presented in this paper is independent of the nature and type of estimation. In this paper, Function point is used as algorithmic model and an attempt is being made to validate the soundness of Neuro fuzzy technique using ISBSG and NASA project data.
The role of Dataset in training ANFIS System for Course AdvisorAM Publications
Adaptive Network based Fuzzy Inference System (ANFIS) is used in the field of decision making to help
the students to choose the best course according to his/her requirements. The structure of ANFIS system and the
datasets used to train the system play a vital role in evaluating the performance of the system. This paper is based on
the design of Sugeno type ANFIS with grid partitioning and the usage of different datasets to train the system using
MATLAB. Results demonstrate that proper dataset is needed for training the ANFIS model
A DECISION SUPPORT SYSTEM FOR ESTIMATING COST OF SOFTWARE PROJECTS USING A HY...ijfcstjournal
One of the major challenges for software, nowadays, is software cost estimation. It refers to estimating the
cost of all activities including software development, design, supervision, maintenance and so on. Accurate
cost-estimation of software projects optimizes the internal and external processes, staff works, efforts and
the overheads to be coordinated with one another. In the management software projects, estimation must
be taken into account so that reduces costs, timing and possible risks to avoid project failure. In this paper,
a decision- support system using a combination of multi-layer artificial neural network and decision tree is
proposed to estimate the cost of software projects. In the model included into the proposed system,
normalizing factors, which is vital in evaluating efforts and costs estimation, is carried out using C4.5
decision tree. Moreover, testing and training factors are done by multi-layer artificial neural network and
the most optimal values are allocated to them. The experimental results and evaluations on Dataset
NASA60 show that the proposed system has less amount of the total average relative error compared with
COCOMO model.
A DECISION SUPPORT SYSTEM FOR ESTIMATING COST OF SOFTWARE PROJECTS USING A HY...ijfcstjournal
One of the major challenges for software, nowadays, is software cost estimation. It refers to estimating the
cost of all activities including software development, design, supervision, maintenance and so on. Accurate
cost-estimation of software projects optimizes the internal and external processes, staff works, efforts and
the overheads to be coordinated with one another. In the management software projects, estimation must
be taken into account so that reduces costs, timing and possible risks to avoid project failure. In this paper,
a decision- support system using a combination of multi-layer artificial neural network and decision tree is
proposed to estimate the cost of software projects. In the model included into the proposed system,
normalizing factors, which is vital in evaluating efforts and costs estimation, is carried out using C4.5
decision tree. Moreover, testing and training factors are done by multi-layer artificial neural network and
the most optimal values are allocated to them. The experimental results and evaluations on Dataset
NASA60 show that the proposed system has less amount of the total average relative error compared with
COCOMO model.
In the present paper, applicability and
capability of A.I techniques for effort estimation prediction has
been investigated. It is seen that neuro fuzzy models are very
robust, characterized by fast computation, capable of handling
the distorted data. Due to the presence of data non-linearity, it is
an efficient quantitative tool to predict effort estimation. The one
hidden layer network has been developed named as OHLANFIS
using MATLAB simulation environment.
Here the initial parameters of the OHLANFIS are
identified using the subtractive clustering method. Parameters of
the Gaussian membership function are optimally determined
using the hybrid learning algorithm. From the analysis it is seen
that the Effort Estimation prediction model developed using
OHLANFIS technique has been able to perform well over normal
ANFIS Model.
From the past many years many software defects prediction models are developed to solve the various issues in software project development. Software reliability is the significant in software quality which evaluates and predicts the quality of the software based on the defects prediction. Many software companies are trying to improve the software quality and also trying to reduce the cost of the software development. Rayleigh model is one of the significant models to analyze the software defects based on the generated data. Analysis of means (ANOM) is statistical technique which gives the quality assurance based on the situations. In this paper, an improved software defect prediction models (ISDPM) are used for predicting defects occur at the time of five phases such as analysis, planning, design, testing and maintenance. To improve the performance of the proposed methodology an order statistics is adopted for better prediction. The experiments are conducted on 2 synthetic projects that are used to analyze the defects.
ENHANCING ENGLISH WRITING SKILLS THROUGH INTERNET-PLUS TOOLS IN THE PERSPECTI...ijfcstjournal
This investigation delves into incorporating a hybridized memetic strategy within the framework of English
composition pedagogy, leveraging Internet Plus resources. The study aims to provide an in-depth analysis
of how this method influences students’ writing competence, their perceptions of writing, and their
enthusiasm for English acquisition. Employing an explanatory research design that combines qualitative
and quantitative methods, the study collects data through surveys, interviews, and observations of students’
writing performance before and after the intervention. Findings demonstrate a beneficial impact of
integrating the memetic approach alongside Internet Plus tools on the writing aptitude of English as a
Foreign Language (EFL) learners. Students reported increased engagement with writing, attributing it to
the use of Internet plus tools. They also expressed that the memetic approach facilitated a deeper
understanding of cultural and social contexts in writing. Furthermore, the findings highlight a significant
improvement in students’ writing skills following the intervention. This study provides significant insights
into the practical implementation of the memetic approach within English writing education, highlighting
the beneficial contribution of Internet Plus tools in enriching students' learning journeys.
A SURVEY TO REAL-TIME MESSAGE-ROUTING NETWORK SYSTEM WITH KLA MODELLINGijfcstjournal
Messages routing over a network is one of the most fundamental concept in communication which requires
simultaneous transmission of messages from a source to a destination. In terms of Real-Time Routing, it
refers to the addition of a timing constraint in which messages should be received within a specified time
delay. This study involves Scheduling, Algorithm Design and Graph Theory which are essential parts of
the Computer Science (CS) discipline. Our goal is to investigate an innovative and efficient way to present
these concepts in the context of CS Education. In this paper, we will explore the fundamental modelling of
routing real-time messages on networks. We study whether it is possible to have an optimal on-line
algorithm for the Arbitrary Directed Graph network topology. In addition, we will examine the message
routing’s algorithmic complexity by breaking down the complex mathematical proofs into concrete, visual
examples. Next, we explore the Unidirectional Ring topology in finding the transmission’s
“makespan”.Lastly, we propose the same network modelling through the technique of Kinesthetic Learning
Activity (KLA). We will analyse the data collected and present the results in a case study to evaluate the
effectiveness of the KLA approach compared to the traditional teaching method.
A COMPARATIVE ANALYSIS ON SOFTWARE ARCHITECTURE STYLESijfcstjournal
Software architecture is the structural solution that achieves the overall technical and operational
requirements for software developments. Software engineers applied software architectures for their
software system developments; however, they worry the basic benchmarks in order to select software
architecture styles, possible components, integration methods (connectors) and the exact application of
each style.
The objective of this research work was a comparative analysis of software architecture styles by its
weakness and benefits in order to select by the programmer during their design time. Finally, in this study,
the researcher has been identified architectural styles, weakness, and Strength and application areas with
its component, connector and Interface for the selected architectural styles.
SYSTEM ANALYSIS AND DESIGN FOR A BUSINESS DEVELOPMENT MANAGEMENT SYSTEM BASED...ijfcstjournal
A design of a sales system for professional services requires a comprehensive understanding of the
dynamics of sale cycles and how key knowledge for completing sales is managed. This research describes
a design model of a business development (sales) system for professional service firms based on the Saudi
Arabian commercial market, which takes into account the new advances in technology while preserving
unique or cultural practices that are an important part of the Saudi Arabian commercial market. The
design model has combined a number of key technologies, such as cloud computing and mobility, as an
integral part of the proposed system. An adaptive development process has also been used in implementing
the proposed design model.
AN ALGORITHM FOR SOLVING LINEAR OPTIMIZATION PROBLEMS SUBJECTED TO THE INTERS...ijfcstjournal
Frank t-norms are parametric family of continuous Archimedean t-norms whose members are also strict
functions. Very often, this family of t-norms is also called the family of fundamental t-norms because of the
role it plays in several applications. In this paper, optimization of a linear objective function with fuzzy
relational inequality constraints is investigated. The feasible region is formed as the intersection of two
inequality fuzzy systems defined by frank family of t-norms is considered as fuzzy composition. First, the
resolution of the feasible solutions set is studied where the two fuzzy inequality systems are defined with
max-Frank composition. Second, some related basic and theoretical properties are derived. Then, a
necessary and sufficient condition and three other necessary conditions are presented to conceptualize the
feasibility of the problem. Subsequently, it is shown that a lower bound is always attainable for the optimal
objective value. Also, it is proved that the optimal solution of the problem is always resulted from the
unique maximum solution and a minimal solution of the feasible region. Finally, an algorithm is presented
to solve the problem and an example is described to illustrate the algorithm. Additionally, a method is
proposed to generate random feasible max-Frank fuzzy relational inequalities. By this method, we can
easily generate a feasible test problem and employ our algorithm to it.
LBRP: A RESILIENT ENERGY HARVESTING NOISE AWARE ROUTING PROTOCOL FOR UNDER WA...ijfcstjournal
Underwater detector network is one amongst the foremost difficult and fascinating analysis arenas that
open the door of pleasing plenty of researchers during this field of study. In several under water based
sensor applications, nodes are square measured and through this the energy is affected. Thus, the mobility
of each sensor nodes are measured through the water atmosphere from the water flow for sensor based
protocol formations. Researchers have developed many routing protocols. However, those lost their charm
with the time. This can be the demand of the age to supply associate degree upon energy-efficient and
ascendable strong routing protocol for under water actuator networks. During this work, the authors tend
to propose a customary routing protocol named level primarily based routing protocol (LBRP), reaching to
offer strong, ascendable and energy economical routing. LBRP conjointly guarantees the most effective use
of total energy consumption and ensures packet transmission which redirects as an additional reliability in
compare to different routing protocols. In this work, the authors have used the level of forwarding node,
residual energy and distance from the forwarding node to the causing node as a proof in multicasting
technique comparisons. Throughout this work, the authors have got a recognition result concerning about
86.35% on the average in node multicasting performances. Simulation has been experienced each in a
wheezy and quiet atmosphere which represents the endorsement of higher performance for the planned
protocol.
STRUCTURAL DYNAMICS AND EVOLUTION OF CAPSULE ENDOSCOPY (PILL CAMERA) TECHNOLO...ijfcstjournal
This research paper examined and re-evaluates the technological innovation, theory, structural dynamics
and evolution of Pill Camera(Capsule Endoscopy) technology in redirecting the response manner of small
bowel (intestine) examination in human. The Pill Camera (Endoscopy Capsule) is made up of sealed
biocompatible material to withstand acid, enzymes and other antibody chemicals in the stomach is a
technology that helps the medical practitioners especially the general physicians and the
gastroenterologists to examine and re-examine the intestine for possible bleeding or infection. Before the
advent of the Pill camera (Endoscopy Capsule) the colonoscopy was the local method used but research
showed that some parts (bowel) of the intestine can’t be reach by mere traditional method hence the need
for Pill Camera. Countless number of deaths from stomach disease such as polyps, inflammatory bowel
(Crohn”s diseases), Cancers, Ulcer, anaemia and tumours of small intestines which ordinary would have
been detected by sophisticated technology like Pill Camera has become norm in the developing nations.
Nevertheless, not only will this paper examine and re-evaluate the Pill Camera Innovation, theory,
Structural dynamics and evolution it unravelled and aimed to create awareness for both medical
practitioners and the public.
AN OPTIMIZED HYBRID APPROACH FOR PATH FINDINGijfcstjournal
Path finding algorithm addresses problem of finding shortest path from source to destination avoiding
obstacles. There exist various search algorithms namely A*, Dijkstra's and ant colony optimization. Unlike
most path finding algorithms which require destination co-ordinates to compute path, the proposed
algorithm comprises of a new method which finds path using backtracking without requiring destination
co-ordinates. Moreover, in existing path finding algorithm, the number of iterations required to find path is
large. Hence, to overcome this, an algorithm is proposed which reduces number of iterations required to
traverse the path. The proposed algorithm is hybrid of backtracking and a new technique(modified 8-
neighbor approach). The proposed algorithm can become essential part in location based, network, gaming
applications. grid traversal, navigation, gaming applications, mobile robot and Artificial Intelligence.
EAGRO CROP MARKETING FOR FARMING COMMUNITYijfcstjournal
The Major Occupation in India is the Agriculture; the people involved in the Agriculture belong to the poor
class and category. The people of the farming community are unaware of the new techniques and Agromachines, which would direct the world to greater heights in the field of agriculture. Though the farmers
work hard, they are cheated by agents in today’s market. This serves as a opportunity to solve
all the problems that farmers face in the current world. The eAgro crop marketing will serve as a better
way for the farmers to sell their products within the country with some mediocre knowledge about using
the website. This would provide information to the farmers about current market rate of agro-products,
their sale history and profits earned in a sale. This site will also help the farmers to know about the market
information and to view agricultural schemes of the Government provided to farmers.
EDGE-TENACITY IN CYCLES AND COMPLETE GRAPHSijfcstjournal
It is well known that the tenacity is a proper measure for studying vulnerability and reliability in graphs.
Here, a modified edge-tenacity of a graph is introduced based on the classical definition of tenacity.
Properties and bounds for this measure are introduced; meanwhile edge-tenacity is calculated for cycle
graphs and also for complete graphs.
COMPARATIVE STUDY OF DIFFERENT ALGORITHMS TO SOLVE N QUEENS PROBLEMijfcstjournal
This Paper provides a brief description of the Genetic Algorithm (GA), the Simulated Annealing (SA)
Algorithm, the Backtracking (BT) Algorithm and the Brute Force (BF) Search Algorithm and attempts to
explain the way as how the Proposed Genetic Algorithm (GA), the Proposed Simulated Annealing (SA)
Algorithm using GA, the Backtracking (BT) Algorithm and the Brute Force (BF) Search Algorithm can be
employed in finding the best solution of N Queens Problem and also, makes a comparison between these
four algorithms. It is entirely a review based work. The four algorithms were written as well as
implemented. From the Results, it was found that, the Proposed Genetic Algorithm (GA) performed better
than the Proposed Simulated Annealing (SA) Algorithm using GA, the Backtracking (BT) Algorithm and
the Brute Force (BF) Search Algorithm and it also provided better fitness value (solution) than the
Proposed Simulated Annealing Algorithm (SA) using GA, the Backtracking (BT) Algorithm and the Brute
Force (BF) Search Algorithm, for different N values. Also, it was noticed that, the Proposed GA took more
time to provide result than the Proposed SA using GA.
PSTECEQL: A NOVEL EVENT QUERY LANGUAGE FOR VANET’S UNCERTAIN EVENT STREAMSijfcstjournal
In recent years, the complex event processing technology has been used to process the VANET’s temporal
and spatial event streams. However, we usually cannot get the accurate data because the device sensing
accuracy limitations of the system. We only can get the uncertain data from the complex and limited
environment of the VANET. Because the VANET’s event streams are consist of the uncertain data, so they
are also uncertain. How effective to express and process these uncertain event streams has become the core
issue for the VANET system. To solve this problem, we propose a novel complex event query language
PSTeCEQL (probabilistic spatio-temporal constraint event query language). Firstly, we give the definition
of the possible world model of VANET’s uncertain event streams. Secondly, we propose an event query
language PSTeCEQL and give the syntax and the operational semantics of the language. Finally, we
illustrate the validity of the PSTeCEQL by an example.
CLUSTBIGFIM-FREQUENT ITEMSET MINING OF BIG DATA USING PRE-PROCESSING BASED ON...ijfcstjournal
Now a day enormous amount of data is getting explored through Internet of Things (IoT) as technologies
are advancing and people uses these technologies in day to day activities, this data is termed as Big Data
having its characteristics and challenges. Frequent Itemset Mining algorithms are aimed to disclose
frequent itemsets from transactional database but as the dataset size increases, it cannot be handled by
traditional frequent itemset mining. MapReduce programming model solves the problem of large datasets
but it has large communication cost which reduces execution efficiency. This proposed new pre-processed
k-means technique applied on BigFIM algorithm. ClustBigFIM uses hybrid approach, clustering using kmeans algorithm to generate Clusters from huge datasets and Apriori and Eclat to mine frequent itemsets
from generated clusters using MapReduce programming model. Results shown that execution efficiency of
ClustBigFIM algorithm is increased by applying k-means clustering algorithm before BigFIM algorithm as
one of the pre-processing technique.
A MUTATION TESTING ANALYSIS AND REGRESSION TESTINGijfcstjournal
Software testing is a testing which conducted a test to provide information to client about the quality of the
product under test. Software testing can also provide an objective, independent view of the software to
allow the business to appreciate and understand the risks of software implementation. In this paper we
focused on two main software testing –mutation testing and mutation testing. Mutation testing is a
procedural testing method, i.e. we use the structure of the code to guide the test program, A mutation is a
little change in a program. Such changes are applied to model low level defects that obtain in the process
of coding systems. Ideally mutations should model low-level defect creation. Mutation testing is a process
of testing in which code is modified then mutated code is tested against test suites. The mutations used in
source code are planned to include in common programming errors. A good unit test typically detects the
program mutations and fails automatically. Mutation testing is used on many different platforms, including
Java, C++, C# and Ruby. Regression testing is a type of software testing that seeks to uncover
new software bugs, or regressions, in existing functional and non-functional areas of a system after
changes such as enhancements, patches or configuration changes, have been made to them. When defects
are found during testing, the defect got fixed and that part of the software started working as needed. But
there may be a case that the defects that fixed have introduced or uncovered a different defect in the
software. The way to detect these unexpected bugs and to fix them used regression testing. The main focus
of regression testing is to verify that changes in the software or program have not made any adverse side
effects and that the software still meets its need. Regression tests are done when there are any changes
made on software, because of modified functions.
GREEN WSN- OPTIMIZATION OF ENERGY USE THROUGH REDUCTION IN COMMUNICATION WORK...ijfcstjournal
Advances in micro fabrication and communication techniques have led to unimaginable proliferation of
WSN applications. Research is focussed on reduction of setup operational energy costs. Bulk of operational
energy costs are linked to communication activities of WSN. Any progress towards energy efficiency has a
potential of huge savings globally. Therefore, every energy efficient step is an endeavour to cut costs and
‘Go Green’. In this paper, we have proposed a framework to reduce communication workload through: Innetwork compression and multiple query synthesis at the base-station and modification of query syntax
through introduction of Static Variables. These approaches are general approaches which can be used in
any WSN irrespective of application.
A NEW MODEL FOR SOFTWARE COSTESTIMATION USING HARMONY SEARCHijfcstjournal
Accurate and realistic estimation is always considered to be a great challenge in software industry.
Software Cost Estimation (SCE) is the standard application used to manage software projects. Determining
the amount of estimation in the initial stages of the project depends on planning other activities of the
project. In fact, the estimation is confronted with a number of uncertainties and barriers’, yet assessing the
previous projects is essential to solve this problem. Several models have been developed for the analysis of
software projects. But the classical reference method is the COCOMO model, there are other methods
which are also applied such as Function Point (FP), Line of Code(LOC); meanwhile, the expert`s opinions
matter in this regard. In recent years, the growth and the combination of meta-heuristic algorithms with
high accuracy have brought about a great achievement in software engineering. Meta-heuristic algorithms
which can analyze data from multiple dimensions and identify the optimum solution between them are
analytical tools for the analysis of data. In this paper, we have used the Harmony Search (HS)algorithm for
SCE. The proposed model which is a collection of 60 standard projects from Dataset NASA60 has been
assessed.The experimental results show that HS algorithm is a good way for determining the weight
similarity measures factors of software effort, and reducing the error of MRE.
AGENT ENABLED MINING OF DISTRIBUTED PROTEIN DATA BANKSijfcstjournal
Mining biological data is an emergent area at the intersection between bioinformatics and data mining
(DM). The intelligent agent based model is a popular approach in constructing Distributed Data Mining
(DDM) systems to address scalable mining over large scale distributed data. The nature of associations
between different amino acids in proteins has also been a subject of great anxiety. There is a strong need to
develop new models and exploit and analyze the available distributed biological data sources. In this study,
we have designed and implemented a multi-agent system (MAS) called Agent enriched Quantitative
Association Rules Mining for Amino Acids in distributed Protein Data Banks (AeQARM-AAPDB). Such
globally strong association rules enhance understanding of protein composition and are desirable for
synthesis of artificial proteins. A real protein data bank is used to validate the system.
International Journal on Foundations of Computer Science & Technology (IJFCST)ijfcstjournal
International Journal on Foundations of Computer Science & Technology (IJFCST) is a Bi-monthly peer-reviewed and refereed open access journal that publishes articles which contribute new results in all areas of the Foundations of Computer Science & Technology. Over the last decade, there has been an explosion in the field of computer science to solve various problems from mathematics to engineering. This journal aims to provide a platform for exchanging ideas in new emerging trends that needs more focus and exposure and will attempt to publish proposals that strengthen our goals. Topics of interest include, but are not limited to the following:
Because the technology is used largely in the last decades; cybercrimes have become a significant
international issue as a result of the huge damage that it causes to the business and even to the ordinary
users of technology. The main aims of this paper is to shed light on digital crimes and gives overview about
what a person who is related to computer science has to know about this new type of crimes. The paper has
three sections: Introduction to Digital Crime which gives fundamental information about digital crimes,
Digital Crime Investigation which presents different investigation models and the third section is about
Cybercrime Law.
DISTRIBUTION OF MAXIMAL CLIQUE SIZE UNDER THE WATTS-STROGATZ MODEL OF EVOLUTI...ijfcstjournal
In this paper, we analyze the evolution of a small-world network and its subsequent transformation to a
random network using the idea of link rewiring under the well-known Watts-Strogatz model for complex
networks. Every link u-v in the regular network is considered for rewiring with a certain probability and if
chosen for rewiring, the link u-v is removed from the network and the node u is connected to a randomly
chosen node w (other than nodes u and v). Our objective in this paper is to analyze the distribution of the
maximal clique size per node by varying the probability of link rewiring and the degree per node (number
of links incident on a node) in the initial regular network. For a given probability of rewiring and initial
number of links per node, we observe the distribution of the maximal clique per node to follow a Poisson
distribution. We also observe the maximal clique size per node in the small-world network to be very close
to that of the average value and close to that of the maximal clique size in a regular network. There is no
appreciable decrease in the maximal clique size per node when the network transforms from a regular
network to a small-world network. On the other hand, when the network transforms from a small-world
network to a random network, the average maximal clique size value decreases significantly
Student information management system project report ii.pdfKamal Acharya
Our project explains about the student management. This project mainly explains the various actions related to student details. This project shows some ease in adding, editing and deleting the student details. It also provides a less time consuming process for viewing, adding, editing and deleting the marks of the students.
Hierarchical Digital Twin of a Naval Power SystemKerry Sado
A hierarchical digital twin of a Naval DC power system has been developed and experimentally verified. Similar to other state-of-the-art digital twins, this technology creates a digital replica of the physical system executed in real-time or faster, which can modify hardware controls. However, its advantage stems from distributing computational efforts by utilizing a hierarchical structure composed of lower-level digital twin blocks and a higher-level system digital twin. Each digital twin block is associated with a physical subsystem of the hardware and communicates with a singular system digital twin, which creates a system-level response. By extracting information from each level of the hierarchy, power system controls of the hardware were reconfigured autonomously. This hierarchical digital twin development offers several advantages over other digital twins, particularly in the field of naval power systems. The hierarchical structure allows for greater computational efficiency and scalability while the ability to autonomously reconfigure hardware controls offers increased flexibility and responsiveness. The hierarchical decomposition and models utilized were well aligned with the physical twin, as indicated by the maximum deviations between the developed digital twin hierarchy and the hardware.
Cosmetic shop management system project report.pdfKamal Acharya
Buying new cosmetic products is difficult. It can even be scary for those who have sensitive skin and are prone to skin trouble. The information needed to alleviate this problem is on the back of each product, but it's thought to interpret those ingredient lists unless you have a background in chemistry.
Instead of buying and hoping for the best, we can use data science to help us predict which products may be good fits for us. It includes various function programs to do the above mentioned tasks.
Data file handling has been effectively used in the program.
The automated cosmetic shop management system should deal with the automation of general workflow and administration process of the shop. The main processes of the system focus on customer's request where the system is able to search the most appropriate products and deliver it to the customers. It should help the employees to quickly identify the list of cosmetic product that have reached the minimum quantity and also keep a track of expired date for each cosmetic product. It should help the employees to find the rack number in which the product is placed.It is also Faster and more efficient way.
Sachpazis:Terzaghi Bearing Capacity Estimation in simple terms with Calculati...Dr.Costas Sachpazis
Terzaghi's soil bearing capacity theory, developed by Karl Terzaghi, is a fundamental principle in geotechnical engineering used to determine the bearing capacity of shallow foundations. This theory provides a method to calculate the ultimate bearing capacity of soil, which is the maximum load per unit area that the soil can support without undergoing shear failure. The Calculation HTML Code included.
Explore the innovative world of trenchless pipe repair with our comprehensive guide, "The Benefits and Techniques of Trenchless Pipe Repair." This document delves into the modern methods of repairing underground pipes without the need for extensive excavation, highlighting the numerous advantages and the latest techniques used in the industry.
Learn about the cost savings, reduced environmental impact, and minimal disruption associated with trenchless technology. Discover detailed explanations of popular techniques such as pipe bursting, cured-in-place pipe (CIPP) lining, and directional drilling. Understand how these methods can be applied to various types of infrastructure, from residential plumbing to large-scale municipal systems.
Ideal for homeowners, contractors, engineers, and anyone interested in modern plumbing solutions, this guide provides valuable insights into why trenchless pipe repair is becoming the preferred choice for pipe rehabilitation. Stay informed about the latest advancements and best practices in the field.
CFD Simulation of By-pass Flow in a HRSG module by R&R Consult.pptxR&R Consult
CFD analysis is incredibly effective at solving mysteries and improving the performance of complex systems!
Here's a great example: At a large natural gas-fired power plant, where they use waste heat to generate steam and energy, they were puzzled that their boiler wasn't producing as much steam as expected.
R&R and Tetra Engineering Group Inc. were asked to solve the issue with reduced steam production.
An inspection had shown that a significant amount of hot flue gas was bypassing the boiler tubes, where the heat was supposed to be transferred.
R&R Consult conducted a CFD analysis, which revealed that 6.3% of the flue gas was bypassing the boiler tubes without transferring heat. The analysis also showed that the flue gas was instead being directed along the sides of the boiler and between the modules that were supposed to capture the heat. This was the cause of the reduced performance.
Based on our results, Tetra Engineering installed covering plates to reduce the bypass flow. This improved the boiler's performance and increased electricity production.
It is always satisfying when we can help solve complex challenges like this. Do your systems also need a check-up or optimization? Give us a call!
Work done in cooperation with James Malloy and David Moelling from Tetra Engineering.
More examples of our work https://www.r-r-consult.dk/en/cases-en/
Overview of the fundamental roles in Hydropower generation and the components involved in wider Electrical Engineering.
This paper presents the design and construction of hydroelectric dams from the hydrologist’s survey of the valley before construction, all aspects and involved disciplines, fluid dynamics, structural engineering, generation and mains frequency regulation to the very transmission of power through the network in the United Kingdom.
Author: Robbie Edward Sayers
Collaborators and co editors: Charlie Sims and Connor Healey.
(C) 2024 Robbie E. Sayers
Governing Equations for Fundamental Aerodynamics_Anderson2010.pdf
A hybrid fuzzy ann approach for software effort estimation
1. International Journal in Foundations of Computer Science & Technology (IJFCST), Vol.4, No.5, September 2014
A HYBRID FUZZY-ANN APPROACH FOR SOFTWARE
EFFORT ESTIMATION
Sheenu Rizvi1, Dr. S.Q. Abbas2 and Dr. Rizwan Beg3
1Department of Computer Science, Amity University, Lucknow, India
2A.I.M.T., Lucknow, India
3 Integral University, Lucknow, India
ABSTRACT
Software development effort estimation is one of the major activities in software project management.
During the project proposal stage there is high probability of estimates being made inaccurate but later on
this inaccuracy decreases. In the field of software development there are certain matrices, based on which
the effort estimation is being made. Till date various methods has been proposed for software effort
estimation, of which the non algorithmic methods, like artificial intelligence techniques have been very
successful. A Hybrid Fuzzy-ANN model, known as Adaptive Neuro Fuzzy Inference System (ANFIS) is more
suitable in such situations. The present paper is concerned with developing software effort estimation
model based on ANFIS. The present study evaluates the efficiency of the proposed ANFIS model, for which
COCOMO81 datasets has been used. The result so obtained has been compared with Artificial Neural
Network (ANN) and Intermediate COCOCMO model developed by Boehm. The results were analyzed using
Magnitude of Relative Error (MRE) and Root Mean Square Error (RMSE). It is observed that the ANFIS
provided better results than ANN and COCOMO model.
KEYWORDS
Software Effort Estimation, RMSE, ANFIS, ANN, COCOMO, MRE.
1. INTRODUCTION
One of the key challenges in software industry is the accurate estimation of the development
effort, which is particularly important for risk evaluation, resource scheduling as well as progress
monitoring. Inaccuracies in estimations lead to problematic results; for instance, overestimation
causes waste of resources, whereas underestimation results in approval of projects that will
exceed their planned budgets. For this many models has been framed so as to make it cost
effective. These models can be examined based on methodologies used: Expert-based, analogy-based
and regression-based. Expert based models depend on the expert knowledge to use past
experience on software projects. Based on a comprehensive review, expert based estimation is
one of the most frequently applied estimation strategy. Alternatively, regression-based methods
use statistical techniques such as least square regression, in the sense that a set of independent
variables explain the dependent variable with minimum error rate. Mathematical models like
Barry Boehm’s COCOMO [1] and COCOMO II [2] are widely investigated regression-based
methods. Parameters of these models are calibrated according to the projects in a company. Thus,
they have the drawback of requiring local calibration. To combat these problems a hybrid Fuzzy-
ANN model known as Adaptive Neuro Fuzzy Inference System (ANFIS) has been dealt in this
paper.
DOI:10.5121/ijfcst.2014.4505 45
2. International Journal in Foundations of Computer Science & Technology (IJFCST), Vol.4, No.5, September 2014
2. DATA USED
The data used is COCOMO 81. The data utilised for ANFIS model development as input and
output variables are given in the Table 1. Total sixteen input variables have been used which
include fifteen effort multipliers and the size measured in thousand delivered lines of code.
Development Effort (DE) has been used as the output of the model measured in man-months. The
data were collected from the analysis of sixty three (63) software projects, as published by Barry
Boehm in 1981[3] [16].
46
Table 1. Input and Output variables for ANFIS model.
Input
Variables
RELY - Required software reliability
DATA - Data base size,
CPLX - Product complexity,
TIME - Execution time,
STOR—main storage constraint,
VIRT—virtual machine volatility
TURN—computer turnaround time,
ACAP—analyst capability,
AEXP—applications experience,
PCAP—programmer capability,
VEXP—virtual machine experience,
LEXP—language experience
MODP—modern programming,
TOOL—use of software tools,
SCED—required development schedule,
SIZE — in KLOC
Output
Variable
Development Effort (DE)
Source: - COCOMO81 Dataset (PROMISE Software Engineering Repository data [16])
3. ANFIS MODEL DEVELOPMENT
3.1. Parameter Selection
ANFIS [9],[10] is a judicious integration of FIS and ANN, capable of learning, high-level
thinking and reasoning and it combines the benefits of these two techniques into a single capsule
[4]. The success for FIS is the finding of the rule base. The reason being that there are no specific
techniques for converting the knowledge of human beings into the rule base and also in order to
maximise the performance of the model and to minimize the output error, further fine tuning of
the membership functions is required. Thus when generating a FIS using ANFIS, it is important
to select proper parameters, including the number of membership functions (MFs) for each
individual antecedent variables. It is also vital to select appropriate parameters for learning and
refining process, including the initial step size (ss). In the present work the commonly used rule
extraction method applied for FIS identification and refinement is subtractive clustering. The
MATLAB Fuzzy Logic Toolbox [7] has been used for ANFIS model development.
Here the initial parameters of the ANFIS are identified using the subtractive clustering method
[5]. However, it is vital to properly define the substractive clustering parameters, of which the
clustering radius is the most important. It is determined through a trial and error approach. By
varying the clustering radius ra with varying step size, the optimal parameters are obtained by
3. International Journal in Foundations of Computer Science & Technology (IJFCST), Vol.4, No.5, September 2014
minimizing the root mean squared error based on the validation datasets. Clustering radius rb is
selected as 1.5ra. Gaussian membership functions are used for each fuzzy set in the fuzzy system.
The number of membership functions and fuzzy rules required for a particular ANFIS is
determined through the subtractive clustering algorithm. Parameters of the Gaussian membership
function are optimally determined using the hybrid learning algorithm. Each ANFIS is trained for
10 epochs.
Gaussian membership function has been used as the input membership function and linear
membership function for the output function. Here separate sets of input and output data has been
used as input arguments. In MATLAB genfis2 generates a Sugeno-type FIS structure using
subtractive clustering. Genfis2 is generally used where there is only one output; hence here it has
been used to generate initial FIS for training the ANFIS. On the other hand genfis2 achieves this
by extracting a set of rules that simulates the data values. In order to determine the number of
rules and antecedent membership functions, subclust function has been used by the rule extraction
methods. Further it uses the linear least squares estimation to determine each rule's consequent
equations.
The parameters used in the model for training ANFIS are given in Table 2 and the rule extraction
method used is given in Table 3. Table 4 summarizes the results of types and values of model
parameters used for training ANFIS
47
Table 2. Parameters used in all the models for training ANFIS
Rule extraction method
used
Subtractive clustering
Input MF type Gaussian membership (‘gaussmf’)
Input partitioning variable
Output MF Type Linear
Number of output MFs one
Training algorithm Hybrid learning
Training epoch number 10
Initial step size 0.01
Table 3. Rule extraction method used for training ANFIS
Rule Extraction Method Type
And method ‘prod’
Or method ‘probor’
Defuzzy method ‘wtever’
Implication method ‘prod’
Aggregation method ‘max’
Table 4. Values of parameters used for training ANFIS
No. of nodes 1311
No. of linear parameters 646
No. of non-linear parameters 1216
Total no. of parameters 1862
No. of training data pairs 40
No. of testing data pairs 23
No. of fuzzy rules 38
4. International Journal in Foundations of Computer Science & Technology (IJFCST), Vol.4, No.5, September 2014
4. RESULT AND DISCUSSION
Here the ANFIS model has been trained tested by ANFIS method and their performance for the
best prediction model are evaluated and compared for training and testing data sets separately.
The RMSE performances of the ANFIS model both for training and testing datasets have been
plotted separately in Fig. 1 & Fig.2 and their corresponding range of values (minimum and
maximum) are summarized in Table 5.
48
Figure 1. Graphical plot of RMSE value variation during training
Figure 2. Graphical plot of RMSE value variation during testing
Table 5. Range of RMSE during training and testing phase
RMSE Value
Minimum Maximum
Training datasets 0.4824 2.8096
Testing datasets 186.41 188.41
5. International Journal in Foundations of Computer Science & Technology (IJFCST), Vol.4, No.5, September 2014
49
Further Table 6 gives the RMSE values using COCOMO, ANN and ANFIS techniques.
Table 6. Performance evaluation using RMSE criteria
RMSE
COCOMO ANN ANFIS
Val.
532.2147 353.1977 112.638
From analysis of Fig. 1 & Fig. 2 and perusal of the data given in tables 5 it is inferred that during
training phase (Fig.1), there is zig zag variation in the RMSE values, having a minimum value of
0.4824 (at epoch 8) and a maximum value of 2.8096 ( epoch 3). Hence during training phase
there is initially a rise in the RMSE value and then there is a fall at epoch no. 8, after which there
is again a slight increase. On the other hand, during testing phase (Fig.2) of ANFIS training
initially upto epoch 4 the RMSE value decreases and reaches upto a minimum of 186.41 and then
there is steep rise in the RMSE value upto 10 epochs, where the maximum value reached is
188.41. From Table 5 it can be inferred that ANFIS has performed better during training phase
than testing phase but its overall RMSE value is 112.638. Which shows a marked improvement
than those calculated in ANN and COCOMO model i.e. 353.1977 and 532.2147 respectively.
(Given above in Table 6).
Further consider the absolute values of Magnitude of Relative Error (MRE) calculated both for
COCOMO and ANFIS models (given below in Table 7) and their comparative plot, both for
training and testing datasets (as given in Fig. 3 & 4). From the perusal of both the data and the
graphical plot, it is seen that during the training as well as testing phase of the ANFIS model
development, the absolute values of the MRE are very less as compared to COCOMO model,
especially during training phase. Since Absolute MRE computes the absolute percentage of error
between the actual and predicted effort for each project, hence from the above data analysis it can
be derived that the absolute percentage of error between the actual and predicted effort using
ANFIS technique is far less than those using COCOMO model.
Thus, it is clear that proper selection of influential radius which affects the cluster results directly
in ANFIS using subtractive clustering rule extraction method has resulted in reduction of RMSE
and MRE both for training and testing data sets. Hence, it is seen that for small size training data,
ANFIS has outperformed ANN and COCOMO model.
Table 7. Comparative chart of Absolute values of MRE for COCOMO and ANFIS Model
S.No. ABS MRE
COCOMO
ABS MRE
ANFIS
1. 8.651813725 0.000103189
2. 73.9110625 0.030832219
3. 1.377489712 0.00195532
4. 2.00825 0.000158388
5. 16.93939394 0.000202853
6. 40.51162791 1.22696E-05
7. 22.125 0.000142747
8. 41.41395349 1.94362E-05
9. 21.04728132 1.11052E-05
10. 14.17757009 5.40767E-05
11. 42.22018349 0.000783969
12. 0.646766169 9.3241E-05
13. 43.78481013 0.000854332 14. 16.41666667 6.95013E-07
7. International Journal in Foundations of Computer Science & Technology (IJFCST), Vol.4, No.5, September 2014
51
Absolute MRE of COCOMO and ANFIS Output for training data
200
100
0
1 4 7 10 13 16 19 22 25 28 31 34 37 40
No. of Projects
A b s o l u t e
M R E
COCOMO MRE
ANFIS MRE
Figure 3. Absolute MRE plot for COCOMO and ANFIS Output for training datasets
MRE of COCOMO and ANFIS output for testing data
200
150
100
50
0
1 3 5 7 9 11 13 15 17 19 21 23
No. of Projects
A b s o l u t e M R E
MRE COCOMO
MRE ANFIS
Figure 4. Absolute MRE plot for COCOMO and ANFIS Output for testing datasets
In order to depict how well ANFIS has performed over ANN and COCOMO model, a
comparative plot of actual effort versus predicted effort, by COCOMO, ANN and ANFIS
technique, has been shown in Fig. 5 using data given in Table 8.. From the graph it is seen that
ANFIS model line almost closely follows the actual effort line than those of COCOMO. This
again depicts the superiority of ANFIS technique over ANN and COCOMO model for effort
estimation.
Table 8. Comparative chart of Actual Effort Versus Estimated Effort using COCOMO, ANN and ANFIS
S. No Actual
Effort
Estimated Effort using
COCOMO ANN ANFIS
1 2040 1863.503 2040.022 2040.002
2 1600 2782.577 3168.456 1599.507
3 243 246.3473 242.8827 242.9952
9. International Journal in Foundations of Computer Science & Technology (IJFCST), Vol.4, No.5, September 2014
53
42 45 109.29 234.8325 195.2396
43 83 103.73 101.074 228.257
44 87 132.87 100.6351 130.0721
45 106 109.2 157.2179 3.31
46 126 213.91 122.6887 343.28
47 36 32.77 7.266029 57.82236
48 1272 2204.63 6.364794 738.6743
49 156 141.51 155.7227 335.0579
50 176 162.46 491.2995 188.5651
51 122 82.74 254.6255 93.75488
52 41 36.46 48.05263 51.03936
53 14 22.41 38.53126 104.7524
54 20 11.78 6.371402 34.6563
55 18 7.51 8.634863 16.71238
56 958 388.88 957.3443 385.3861
57 237 277.35 238.0535 177.1851
58 130 145.19 1540.691 282.375
59 70 82.78 6.243794 85.83885
60 57 50.11 132.3261 119.6359
61 50 47.26 6.030985 40.99599
62 38 41.18 38.24981 140.7745
63 15 17.13 6.164915 19.69363
Finally, Figure 6, 7 & 8 shows the scatter plot of Actual Effort versus Estimated Effort using
ANFIS, ANN and COCOMO models. The figures show that the model performance is generally
precise in case of ANFIS, where all data points follow a linear trend line and the model using
ANFIS is better than ANN and COCOMO.
15000
10000
5000
0
Actual Effort
Estimated Effort using COCOMO
Estimated Effort using ANN
Estimated Effort using ANFIS
1 7 13 19 25 31 37 43 49 55 61
Figure 5. Comparative plot of Actual Effort, COCOMO, ANN and ANFIS Output
10. International Journal in Foundations of Computer Science & Technology (IJFCST), Vol.4, No.5, September 2014
54
Using ANFIS
15000
10000
5000
0
0 5000 10000 15000
Actual Effort
Estimated Effort
Figure 6. Scatter Plot of Actual vs. Estimated Effort using ANFIS
Using ANN
15000
10000
5000
0
0 2000 4000 6000 8000 1000
0
1200
0
Actual Effort
Estimated Effort
Figure 7. Scatter Plot of Actual vs. Estimated Effort using ANN
11. International Journal in Foundations of Computer Science & Technology (IJFCST), Vol.4, No.5, September 2014
55
Using COCOMO
15000
10000
5000
0
0 5000 10000 15000
Actual Effort
Estimated Effort
Figure 8. Scatter Plot of Actual vs. Estimated Effort using COCOMO
5. CONCLUSION
Here, in the present paper, applicability and capability of ANFIS techniques for effort estimation
prediction has been investigated. It is seen that ANFIS models are very robust, characterized by
fast computation, capable of handling the noisy and approximate data that are typical of data used
here for the present study. Due to the presence of non-linearity in the data, it is an efficient
quantitative tool to predict effort estimation. The studies have been carried out using MATLAB
simulation environment. In all sixteen input variable were used, consisting of fifteen Effort
Adjustment Factors and size of the project and one output variable as Effort.
Here the initial parameters of the ANFIS are identified using the subtractive clustering method.
Gaussian membership functions (given in earlier section) are used for each fuzzy set in the fuzzy
system. Subtractive clustering algorithm has been used to determine the number of membership
functions and fuzzy rules required for ANFIS development. Here hybrid learning algorithm has
been used to determine the parameters of the Gaussian membership function. Each ANFIS has
been trained for 10 epochs.
From the analysis of the above results, given under heading Results and Discussions, it is seen
that the Effort Estimation prediction model developed using ANFIS technique has been able to
perform well over ANN and COCOMO Model. This can be concluded from the analysis of the
results given in Tables 5, 6, 7 and 8. The RMSE value obtained from ANFIS model (112.638) is
lower than those from ANN (353.1977) and COCOMO Model (532.2147). Further from Fig. 6, 7
& 8 and Table 8 it is seen that ANFIS model line almost closely follows the actual effort line than
those of ANN and COCOMO. This again depicts the superiority of ANFIS technique over ANN
and COCOMO model for effort estimation.
REFERENCES
[1]. Alpaydın,E. 2004. Introduction to machine learning. Cambridge: MIT Press.
[2]. Boehm,B., Abts, C., Chulani, S. 2000. Software development cost estimation approaches: A survey.
[3]. Annals of Software Engineering (10): 177–205.
12. International Journal in Foundations of Computer Science & Technology (IJFCST), Vol.4, No.5, September 2014
[4]. Boehm,B.W. 1981. Software Engineering Economics. Upper Saddle River, NJ, USA: Prentice Hall
56
PTR.
[5]. Chen,D.W. And Zhang, J.P., (2005), “Time series prediction based on ensemble ANFIS”,
Proceedings of the fourth International Conference on Machine Learning and Cybernetics, IEEE, pp
3552-3556.10
[6]. Chiu,S.,(1994), “Fuzzy Model Identification based on cluster estimation”, Journal of Intelligent and
Fuzzy Systems, 2 (3), pp 267–278.11
[7] .Fuller,R.,(1995), “Neural Fuzzy Systems”, ISBN 951-650-624-0, ISSN 0358-5654.17
[8]. “Fuzzy Logic Toolbox”, MATLAB version R2013a.
[9]. Hammouda, K. A., “Comparative Study of Data Clustering Techniques”.
[10]. Jang,J-S.R.,(1992),“Neuro-Fuzzy Modelling: Architecture, Analyses and Applications”, P.hd. Thesis.
[11]. Jang,J-S.R.,(1993),“ANFIS-Adaptive-Network Based Fuzzy Inference System”, IEEE Transactions
on Systems, Man and Cybernetics, 23(3), pp 665-685.
[12]. Jang, J-S. R., SUN, C.-T., (1995), “Neuro-fuzzy modelling and control”, Proceedings IEEE,. 83 (3),
pp 378–406.
[13]. Jantzen,J.,(1998), “Neurofuzzy Modelling. Technical Report no. 98-H-874(nfmod)”, Department of
Automation. Technical University of Denmark.1-28.
[14]. Pendharkar, Parag C., et. al., (2005), “A Probabilistic Model for Predicting Software Development
Effort”, IEEE Transactions On Software Engineering, Vol. 31, NO. 7.
[15]. Priyono, A. Ridwan, M., et. al. (2005), “Generation of fuzzy rules with subtractive clustering”,
Journal Teknologi., 43(D), pp 143-153.
[16]. Sayyad Shirabad, J. and Menzies, T.J. (2005) The PROMISE Repository of Software Engineering
Databases. School of Information Technology and Engineering, University of Ottawa, Canada.
Available: http://promise.site.uottawa.ca/SERepository
[17]. Tagaki, T. And Sugeno, M. , (1983), “Derivation of fuzzy control rules from human operators control
actions”, Proc. IFAC Symp. Fuzzy Inform, Knowledge Representation and Decision Analysis, pp 55-
60.
[18]. Vaidehi, V., Monica, S., Mohammad Sheikh Safeer, S.,Deepika, M. And Sangeetha, S., (2008), “A
Prediction System Based on Fuzzy Logic”, Proceedings of World Congress on Engineering and
Computer Science. 38
[19]. Zadeh, L.A., 1965), “Fuzzy sets, Information and Control”, 8, pp 338–353.36.
Authors
Sheenu Rizvi, Assistant Professor, Amity School of Engineering and Technology
Lucknow, India. He received his M.Tech degree in Information Technology in 2005 and
Persuing Ph.D in Computer Application from Integral University.
Syed Qamar Abbas completed his Master of Science (MS) from BITS Pilani. His PhD
was on computer-oriented study on Queueing models. He has more than 20 years of
teaching and research experience in the field of Computer Science and Information
Technology. Currently, he is Director of Ambalika Institute of Management and
Technology, Lucknow.
Prof. Dr. M. Rizwan Beg is M.Tech & Ph.D in Computer Sc. & Engg. Presently he is
working as Controller of Examination in Integral University Luck now, Uttar Pradesh,
India He is having more than 16 years of experience which includes around 14 years of
teaching experience. His area of expertise is Software Engg., Requirement Engineering,
Software Quality, and Software Project Management. He has published more than 40
Research papers in International Journals & Conferences. Presently 8 research scholars
are pursuing their Ph.D in his supervision.