The aim of this article is to present perdition and risk accuracy analysis of default customer in the banking sector. The neural network is a learning model inspired by biological neuron it is used to estimate and predict that can depend on a large number of inputs. The bank customer dataset from UCI repository, used for data analysis method to extract informative data set from a large volume of the dataset. This dataset is used in the neural network for training data and testing data. In a training of data, the data set is iterated till the desired output. This training data is cross check with test data. This paper focuses on predicting default customer by using deep learning neural network (DNN) algorithm.
Extended pso algorithm for improvement problems k means clustering algorithmIJMIT JOURNAL
The clustering is a without monitoring process and one of the most common data mining techniques. The
purpose of clustering is grouping similar data together in a group, so were most similar to each other in a
cluster and the difference with most other instances in the cluster are. In this paper we focus on clustering
partition k-means, due to ease of implementation and high-speed performance of large data sets, After 30
year it is still very popular among the developed clustering algorithm and then for improvement problem of
placing of k-means algorithm in local optimal, we pose extended PSO algorithm, that its name is ECPSO.
Our new algorithm is able to be cause of exit from local optimal and with high percent produce the
problem’s optimal answer. The probe of results show that mooted algorithm have better performance
regards as other clustering algorithms specially in two index, the carefulness of clustering and the quality
of clustering.
Identification of important features and data mining classification technique...IJECEIAES
Employees absenteeism at the work costs organizations billions a year. Prediction of employees’ absenteeism and the reasons behind their absence help organizations in reducing expenses and increasing productivity. Data mining turns the vast volume of human resources data into information that can help in decision-making and prediction. Although the selection of features is a critical step in data mining to enhance the efficiency of the final prediction, it is not yet known which method of feature selection is better. Therefore, this paper aims to compare the performance of three well-known feature selection methods in absenteeism prediction, which are relief-based feature selection, correlation-based feature selection and information-gain feature selection. In addition, this paper aims to find the best combination of feature selection method and data mining technique in enhancing the absenteeism prediction accuracy. Seven classification techniques were used as the prediction model. Additionally, cross-validation approach was utilized to assess the applied prediction models to have more realistic and reliable results. The used dataset was built at a courier company in Brazil with records of absenteeism at work. Regarding experimental results, correlationbased feature selection surpasses the other methods through the performance measurements. Furthermore, bagging classifier was the best-performing data mining technique when features were selected using correlation-based feature selection with an accuracy rate of (92%).
Clustering Prediction Techniques in Defining and Predicting Customers Defecti...IJECEIAES
With the growth of the e-commerce sector, customers have more choices, a fact which encourages them to divide their purchases amongst several ecommerce sites and compare their competitors‟ products, yet this increases high risks of churning. A review of the literature on customer churning models reveals that no prior research had considered both partial and total defection in non-contractual online environments. Instead, they focused either on a total or partial defect. This study proposes a customer churn prediction model in an e-commerce context, wherein a clustering phase is based on the integration of the k-means method and the Length-RecencyFrequency-Monetary (LRFM) model. This phase is employed to define churn followed by a multi-class prediction phase based on three classification techniques: Simple decision tree, Artificial neural networks and Decision tree ensemble, in which the dependent variable classifies a particular customer into a customer continuing loyal buying patterns (Non-churned), a partial defector (Partially-churned), and a total defector (Totally-churned). Macroaveraging measures including average accuracy, macro-average of Precision, Recall, and F-1 are used to evaluate classifiers‟ performance on 10-fold cross validation. Using real data from an online store, the results show the efficiency of decision tree ensemble model over the other models in identifying both future partial and total defection.
Projection pursuit Random Forest using discriminant feature analysis model fo...IJECEIAES
A major and demand issue in the telecommunications industry is the prediction of churn customers. Churn describes the customer who attrites from the current provider to competitors searching for better service offers. Companies from the Telco sector frequently have customer relationship management offices it is the main objective in how to win back defecting clients because preserve long-term customers can be much more beneficial than gain newly recruited customers. Researchers and practitioners are paying great attention to developing a robust customer churn prediction model, especially in the telecommunication business by proposed numerous machine learning approaches. Many approaches of Classification are established, but the most effective in recent times is a tree-based method. The main contribution of this research is to predict churners/non-churners in the Telecom sector based on project pursuit Random Forest (PPForest) that uses discriminant feature analysis as a novelty extension of the conventional Random Forest for learning oblique Project Pursuit tree (PPtree). The proposed methodology leverages the advantage of two discriminant analysis methods to calculate the project index used in the construction of PPtree. The first method used Support Vector Machines (SVM) while, the second method used Linear Discriminant Analysis (LDA) to achieve linear splitting of variables during oblique PPtree construction to produce individual classifiers that are robust and more diverse than classical Random Forest. It is found that the proposed methods enjoy the best performance measurements e.g. Accuracy, hit rate, ROC curve, Lift, H-measure, AUC. Moreover, PPForest based on LDA delivers effective evaluators in the prediction model.
On the benefit of logic-based machine learning to learn pairwise comparisonsjournalBEEI
In recent years, many daily processes such as internet web searching, e-mail filter-ing, social media services, e-commerce have benefited from machine learning tech-niques (ML). The implementation of ML techniques has been largely focused on blackbox methods where the general conclusions are not easily interpretable. Hence, theelaboration with other declarative software models to identify the correctness and com-pleteness of the models is not easy to perform. On the other hand, the emerge of somelogic-based machine learning techniques with their advantage of white box approachhave been proven to be well-suited for many software engineering tasks. In this paper,we propose the use of a logic-based approach to learn user preference in the form ofpairwise comparisons. APARELL as a novel approach of inductive learning is able tomodel the user’s preferences in description logic representation. This offers a rich, re-lational representation which is then can be used to produce a set of recommendations.A user study has been performed in our experiment to evaluate the implementation ofpairwise preference recommender system when compared to a standard list interface.The result of the experiment shows that the pairwise interface was significantly betterthan the other interface in many ways.
The use of genetic algorithm, clustering and feature selection techniques in ...IJMIT JOURNAL
Decision tree modelling, as one of data mining techniques, is used for credit scoring of bank customers.
The main problem is the construction of decision trees that could classify customers optimally. This study
presents a new hybrid mining approach in the design of an effective and appropriate credit scoring model.
It is based on genetic algorithm for credit scoring of bank customers in order to offer credit facilities to
each class of customers. Genetic algorithm can help banks in credit scoring of customers by selecting
appropriate features and building optimum decision trees. The new proposed hybrid classification model is
established based on a combination of clustering, feature selection, decision trees, and genetic algorithm
techniques. We used clustering and feature selection techniques to pre-process the input samples to
construct the decision trees in the credit scoring model. The proposed hybrid model choices and combines
the best decision trees based on the optimality criteria. It constructs the final decision tree for credit
scoring of customers. Using one credit dataset, results confirm that the classification accuracy of the
proposed hybrid classification model is more than almost the entire classification models that have been
compared in this paper. Furthermore, the number of leaves and the size of the constructed decision tree
(i.e. complexity) are less, compared with other decision tree models. In this work, one financial dataset was
chosen for experiments, including Bank Mellat credit dataset.
Extended pso algorithm for improvement problems k means clustering algorithmIJMIT JOURNAL
The clustering is a without monitoring process and one of the most common data mining techniques. The
purpose of clustering is grouping similar data together in a group, so were most similar to each other in a
cluster and the difference with most other instances in the cluster are. In this paper we focus on clustering
partition k-means, due to ease of implementation and high-speed performance of large data sets, After 30
year it is still very popular among the developed clustering algorithm and then for improvement problem of
placing of k-means algorithm in local optimal, we pose extended PSO algorithm, that its name is ECPSO.
Our new algorithm is able to be cause of exit from local optimal and with high percent produce the
problem’s optimal answer. The probe of results show that mooted algorithm have better performance
regards as other clustering algorithms specially in two index, the carefulness of clustering and the quality
of clustering.
Identification of important features and data mining classification technique...IJECEIAES
Employees absenteeism at the work costs organizations billions a year. Prediction of employees’ absenteeism and the reasons behind their absence help organizations in reducing expenses and increasing productivity. Data mining turns the vast volume of human resources data into information that can help in decision-making and prediction. Although the selection of features is a critical step in data mining to enhance the efficiency of the final prediction, it is not yet known which method of feature selection is better. Therefore, this paper aims to compare the performance of three well-known feature selection methods in absenteeism prediction, which are relief-based feature selection, correlation-based feature selection and information-gain feature selection. In addition, this paper aims to find the best combination of feature selection method and data mining technique in enhancing the absenteeism prediction accuracy. Seven classification techniques were used as the prediction model. Additionally, cross-validation approach was utilized to assess the applied prediction models to have more realistic and reliable results. The used dataset was built at a courier company in Brazil with records of absenteeism at work. Regarding experimental results, correlationbased feature selection surpasses the other methods through the performance measurements. Furthermore, bagging classifier was the best-performing data mining technique when features were selected using correlation-based feature selection with an accuracy rate of (92%).
Clustering Prediction Techniques in Defining and Predicting Customers Defecti...IJECEIAES
With the growth of the e-commerce sector, customers have more choices, a fact which encourages them to divide their purchases amongst several ecommerce sites and compare their competitors‟ products, yet this increases high risks of churning. A review of the literature on customer churning models reveals that no prior research had considered both partial and total defection in non-contractual online environments. Instead, they focused either on a total or partial defect. This study proposes a customer churn prediction model in an e-commerce context, wherein a clustering phase is based on the integration of the k-means method and the Length-RecencyFrequency-Monetary (LRFM) model. This phase is employed to define churn followed by a multi-class prediction phase based on three classification techniques: Simple decision tree, Artificial neural networks and Decision tree ensemble, in which the dependent variable classifies a particular customer into a customer continuing loyal buying patterns (Non-churned), a partial defector (Partially-churned), and a total defector (Totally-churned). Macroaveraging measures including average accuracy, macro-average of Precision, Recall, and F-1 are used to evaluate classifiers‟ performance on 10-fold cross validation. Using real data from an online store, the results show the efficiency of decision tree ensemble model over the other models in identifying both future partial and total defection.
Projection pursuit Random Forest using discriminant feature analysis model fo...IJECEIAES
A major and demand issue in the telecommunications industry is the prediction of churn customers. Churn describes the customer who attrites from the current provider to competitors searching for better service offers. Companies from the Telco sector frequently have customer relationship management offices it is the main objective in how to win back defecting clients because preserve long-term customers can be much more beneficial than gain newly recruited customers. Researchers and practitioners are paying great attention to developing a robust customer churn prediction model, especially in the telecommunication business by proposed numerous machine learning approaches. Many approaches of Classification are established, but the most effective in recent times is a tree-based method. The main contribution of this research is to predict churners/non-churners in the Telecom sector based on project pursuit Random Forest (PPForest) that uses discriminant feature analysis as a novelty extension of the conventional Random Forest for learning oblique Project Pursuit tree (PPtree). The proposed methodology leverages the advantage of two discriminant analysis methods to calculate the project index used in the construction of PPtree. The first method used Support Vector Machines (SVM) while, the second method used Linear Discriminant Analysis (LDA) to achieve linear splitting of variables during oblique PPtree construction to produce individual classifiers that are robust and more diverse than classical Random Forest. It is found that the proposed methods enjoy the best performance measurements e.g. Accuracy, hit rate, ROC curve, Lift, H-measure, AUC. Moreover, PPForest based on LDA delivers effective evaluators in the prediction model.
On the benefit of logic-based machine learning to learn pairwise comparisonsjournalBEEI
In recent years, many daily processes such as internet web searching, e-mail filter-ing, social media services, e-commerce have benefited from machine learning tech-niques (ML). The implementation of ML techniques has been largely focused on blackbox methods where the general conclusions are not easily interpretable. Hence, theelaboration with other declarative software models to identify the correctness and com-pleteness of the models is not easy to perform. On the other hand, the emerge of somelogic-based machine learning techniques with their advantage of white box approachhave been proven to be well-suited for many software engineering tasks. In this paper,we propose the use of a logic-based approach to learn user preference in the form ofpairwise comparisons. APARELL as a novel approach of inductive learning is able tomodel the user’s preferences in description logic representation. This offers a rich, re-lational representation which is then can be used to produce a set of recommendations.A user study has been performed in our experiment to evaluate the implementation ofpairwise preference recommender system when compared to a standard list interface.The result of the experiment shows that the pairwise interface was significantly betterthan the other interface in many ways.
The use of genetic algorithm, clustering and feature selection techniques in ...IJMIT JOURNAL
Decision tree modelling, as one of data mining techniques, is used for credit scoring of bank customers.
The main problem is the construction of decision trees that could classify customers optimally. This study
presents a new hybrid mining approach in the design of an effective and appropriate credit scoring model.
It is based on genetic algorithm for credit scoring of bank customers in order to offer credit facilities to
each class of customers. Genetic algorithm can help banks in credit scoring of customers by selecting
appropriate features and building optimum decision trees. The new proposed hybrid classification model is
established based on a combination of clustering, feature selection, decision trees, and genetic algorithm
techniques. We used clustering and feature selection techniques to pre-process the input samples to
construct the decision trees in the credit scoring model. The proposed hybrid model choices and combines
the best decision trees based on the optimality criteria. It constructs the final decision tree for credit
scoring of customers. Using one credit dataset, results confirm that the classification accuracy of the
proposed hybrid classification model is more than almost the entire classification models that have been
compared in this paper. Furthermore, the number of leaves and the size of the constructed decision tree
(i.e. complexity) are less, compared with other decision tree models. In this work, one financial dataset was
chosen for experiments, including Bank Mellat credit dataset.
Biometric Identification and Authentication Providence using Fingerprint for ...IJECEIAES
The raise in the recent security incidents of cloud computing and its challenges is to secure the data. To solve this problem, the integration of mobile with cloud computing, Mobile biometric authentication in cloud computing is presented in this paper. To enhance the security, the biometric authentication is being used, since the Mobile cloud computing is popular among the mobile user. This paper examines how the mobile cloud computing (MCC) is used in security issue with finger biometric authentication model. Through this fingerprint biometric, the secret code is generated by entropy value. This enables the person to request for accessing the data in the desk computer. When the person requests the access to the authorized user through Bluetooth in mobile, the Authorized user sends the permit access through fingerprint secret code. Finally this fingerprint is verified with the database in the Desk computer. If it is matched, then the computer can be accessed by the requested person.
Introduction to feature subset selection methodIJSRD
Data Mining is a computational progression to ascertain patterns in hefty data sets. It has various important techniques and one of them is Classification which is receiving great attention recently in the database community. Classification technique can solve several problems in different fields like medicine, industry, business, science. PSO is based on social behaviour for optimization problem. Feature Selection (FS) is a solution that involves finding a subset of prominent features to improve predictive accuracy and to remove the redundant features. Rough Set Theory (RST) is a mathematical tool which deals with the uncertainty and vagueness of the decision systems.
REVIEWING PROCESS MINING APPLICATIONS AND TECHNIQUES IN EDUCATIONijaia
Process Mining (PM) emerged from business process management but has recently been applied to
educational data and has been found to facilitate the understanding of the educational process.
Educational Process Mining (EPM) bridges the gap between process analysis and data analysis, based on
the techniques of model discovery, conformance checking and extension of existing process models. We
present a systematic review of the recent and current status of research in the EPM domain, focusing on
application domains, techniques, tools and models, to highlight the use of EPM in comprehending and
improving educational processes.
Performance Analysis of Selected Classifiers in User Profilingijdmtaiir
User profiles can serve as indicators of personal
preferences which can be effectively used while providing
personalized services. Building user files which can capture
accurate information of individuals has been a daunting task.
Several attempts have been made by researchers to extract
information from different data sources to build user profiles
on different application domains. Towards this end, in this
paper we employ different classification algorithmsto create
accurate user profiles based on information gathered from
demographic data. The aim of this work is to analyze the
performance of five most effective classification methods,
namely Bayesian Network(BN), Naïve Bayesian(NB), Naives
Bayes Updateable(NBU), J48, and Decision Table(DT). Our
simulation results show that, in general, the J48has the highest
classification accuracy performance with the lowest error rate.
On the other hand, it is found that Naïve Bayesian and Naives
Bayes Updateable classifiers have the lowest time requirement
to build the classification model
MITIGATION TECHNIQUES TO OVERCOME DATA HARM IN MODEL BUILDING FOR MLijaia
Given the impact of Machine Learning (ML) on individuals and the society, understanding how harm might
be occur throughout the ML life cycle becomes critical more than ever. By offering a framework to
determine distinct potential sources of downstream harm in ML pipeline, the paper demonstrates the
importance of choices throughout distinct phases of data collection, development, and deployment that
extend far beyond just model training. Relevant mitigation techniques are also suggested for being used
instead of merely relying on generic notions of what counts as fairness.
Paper Annotated: SinGAN-Seg: Synthetic Training Data Generation for Medical I...Devansh16
YouTube video: https://www.youtube.com/watch?v=Ao-19L0sLOI
SinGAN-Seg: Synthetic Training Data Generation for Medical Image Segmentation
Vajira Thambawita, Pegah Salehi, Sajad Amouei Sheshkal, Steven A. Hicks, Hugo L.Hammer, Sravanthi Parasa, Thomas de Lange, Pål Halvorsen, Michael A. Riegler
Processing medical data to find abnormalities is a time-consuming and costly task, requiring tremendous efforts from medical experts. Therefore, Ai has become a popular tool for the automatic processing of medical data, acting as a supportive tool for doctors. AI tools highly depend on data for training the models. However, there are several constraints to access to large amounts of medical data to train machine learning algorithms in the medical domain, e.g., due to privacy concerns and the costly, time-consuming medical data annotation process. To address this, in this paper we present a novel synthetic data generation pipeline called SinGAN-Seg to produce synthetic medical data with the corresponding annotated ground truth masks. We show that these synthetic data generation pipelines can be used as an alternative to bypass privacy concerns and as an alternative way to produce artificial segmentation datasets with corresponding ground truth masks to avoid the tedious medical data annotation process. As a proof of concept, we used an open polyp segmentation dataset. By training UNet++ using both the real polyp segmentation dataset and the corresponding synthetic dataset generated from the SinGAN-Seg pipeline, we show that the synthetic data can achieve a very close performance to the real data when the real segmentation datasets are large enough. In addition, we show that synthetic data generated from the SinGAN-Seg pipeline improving the performance of segmentation algorithms when the training dataset is very small. Since our SinGAN-Seg pipeline is applicable for any medical dataset, this pipeline can be used with any other segmentation datasets.
Subjects: Image and Video Processing (eess.IV); Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG)
Cite as: arXiv:2107.00471 [eess.IV]
(or arXiv:2107.00471v1 [eess.IV] for this version)
Reach out to me:
Check out my other articles on Medium. : https://machine-learning-made-simple....
My YouTube: https://rb.gy/88iwdd
Reach out to me on LinkedIn: https://www.linkedin.com/in/devansh-d...
My Instagram: https://rb.gy/gmvuy9
My Twitter: https://twitter.com/Machine01776819
My Substack: https://devanshacc.substack.com/
Live conversations at twitch here: https://rb.gy/zlhk9y
Get a free stock on Robinhood: https://join.robinhood.com/fnud75
When deep learners change their mind learning dynamics for active learningDevansh16
Abstract:
Active learning aims to select samples to be annotated that yield the largest performance improvement for the learning algorithm. Many methods approach this problem by measuring the informativeness of samples and do this based on the certainty of the network predictions for samples. However, it is well-known that neural networks are overly confident about their prediction and are therefore an untrustworthy source to assess sample informativeness. In this paper, we propose a new informativeness-based active learning method. Our measure is derived from the learning dynamics of a neural network. More precisely we track the label assignment of the unlabeled data pool during the training of the algorithm. We capture the learning dynamics with a metric called label-dispersion, which is low when the network consistently assigns the same label to the sample during the training of the network and high when the assigned label changes frequently. We show that label-dispersion is a promising predictor of the uncertainty of the network, and show on two benchmark datasets that an active learning algorithm based on label-dispersion obtains excellent results.
DEEP-LEARNING-BASED HUMAN INTENTION PREDICTION WITH DATA AUGMENTATIONijaia
Data augmentation has been broadly applied in training deep-learning models to increase the diversity of
data. This study ingestigates the effectiveness of different data augmentation methods for deep-learningbased human intention prediction when only limited training data is available. A human participant pitches
a ball to nine potential targets in our experiment. We expect to predict which target the participant pitches
the ball to. Firstly, the effectiveness of 10 data augmentation groups is evaluated on a single-participant
data set using RGB images. Secondly, the best data augmentation method (i.e., random cropping) on the
single-participant data set is further evaluated on a multi-participant data set to assess its generalization
ability. Finally, the effectiveness of random cropping on fusion data of RGB images and optical flow is
evaluated on both single- and multi-participant data sets. Experiment results show that: 1) Data
augmentation methods that crop or deform images can improve the prediction performance; 2) Random
cropping can be generalized to the multi-participant data set (prediction accuracy is improved from 50%
to 57.4%); and 3) Random cropping with fusion data of RGB images and optical flow can further improve
the prediction accuracy from 57.4% to 63.9% on the multi-participant data set.
Framework for opinion as a service on review data of customer using semantics...IJECEIAES
At opinion mining plays a significant role in representing the original and unbiased perception of the products/services. However, there are various challenges associated with performing an effective opinion mining in the present era of distributed computing system with dynamic behaviour of users. Existing approaches is more laborious towards extracting knowledge from the reviews of user which is further subjected to various rounds of operation with complex procedures. The proposed system addresses the problem by introducing a novel framework called as opinion-as-a-service which is meant for direct utilization of the extracted knowledge in most user friendly manner. The proposed system introduces a set of three sequential algorithm that performs aggregated of incoming stream of opinion data, performing indexing, followed by applying semantics for extracting knowledge. The study outcome shows that proposed system is better than existing system in mining performance.
DATA AUGMENTATION TECHNIQUES AND TRANSFER LEARNING APPROACHES APPLIED TO FACI...ijaia
The face expression is the first thing we pay attention to when we want to understand a person’s state of
mind. Thus, the ability to recognize facial expressions in an automatic way is a very interesting research
field. In this paper, because the small size of available training datasets, we propose a novel data
augmentation technique that improves the performances in the recognition task. We apply geometrical
transformations and build from scratch GAN models able to generate new synthetic images for each
emotion type. Thus, on the augmented datasets we fine tune pretrained convolutional neural networks with
different architectures. To measure the generalization ability of the models, we apply extra-database
protocol approach, namely we train models on the augmented versions of training dataset and test them on
two different databases. The combination of these techniques allows to reach average accuracy values of
the order of 85% for the InceptionResNetV2 model.
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
Abstract Learning Analytics by nature relies on computational information processing activities intended to extract from raw data some interesting aspects that can be used to obtain insights into the behaviors of learners, the design of learning experiences, etc. There is a large variety of computational techniques that can be employed, all with interesting properties, but it is the interpretation of their results that really forms the core of the analytics process. As a rising subject, data mining and business intelligence are playing an increasingly important role in the decision support activity of every walk of life. The Variance Rover System (VRS) mainly focused on the large data sets obtained from online web visiting and categorizing this into clusters according some similarity and the process of predicting customer behavior and selecting actions to influence that behavior to benefit the company, so as to take optimized and beneficial decisions of business expansion. Keywords: Analytics, Business intelligence, Clustering, Data Mining, Standard K-means, Optimized K-means
Biometric Identification and Authentication Providence using Fingerprint for ...IJECEIAES
The raise in the recent security incidents of cloud computing and its challenges is to secure the data. To solve this problem, the integration of mobile with cloud computing, Mobile biometric authentication in cloud computing is presented in this paper. To enhance the security, the biometric authentication is being used, since the Mobile cloud computing is popular among the mobile user. This paper examines how the mobile cloud computing (MCC) is used in security issue with finger biometric authentication model. Through this fingerprint biometric, the secret code is generated by entropy value. This enables the person to request for accessing the data in the desk computer. When the person requests the access to the authorized user through Bluetooth in mobile, the Authorized user sends the permit access through fingerprint secret code. Finally this fingerprint is verified with the database in the Desk computer. If it is matched, then the computer can be accessed by the requested person.
Introduction to feature subset selection methodIJSRD
Data Mining is a computational progression to ascertain patterns in hefty data sets. It has various important techniques and one of them is Classification which is receiving great attention recently in the database community. Classification technique can solve several problems in different fields like medicine, industry, business, science. PSO is based on social behaviour for optimization problem. Feature Selection (FS) is a solution that involves finding a subset of prominent features to improve predictive accuracy and to remove the redundant features. Rough Set Theory (RST) is a mathematical tool which deals with the uncertainty and vagueness of the decision systems.
REVIEWING PROCESS MINING APPLICATIONS AND TECHNIQUES IN EDUCATIONijaia
Process Mining (PM) emerged from business process management but has recently been applied to
educational data and has been found to facilitate the understanding of the educational process.
Educational Process Mining (EPM) bridges the gap between process analysis and data analysis, based on
the techniques of model discovery, conformance checking and extension of existing process models. We
present a systematic review of the recent and current status of research in the EPM domain, focusing on
application domains, techniques, tools and models, to highlight the use of EPM in comprehending and
improving educational processes.
Performance Analysis of Selected Classifiers in User Profilingijdmtaiir
User profiles can serve as indicators of personal
preferences which can be effectively used while providing
personalized services. Building user files which can capture
accurate information of individuals has been a daunting task.
Several attempts have been made by researchers to extract
information from different data sources to build user profiles
on different application domains. Towards this end, in this
paper we employ different classification algorithmsto create
accurate user profiles based on information gathered from
demographic data. The aim of this work is to analyze the
performance of five most effective classification methods,
namely Bayesian Network(BN), Naïve Bayesian(NB), Naives
Bayes Updateable(NBU), J48, and Decision Table(DT). Our
simulation results show that, in general, the J48has the highest
classification accuracy performance with the lowest error rate.
On the other hand, it is found that Naïve Bayesian and Naives
Bayes Updateable classifiers have the lowest time requirement
to build the classification model
MITIGATION TECHNIQUES TO OVERCOME DATA HARM IN MODEL BUILDING FOR MLijaia
Given the impact of Machine Learning (ML) on individuals and the society, understanding how harm might
be occur throughout the ML life cycle becomes critical more than ever. By offering a framework to
determine distinct potential sources of downstream harm in ML pipeline, the paper demonstrates the
importance of choices throughout distinct phases of data collection, development, and deployment that
extend far beyond just model training. Relevant mitigation techniques are also suggested for being used
instead of merely relying on generic notions of what counts as fairness.
Paper Annotated: SinGAN-Seg: Synthetic Training Data Generation for Medical I...Devansh16
YouTube video: https://www.youtube.com/watch?v=Ao-19L0sLOI
SinGAN-Seg: Synthetic Training Data Generation for Medical Image Segmentation
Vajira Thambawita, Pegah Salehi, Sajad Amouei Sheshkal, Steven A. Hicks, Hugo L.Hammer, Sravanthi Parasa, Thomas de Lange, Pål Halvorsen, Michael A. Riegler
Processing medical data to find abnormalities is a time-consuming and costly task, requiring tremendous efforts from medical experts. Therefore, Ai has become a popular tool for the automatic processing of medical data, acting as a supportive tool for doctors. AI tools highly depend on data for training the models. However, there are several constraints to access to large amounts of medical data to train machine learning algorithms in the medical domain, e.g., due to privacy concerns and the costly, time-consuming medical data annotation process. To address this, in this paper we present a novel synthetic data generation pipeline called SinGAN-Seg to produce synthetic medical data with the corresponding annotated ground truth masks. We show that these synthetic data generation pipelines can be used as an alternative to bypass privacy concerns and as an alternative way to produce artificial segmentation datasets with corresponding ground truth masks to avoid the tedious medical data annotation process. As a proof of concept, we used an open polyp segmentation dataset. By training UNet++ using both the real polyp segmentation dataset and the corresponding synthetic dataset generated from the SinGAN-Seg pipeline, we show that the synthetic data can achieve a very close performance to the real data when the real segmentation datasets are large enough. In addition, we show that synthetic data generated from the SinGAN-Seg pipeline improving the performance of segmentation algorithms when the training dataset is very small. Since our SinGAN-Seg pipeline is applicable for any medical dataset, this pipeline can be used with any other segmentation datasets.
Subjects: Image and Video Processing (eess.IV); Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG)
Cite as: arXiv:2107.00471 [eess.IV]
(or arXiv:2107.00471v1 [eess.IV] for this version)
Reach out to me:
Check out my other articles on Medium. : https://machine-learning-made-simple....
My YouTube: https://rb.gy/88iwdd
Reach out to me on LinkedIn: https://www.linkedin.com/in/devansh-d...
My Instagram: https://rb.gy/gmvuy9
My Twitter: https://twitter.com/Machine01776819
My Substack: https://devanshacc.substack.com/
Live conversations at twitch here: https://rb.gy/zlhk9y
Get a free stock on Robinhood: https://join.robinhood.com/fnud75
When deep learners change their mind learning dynamics for active learningDevansh16
Abstract:
Active learning aims to select samples to be annotated that yield the largest performance improvement for the learning algorithm. Many methods approach this problem by measuring the informativeness of samples and do this based on the certainty of the network predictions for samples. However, it is well-known that neural networks are overly confident about their prediction and are therefore an untrustworthy source to assess sample informativeness. In this paper, we propose a new informativeness-based active learning method. Our measure is derived from the learning dynamics of a neural network. More precisely we track the label assignment of the unlabeled data pool during the training of the algorithm. We capture the learning dynamics with a metric called label-dispersion, which is low when the network consistently assigns the same label to the sample during the training of the network and high when the assigned label changes frequently. We show that label-dispersion is a promising predictor of the uncertainty of the network, and show on two benchmark datasets that an active learning algorithm based on label-dispersion obtains excellent results.
DEEP-LEARNING-BASED HUMAN INTENTION PREDICTION WITH DATA AUGMENTATIONijaia
Data augmentation has been broadly applied in training deep-learning models to increase the diversity of
data. This study ingestigates the effectiveness of different data augmentation methods for deep-learningbased human intention prediction when only limited training data is available. A human participant pitches
a ball to nine potential targets in our experiment. We expect to predict which target the participant pitches
the ball to. Firstly, the effectiveness of 10 data augmentation groups is evaluated on a single-participant
data set using RGB images. Secondly, the best data augmentation method (i.e., random cropping) on the
single-participant data set is further evaluated on a multi-participant data set to assess its generalization
ability. Finally, the effectiveness of random cropping on fusion data of RGB images and optical flow is
evaluated on both single- and multi-participant data sets. Experiment results show that: 1) Data
augmentation methods that crop or deform images can improve the prediction performance; 2) Random
cropping can be generalized to the multi-participant data set (prediction accuracy is improved from 50%
to 57.4%); and 3) Random cropping with fusion data of RGB images and optical flow can further improve
the prediction accuracy from 57.4% to 63.9% on the multi-participant data set.
Framework for opinion as a service on review data of customer using semantics...IJECEIAES
At opinion mining plays a significant role in representing the original and unbiased perception of the products/services. However, there are various challenges associated with performing an effective opinion mining in the present era of distributed computing system with dynamic behaviour of users. Existing approaches is more laborious towards extracting knowledge from the reviews of user which is further subjected to various rounds of operation with complex procedures. The proposed system addresses the problem by introducing a novel framework called as opinion-as-a-service which is meant for direct utilization of the extracted knowledge in most user friendly manner. The proposed system introduces a set of three sequential algorithm that performs aggregated of incoming stream of opinion data, performing indexing, followed by applying semantics for extracting knowledge. The study outcome shows that proposed system is better than existing system in mining performance.
DATA AUGMENTATION TECHNIQUES AND TRANSFER LEARNING APPROACHES APPLIED TO FACI...ijaia
The face expression is the first thing we pay attention to when we want to understand a person’s state of
mind. Thus, the ability to recognize facial expressions in an automatic way is a very interesting research
field. In this paper, because the small size of available training datasets, we propose a novel data
augmentation technique that improves the performances in the recognition task. We apply geometrical
transformations and build from scratch GAN models able to generate new synthetic images for each
emotion type. Thus, on the augmented datasets we fine tune pretrained convolutional neural networks with
different architectures. To measure the generalization ability of the models, we apply extra-database
protocol approach, namely we train models on the augmented versions of training dataset and test them on
two different databases. The combination of these techniques allows to reach average accuracy values of
the order of 85% for the InceptionResNetV2 model.
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
Abstract Learning Analytics by nature relies on computational information processing activities intended to extract from raw data some interesting aspects that can be used to obtain insights into the behaviors of learners, the design of learning experiences, etc. There is a large variety of computational techniques that can be employed, all with interesting properties, but it is the interpretation of their results that really forms the core of the analytics process. As a rising subject, data mining and business intelligence are playing an increasingly important role in the decision support activity of every walk of life. The Variance Rover System (VRS) mainly focused on the large data sets obtained from online web visiting and categorizing this into clusters according some similarity and the process of predicting customer behavior and selecting actions to influence that behavior to benefit the company, so as to take optimized and beneficial decisions of business expansion. Keywords: Analytics, Business intelligence, Clustering, Data Mining, Standard K-means, Optimized K-means
A large number of techniques has been developed so far to tell the diversity of machine learning. Machine learning is categorized into supervised, unsupervised and reinforcement learning .Every instance in given data-set used by Machine learning algorithms is represented same set of features .On basis of label of instances it is divided into category. In this review paper our main focus is on Supervised, unsupervised learning techniques and its performance parameters.
Object Detection using Deep Learning with Hierarchical Multi Swarm Optimizationrahulmonikasharma
Till now there is a huge research had in the field of visual information retrieval, but with the growth of data and with less processing speed we are not meeting the needs of current problem. The main focus of this paper is to identify the objects with salient features and object highlighting. Till now object identification is done with the pixel based or with the region based. Different methodologies are compared in this work and they will be processed with the learning work. Multi scale contrast is one of the pixel based technology where object borders are identified but not the object. This can be done with the histogram contrast. Still it is not covering all the features of the object and it is not clear in identifying the objects at high contrast regions. To solve this issue region based contrasting method is used which is the better solution for all this object identification. After extracting the features and identifying the object, now auto classification or identification of the object should be done. The other part of the work mainly concentrates on the learning system which uses most popular neural network algorithms. Identifying the drawbacks of neural network algorithms and proposing the new methodology identify the objects is done in this paper.
Anomaly detection is usually associate degree identification of associate degree odd or abnormal
information typically even known as as an outlier from a offer pattern of information. It involves machine learning
technique to be told the info and verify the outliers supported a likelihood condition. Machine learning, a branch
of AI plays a significant role in analyzing the info and identifies the outliers with a decent likelihood. The target of this
paper is to work out the outlier supported anomaly detection techniques and describe the quality standards of the actual
trade. We have a tendency to describe associate degree approach to analyzing anomalies in trade information
supported the identification of cluster outliers.
Abstract In this paper, the concept of data mining was summarized and its significance towards its methodologies was illustrated. The data mining based on Neural Network and Genetic Algorithm is researched in detail and the key technology and ways to achieve the data mining on Neural Network and Genetic Algorithm are also surveyed. This paper also conducts a formal review of the area of rule extraction from ANN and GA. Keywords: Data Mining, Neural Network, Genetic Algorithm, Rule Extraction.
PROVIDING A METHOD FOR DETERMINING THE INDEX OF CUSTOMER CHURN IN INDUSTRYIJITCA Journal
Churn customer, one of the most important issues in customer relationship management and marketing is especially in industries such as telecommunications, the financial and insurance. In recent decades much
research has been done in this area. In this research, the index set for the reasons set reason churn customers for our customers is of particular importance. In this study we are intended to provide a formula for the index churn customers, the better to understand the reasons for customers to provide churn. Therefore, in order to evaluate the formula provided through six Classification methods (Decision tree QUEST, Decision tree C5.0, Decision tree CHAID, Decision trees CART, Bayesian network, Neural network) to evaluate the formula will be involved with individual indicators
Applying Convolutional-GRU for Term Deposit Likelihood PredictionVandanaSharma356
Banks are normally offered two kinds of deposit accounts. It consists of deposits like current/saving account and term deposits like fixed or recurring deposits.For enhancing the maximized profit from bank as well as customer perspective, term deposit can accelerate uplifting of finance fields. This paper focuses on likelihood of term deposit subscription taken by the customers. Bank campaign efforts and customer detail analysis caninfluence term deposit subscription chances. An automated system is approached in this paper that works towards prediction of term deposit investment possibilities in advance. This paper proposes deep learning based hybrid model that stacks Convolutional layers and Recurrent Neural Network (RNN) layers as predictive model. For RNN, Gated Recurrent Unit (GRU) is employed. The proposed predictive model is later compared with other benchmark classifiers such as k-Nearest Neighbor (k-NN), Decision tree classifier (DT), and Multi-layer perceptron classifier (MLP). Experimental study concludesthat proposed model attainsan accuracy of 89.59% and MSE of 0.1041 which outperform wellother baseline models.
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
Predicting churn with filter-based techniques and deep learningIJECEIAES
Customer churn prediction is of utmost importance in the telecommunications industry. Retaining customers through effective churn prevention strategies proves to be more cost-efficient. In this study, attribute selection analysis and deep learning are integrated to develop a customer churn prediction model to improve performance while reducing feature dimensions. The study includes the analysis of customer data attributes, exploratory data analysis, and data preprocessing for data quality enhancement. Next, significant features are selected using two attribute selection techniques, which are chi-square and analysis of variance (ANOVA). The selected features are fed into an artificial neural network (ANN) model for analysis and prediction. To enhance prediction performance and stability, a learning rate scheduler is deployed. Implementing the learning rate scheduler in the model can help prevent overfitting and enhance convergence speed. By dynamically adjusting the learning rate during the training process, the scheduler ensures that the model optimally adapts to the data while avoiding overfitting. The proposed model is evaluated using the Cell2Cell telecom database, and the results demonstrate that the proposed model exhibits a promising performance, showcasing its potential as an effective churn prediction solution in the telecommunications industry.
Similar to Prediction of Default Customer in Banking Sector using Artificial Neural Network (20)
Data Mining is a significant field in today’s data-driven world. Understanding and implementing its concepts can lead to discovery of useful insights. This paper discusses the main concepts of data mining, focusing on two main concepts namely Association Rule Mining and Time Series Analysis
A Review on Real Time Integrated CCTV System Using Face Detection for Vehicle...rahulmonikasharma
We are describes the technique for real time human face detection and counting the number of passengers in vehicle and also gender of the passengers.The Image processing technology is very popular,now at present all are going to use it for various purpose. It can be applied to various applications for detecting and processing the digital images. Face detection is a part of image processing. It is used for finding the face of human in a given area. Face detection is used in many applications such as face recognition, people tracking, or photography. In this paper,The webcam is installed in public vehicle and connected with Raspberry Pi model. We use face detection technique for detecting and counting the number of passengers in public vehicle via webcam with the help of image processing and Raspberry Pi.
Considering Two Sides of One Review Using Stanford NLP Frameworkrahulmonikasharma
Sentiment analysis is a type of natural language processing for tracking the mood of the public about a particular product or a topic and is useful in several ways. Polarity shift is the most classical task which aims at classifying the reviews either positive or negative. But in many cases, in addition to the positive and negative reviews, there still many neutral reviews exist. However, the performance sometimes limited due to the fundamental deficiencies in handling the polarity shift problem. We propose an Improvised Dual Sentiment Analysis (IDSA) model to address this problem for sentiment classification. We first propose a novel data expansion technique by creating sentiment-reversed review for each training and test review. We develop a corpus based method to construct a pseudo-antonym dictionary. It removes DSA’s dependency on an external antonym dictionary for review reversion. We conduct a range of experiments and the results demonstrates the effectiveness of DSA in addressing the polarity shift in sentiment classification. .
A New Detection and Decoding Technique for (2×N_r ) MIMO Communication Systemsrahulmonikasharma
The requirements of fifth generation new radio (5G- NR) access networks are very high capacity and ultra-reliability. In this paper, we proposed a V-BLAST2 × N_r MIMO system that is analyzed, improved, and expected to achieve both very high throughput and ultra- reliability simultaneously.A new detection technique called parallel detection algorithm is proposed. The performance of the proposed algorithm compared with existing linear detection algorithms. It was seen that the proposed technique increases the speed of signal transmission and prevents error propagation which may be present in serial decoding techniques. The new algorithm reduces the bit error probability and increases the capacity simultaneouslywithout using a standard STC technique. However, it was seen that the BER of systems using the proposed algorithm is slightly higher than a similar system using only STC technique. Simulation results show the advantages of using the proposed technique.
Broadcasting Scenario under Different Protocols in MANET: A Surveyrahulmonikasharma
A wireless network enables people to communicate and access applications and information without wires. This provides freedom of movement and the ability to extend applications to different parts of a building, city, or nearly anywhere in the world. Wireless networks allow people to interact with e-mail or browse the Internet from a location that they prefer. Adhoc Networks are self-organizing wireless networks, absent any fixed infrastructure. broadcasting of data through proper channel is essential. Various protocols are designed to avoid the loss of data. In this paper an overview of different broadcast protocols are discussed.
Sybil Attack Analysis and Detection Techniques in MANETrahulmonikasharma
Security is important for many sensor network applications. A particularly harmful attack against sensor and ad hoc networks is known as the Sybil attack [6], where a node Illegitimately claims multiple identities.Mobility cause a main problem when we talk about security in Mobile Ad-hoc networks. It doesn’t depend on fixed architecture, the nodes are continuously moving in a random fashion. In this article we will focus on identifying the Sybil attack in MANET. It uses air medium for communication so it is more prone to the attack. Sybil attack is one in which single node present multiple fake identities to other nodes, which cause destruction.
A Landmark Based Shortest Path Detection by Using A* and Haversine Formularahulmonikasharma
In 1900, less than 20 percent of the world populace lived in cities, in 2007, fair more than 50 percent of the world populace lived in cities. In 2050, it has been anticipated that more than 70 percent of the worldwide population (about 6.4 billion individuals) will be city tenants. There's more weight being set on cities through this increment in population [1]. With approach of keen cities, data and communication technology is progressively transforming the way city regions and city inhabitants organize and work in reaction to urban development. In this paper, we create a nonspecific plot for navigating a route throughout city A asked route is given by utilizing combination of A* Algorithm and Haversine equation. Haversine Equation gives least distance between any two focuses on spherical body by utilizing latitude and longitude. This least distance is at that point given to A* calculation to calculate minimum distance. The method for identifying the shortest path is specify in this paper.
Processing Over Encrypted Query Data In Internet of Things (IoTs) : CryptDBs,...rahulmonikasharma
Internet of Things (IoT) is the developing technologies that would be the biggest agents to modify the current world. Machine-to-machine communications perform with virtual, mobile and instantaneous connections. In IoT system, it consists of data-gathering sensors various other household devices. Intended for protecting IoT system, the end-to-end secure communication is a necessary measure to protect against unauthorized entities (e.g., modification attacks and eavesdropping,) and the data unprotected on the Cloud. The most important concern hereby is how to preserve the insightful information and to provide the privacy of user data. In IoT, the encrypted data computing is based on techniques appear to be promising approaches. In this paper, we discuss about the recent secure database systems, which are capable to execute SQL queries over encrypted data.
Quality Determination and Grading of Tomatoes using Raspberry Pirahulmonikasharma
In India cultivation of tomatoes is carried out by traditional methods and techniques. Today tremendous improvement in field of agriculture technologies and products can be seen. The tomatoes affect the overall production drastically. Image processing technique can be key technique for finding good qualities of tomatoes and grading. This work aimed to study different types of algorithms used for quality grading and sorting of fruit from the acquire image. In previous years several types of techniques are applied to analyses the good quality fruits. A simple system can be implemented using Raspberry pi with computer vision technology and image processing algorithms.
Comparative of Delay Tolerant Network Routings and Scheduling using Max-Weigh...rahulmonikasharma
Network management and Routing is supportively done by performing with the nodes, due to infrastructure-less nature of the network in Ad hoc networks or MANET. The nodes are maintained itself from the functioning of the network, for that reason the MANET security challenges several defects. Routing process and Scheduling is a significant idea to enhance the security in MANET. Other than, scheduling has been recognized to be a key issue for implementing throughput/capacity optimization in Ad hoc networks. Designed underneath conventional (LT) light tailed assumptions, traffic fundamentally faces Heavy-tailed (HT) assumption of the validity of scheduling algorithms. Scheduling policies are utilized for communication networks such as Max-Weight, backpressure and ACO, which are provably throughput optimality and the Pareto frontier of the feasible throughput region under maximal throughput vector. In wireless ad-hoc network, the issue of routing and optimal scheduling performs with time varying channel reliability and multiple traffic streams. Depending upon the security issues within MANETs in this paper presents a comparative analysis of existing scheduling policies based on their performance to progress the delay performance in most scenarios. The security issues of MANETs considered from this paper presents a relative analysis of existing scheduling policies depend on their performance to progress the delay performance in most developments.
DC Conductivity Study of Cadmium Sulfide Nanoparticlesrahulmonikasharma
The dc conductivity of consolidated nanoparticle of CdS has been studied over the temperature range from 303 K to 523 K and the conductivity has been found to be much larger than that of single crystals.
A Survey on Peak to Average Power Ratio Reduction Methods for LTE-OFDMrahulmonikasharma
OFDM (Orthogonal Frequency Division Multiplexing) is generally preferred for high data rate transmission in digital communication. The Long-Term Evolution (LTE) standards for the fourth generation (4G) wireless communication systems. Orthogonal Frequency Division Multiple Access (OFDMA) and Single Carrier Frequency Division Multiple Access (SC-FDMA) are the two multiple access techniques which are generally used in LTE.OFDM system has a major shortcoming of high peak to average power ratio (PAPR) value. This paper explains different PAPR reduction techniques and presents a comparison of the various techniques based on theoretical results. It also presents a survey of the various PAPR reduction techniques and the state of the art in this area.
IOT Based Home Appliance Control System, Location Tracking and Energy Monitoringrahulmonikasharma
Home automation has been a dream of sciences for so many years. It could wind up conceivable in twentieth century simply after power all family units and web administrations were begun being utilized on across the board level. The point of home robotization is to give enhanced accommodation, comfort, vitality effectiveness and security. Vitality checking and protection holds prime significance in this day and age in view of the irregularity between control age and request observing frameworks accessible in the market. Ordinarily, customers are disappointed with the power charge as it doesn't demonstrate the power devoured at the gadget level. This paper shows the outline and execution of a vitality meter utilizing Arduino microcontroller which can be utilized to gauge the power devoured by any individual electrical apparatus. The primary expectation of the proposed vitality meter is to screen the power utilization at the gadget level, transfer it to the server and build up remote control of any apparatus. So we can screen the power utilization remotely and close down gadgets if vital. The car segment is additionally one of the application spaces where vehicle can be made keen by utilizing "IOT". So a vehicle following framework is additionally executed to screen development of vehicles remotely.
Thermal Radiation and Viscous Dissipation Effects on an Oscillatory Heat and ...rahulmonikasharma
An anticipated outcome that is intended chapter is to investigate effects of magnetic field on an oscillatory flow of a viscoelastic fluid with thermal radiation, viscous dissipation with Ohmic heating which bounded by a vertical plane surface, have been studied. Analytical solutions for the quasi – linear hyperbolic partial differential equations are obtained by perturbation technique. Solutions for velocity and temperature distributions are discussed for various values of physical parameters involving in the problem. The effects of cooling and heating of a viscoelastic fluid compared to the Newtonian fluid have been discussed.
Advance Approach towards Key Feature Extraction Using Designed Filters on Dif...rahulmonikasharma
In fast growing database repository system, image as data is one of the important concern despite text or numeric. Still we can’t replace test on any cost but for advancement, information may be managed with images. Therefore image processing is a wide area for the researcher. Many stages of processing of image provide researchers with new ideas to keep information safe with better way. Feature extraction, segmentation, recognition are the key areas of the image processing which helps to enhance the quality of working with images. Paper presents the comparison between image formats like .jpg, .png, .bmp, .gif. This paper is focused on the feature extraction and segmentation stages with background removal process. There are two filters, one is integer filter and second one is floating point Filter, which is used for the key feature extraction from image. These filters applied on the different images of different formats and visually compare the results.
Alamouti-STBC based Channel Estimation Technique over MIMO OFDM Systemrahulmonikasharma
The examination on various looks into on MIMO STBC framework in order to accomplish the higher framework execution is standard that the execution of the remote correspondence frameworks can be improved by usage numerous transmit and get radio wires, that is normally gathered on the grounds that the MIMO procedure, and has been incorporated. The Alamouti STBC might be a promising because of notice the pick up inside the remote interchanges framework misuse MIMO. To broaden the code rate and furthermore the yield of the symmetrical zone time square code for more than 4 transmit reception apparatuses is examined. The outlined framework is beated once forced with M-PSK (i.e upto 32-PSK) regulation. The channel estimation examine in these conditions.
Empirical Mode Decomposition Based Signal Analysis of Gear Fault Diagnosisrahulmonikasharma
A vibration investigation is about the specialty of searching for changes in the vibration example, and after that relating those progressions back to the machines mechanical outline. The level of vibration and the example of the vibration reveal to us something about the interior state of the turning segment. The vibration example can let us know whether the machine is out of adjust or twisted. Al-so blames with the moving components and coupling issues can be distinguished. This paper shows an approach for equip blame investigation utilizing signal handling plans. The information has been taken from college of ohio, joined states. The investigation has done utilizing MATLAB software.
This paper discusses a new algorithm of a univariate method, which is vitally important to develop a short-term load forecasting module for planning and operation of distribution system. It has many applications including purchasing of energy, generation and infrastructure development etc. We have discussed different time series forecasting approaches in this paper. But ARIMA has proved itself as the most appropriate method in forecasting of the load profile for West Bengal using the historical data of the year of 2017. Auto Regressive Integrated Moving Average model gives more accuracy level of load forecast than any other techniques. Mean Absolute Percentage Error (MAPE) has been calculated for the mentioned forecasted model.
Impact of Coupling Coefficient on Coupled Line Couplerrahulmonikasharma
The coupled line coupler is a type of directional coupler which finds practical utility. It is mainly used for sampling the microwave power. In this paper, 3 couplers A,B & C are designed with different values of coupling coefficient 6dB,10dB & 18dB respectively at a frequency of 2.5GHz using ADS tool. The return loss, isolation loss & transmission loss are determined. The design & simulation is done using microstrip line technology.
Design Evaluation and Temperature Rise Test of Flameproof Induction Motorrahulmonikasharma
The ignition of flammable gases, vapours or dust in presence of oxygen contained in the surrounding atmosphere may lead to explosion. Flameproof three phase induction motors are the most common and frequently used in the process industries such as oil refineries, oil rigs, petrochemicals, fertilizers, etc. The design of flameproof motor is such that it allows and sustain explosion within the enclosure caused by ignition of hazardous gases without transmitting it to the external flammable atmosphere. The enclosure is mechanically strong enough to withstand the explosion pressure developed inside it. To prevent an explosion due to hot spot on the surface of the motor, flameproof induction motors are subjected to heat run test to determine the maximum surface temperature and temperature class with respect to the ignition temperature of the surrounding flammable gas atmosphere. This paper highlights the design features of flameproof motors and their surface temperature classification for different sizes.
Explore the innovative world of trenchless pipe repair with our comprehensive guide, "The Benefits and Techniques of Trenchless Pipe Repair." This document delves into the modern methods of repairing underground pipes without the need for extensive excavation, highlighting the numerous advantages and the latest techniques used in the industry.
Learn about the cost savings, reduced environmental impact, and minimal disruption associated with trenchless technology. Discover detailed explanations of popular techniques such as pipe bursting, cured-in-place pipe (CIPP) lining, and directional drilling. Understand how these methods can be applied to various types of infrastructure, from residential plumbing to large-scale municipal systems.
Ideal for homeowners, contractors, engineers, and anyone interested in modern plumbing solutions, this guide provides valuable insights into why trenchless pipe repair is becoming the preferred choice for pipe rehabilitation. Stay informed about the latest advancements and best practices in the field.
Welcome to WIPAC Monthly the magazine brought to you by the LinkedIn Group Water Industry Process Automation & Control.
In this month's edition, along with this month's industry news to celebrate the 13 years since the group was created we have articles including
A case study of the used of Advanced Process Control at the Wastewater Treatment works at Lleida in Spain
A look back on an article on smart wastewater networks in order to see how the industry has measured up in the interim around the adoption of Digital Transformation in the Water Industry.
Industrial Training at Shahjalal Fertilizer Company Limited (SFCL)MdTanvirMahtab2
This presentation is about the working procedure of Shahjalal Fertilizer Company Limited (SFCL). A Govt. owned Company of Bangladesh Chemical Industries Corporation under Ministry of Industries.
Hybrid optimization of pumped hydro system and solar- Engr. Abdul-Azeez.pdffxintegritypublishin
Advancements in technology unveil a myriad of electrical and electronic breakthroughs geared towards efficiently harnessing limited resources to meet human energy demands. The optimization of hybrid solar PV panels and pumped hydro energy supply systems plays a pivotal role in utilizing natural resources effectively. This initiative not only benefits humanity but also fosters environmental sustainability. The study investigated the design optimization of these hybrid systems, focusing on understanding solar radiation patterns, identifying geographical influences on solar radiation, formulating a mathematical model for system optimization, and determining the optimal configuration of PV panels and pumped hydro storage. Through a comparative analysis approach and eight weeks of data collection, the study addressed key research questions related to solar radiation patterns and optimal system design. The findings highlighted regions with heightened solar radiation levels, showcasing substantial potential for power generation and emphasizing the system's efficiency. Optimizing system design significantly boosted power generation, promoted renewable energy utilization, and enhanced energy storage capacity. The study underscored the benefits of optimizing hybrid solar PV panels and pumped hydro energy supply systems for sustainable energy usage. Optimizing the design of solar PV panels and pumped hydro energy supply systems as examined across diverse climatic conditions in a developing country, not only enhances power generation but also improves the integration of renewable energy sources and boosts energy storage capacities, particularly beneficial for less economically prosperous regions. Additionally, the study provides valuable insights for advancing energy research in economically viable areas. Recommendations included conducting site-specific assessments, utilizing advanced modeling tools, implementing regular maintenance protocols, and enhancing communication among system components.
Final project report on grocery store management system..pdfKamal Acharya
In today’s fast-changing business environment, it’s extremely important to be able to respond to client needs in the most effective and timely manner. If your customers wish to see your business online and have instant access to your products or services.
Online Grocery Store is an e-commerce website, which retails various grocery products. This project allows viewing various products available enables registered users to purchase desired products instantly using Paytm, UPI payment processor (Instant Pay) and also can place order by using Cash on Delivery (Pay Later) option. This project provides an easy access to Administrators and Managers to view orders placed using Pay Later and Instant Pay options.
In order to develop an e-commerce website, a number of Technologies must be studied and understood. These include multi-tiered architecture, server and client-side scripting techniques, implementation technologies, programming language (such as PHP, HTML, CSS, JavaScript) and MySQL relational databases. This is a project with the objective to develop a basic website where a consumer is provided with a shopping cart website and also to know about the technologies used to develop such a website.
This document will discuss each of the underlying technologies to create and implement an e- commerce website.
Immunizing Image Classifiers Against Localized Adversary Attacksgerogepatton
This paper addresses the vulnerability of deep learning models, particularly convolutional neural networks
(CNN)s, to adversarial attacks and presents a proactive training technique designed to counter them. We
introduce a novel volumization algorithm, which transforms 2D images into 3D volumetric representations.
When combined with 3D convolution and deep curriculum learning optimization (CLO), itsignificantly improves
the immunity of models against localized universal attacks by up to 40%. We evaluate our proposed approach
using contemporary CNN architectures and the modified Canadian Institute for Advanced Research (CIFAR-10
and CIFAR-100) and ImageNet Large Scale Visual Recognition Challenge (ILSVRC12) datasets, showcasing
accuracy improvements over previous techniques. The results indicate that the combination of the volumetric
input and curriculum learning holds significant promise for mitigating adversarial attacks without necessitating
adversary training.
Prediction of Default Customer in Banking Sector using Artificial Neural Network
1. International Journal on Recent and Innovation Trends in Computing and Communication ISSN: 2321-8169
Volume: 5 Issue: 7 293 – 296
_______________________________________________________________________________________________
293
IJRITCC | July 2017, Available @ http://www.ijritcc.org
_______________________________________________________________________________________
Prediction of Default Customer in Banking Sector using Artificial Neural
Network
W.S.Kanmani
Research Scholar PG and Research Department of Computer
Science, Dr.Ambethkar Government Arts College,
Chennai – 600 039, India.
e-mail: kanmanimalliga@gmail.com
B.Jayapradha
Assistant Professor PG and Research Department of Computer
Science, Dr.Ambedar Government Arts College,
Chennai – 600 039, India.
e-mail:jayapradha_bhaskar@rediffmail.com
Abstract— The aim of this article is to present perdition and risk accuracy analysis of default customer in the banking sector. The neural network
is a learning model inspired by biological neuron it is used to estimate and predict that can depend on a large number of inputs. The bank
customer dataset from UCI repository, used for data analysis method to extract informative data set from a large volume of the dataset. This
dataset is used in the neural network for training data and testing data. In a training of data, the data set is iterated till the desired output. This
training data is cross check with test data. This paper focuses on predicting default customer by using deep learning neural network (DNN)
algorithm.
Keywords-Neural Network, Deep learning neural network(DNN), Rapid Miner, DefaultCustomer
__________________________________________________*****_________________________________________________
I. INTRODUCTION
In recent research, works evolve around with neural
network. So the neural network has enormous developing in
every field because Neural Network learns to do the task by
observational data without task specification programs. It
typically has a number of neuron and connections those are
divided by layers. Deep learning Neural Network (DNN) is a
powerful technique for a neural network and it provides the
best solution for many problems. At present in the field of
business, customer relation management’s first step is to
create a customer retention analysis model. Most of the
businesses fields hope to figure out the real cause of lost a
customer, or even to be told that they are about to lose the
customer by some clues before it occurs, and then they can
propose or make some new sales strategies against the loss in
advance [7].Banks have realized that customer relations are a
very important factor for their success. Constants increase in
customers, increase the volume of data stored in banking
database. Those data may not provide sufficient information
so the bank faces a challenge to find the default customer and
how to retain most profitable customers. Data analysis
techniques are resourceful for extracting the needed
information [3]. This paper focuses on predicting the default
customer from customer data set in banking sector using
Deep Learning Neural Network with help of Rapid Miner
studio software package which provides integrated solution
for neural network and data analysis. The hypothesis was that
customer who often uses bank transaction are more loyal and
the bank should focus on a customer who uses the least
transaction and make them turn to be a profitable customer.
II. LITERATURE REVIEW
Burez and Van den Poel indicate that there are two
types of targeted approaches to managing customer churn:
reactive and proactive. When a company adopts a reactive
approach, it waits until customers ask the company to cancel
their service relationship. On the other hand, when a
company adopts a proactive approach, it tries to identify
customers who are likely to default before they do so.
Targeted proactive programs have potential advantages of
having lower incentive costs [4]. In the financial area, the
multilayer perceptron (MLP) network trained by the back-
propagation learning algorithm is the most used technique for
financial decision-making problems [10]. In existing system
propose a method to solve such a problem, and according to
the method, to find out behavior patterns of losing customers
or clues before they stop using some products through
Mining Sequential Patterns [7]. In existing paper Multilayer
perceptron used to find default customer. It has two different
approaches based on neural networks for identifying
important variables. The first is based on error change and
the second is based on weights contribution in the network
[5]. Another existing work proposed back propagation feed
forward to predict customers. It also provides a solution for
overtraining problem that occurs in the neural networks[6].
III. RELATED WORK
According to research literature about default customer,
most of the related work focuses on using only one data
mining method such as classification or clustering to mine
the customer retention data. From the review of the
literature, we conclude that neural network can predict
default customer in different domain like cellular network
2. International Journal on Recent and Innovation Trends in Computing and Communication ISSN: 2321-8169
Volume: 5 Issue: 7 293 – 296
_______________________________________________________________________________________________
294
IJRITCC | July 2017, Available @ http://www.ijritcc.org
_______________________________________________________________________________________
[6], pay TV [4], telecom [5], credit risk analysis [8], and
bankrupt [9].
This paper proposed deep neural network based approach
to predict the default customer in banking and following
hypotheses question are consider for predictions
• How much transaction does the customer perform per
month?
• What is the occupation of the customer?
• What is a profile for the high-risk borrower?
• How much Balance maintain each month?
• What is the private status of a customer?
IV. METHODOLOGY
a. Rapid Miner
Rapid Miner is a commercial data science software
platform developed by the company of the same name
that provides an integrated environment for machine
learning, deep learning, text mining, and predictive
analytics. It is used for business and commercial
applications as well as for research, education, training,
rapid prototyping, and application development and
supports all steps of the machine learning process
including data preparation, results in visualization,
validation, and optimization.
b. Data Analysis
The used data set contain a large number of data
collection which consists of noisy, irrelevant, missing
values. We extract need data set by using one of data
analysis method, this data set which can apply in neural
network training. The extracted data set is 3846 based on
the client age, job, day and month which most frequent.
c. Deep Learning Neural Network
Deep learning is part of machine learning, is an
application to learn tasks of artificial neural network
(ANN). It allows computational models that are
composed of multiple processing layers to learn
representations of data with multiple levels of
abstraction. These methods have dramatically improved
in many domains like speech recognition, visual object
recognition, object detection and other domains [12]. It
is very best for supervised learning. DNN has common
issues are overfitting and computation time this is due to
naively training of data [13]. These can be overcome
with well train data sets.
V. RESULT ANALYSIS
The paper used rapid miner software package which
supports neural network to detect default customer. After
selecting a data set, the software goes to data analysis phase,
we are extracting need data and defining target output field
which applied in a neural network for prediction.
In network training phase we are selecting number
hidden layers and number of training cycles. The software
package offers the best model which we can edit. Figure.1
shows our work uses three hidden layers with 9(8 input+1
thresholds) input neuron and two output results.
Figure 1. Network Topology in Rapid Miner
In figure 2 shows the deep learning neural network training
and testing phase with parameter setting.
Figure 2. Deep learning in rapid miner
After training complete we get the results that shown in
Table 1. With overall Accuracy: 78.18%
Table 1.Performance Accuracy Table
True no True yes
Class
Precision
Pred.No 2775 305 90.10%
Pred.Yes 534 231 30.20%
Class recall 83.86% 43.10%
Table 1. shows the Overall accuracy of actual prediction
achieves by with three hidden layer models with six neurons
is roughly trained with the result of 78.18%. the dataset is
divided into 10 sets of the model and one testing model.
each learning is iterated 10 times with one testing models on
each set. The prediction made rough of 1000 example after
being trained on nearly about 50 is nearly 78%. +/-1.09% is
standard deviation calculated from 10 individual model
3. International Journal on Recent and Innovation Trends in Computing and Communication ISSN: 2321-8169
Volume: 5 Issue: 7 293 – 296
_______________________________________________________________________________________________
295
IJRITCC | July 2017, Available @ http://www.ijritcc.org
_______________________________________________________________________________________
accuracy errors on 10 models which are trained. The class
precision shows the work of correct prediction for each
label. Figure 3 shows the chart result.
Figure 3. Prediction Results
Confusion matrix calculation is shown in Table 2 with
respected bar chart result shown in Figure 3.
Table 2. Confusion matrix table
Predict. No Predict. Yes Error
Actual. No 2846 463 0.1399
Actual.
Yes
284 252 0.5299
Totals 3130 715 0.1943
In Table 2. shows the confusion matrix for the actual and
prediction with error mean values for each prediction. The
total outcome of 3130 data the error is 0.19% its shows the
efficient in reducing error for better prediction results
Figure 4. Confusion Matrix Chart With Error Mean
VI. DISCUSSION
This work concludes that there is ʹa problematic groupʹ of
people (self-employed, entrepreneur) with less transaction in
a month also have credit balances, who in the future can
become very important and very valuable clients. The bank
should focus on offering services to these clients. For
example, Bank could introduce new scheme tailored to self-
employed and entrepreneur needs such as car loan with
insurance, favorable interest rates, promotional use of
internet banking, etc. Also, the DNN topology of the neural
network gives the better results.
VII. CONCLUSION AND FUTURE ENHANCEMENT
Building an effective default customer prediction model
using various techniques has become a significant topic for
business and academics in recent years. Default prediction
and management is a curial task in the banking sector. In
order to compete in the financial field, the bank has to be
able to predict possible defaults and take proactive actions to
retain the valuable loyal customer. The results show that The
customer who uses bank often are more loyal and who uses
bank very less turns to be defaulters. It also shows the
accuracy rate of the defaulter. In future work, several issues
can be considered and it can be implemented along with
neural network several other popular prediction techniques
can be applied in combination such as vector machine,
machine learning to develop hybrid models.
References
[1] AlisaBilalZorić Interdisciplinary Description of Complex
Systems ―Predicting Customer Churn In Banking Industry
Using Neural Networks‖ vol 14(2), pages116-124, 10
March 2016.
[2] Sumathi, S. and Sivanandam, S.N.: Introduction to Data
Mining Principles. Studies in Computational Intelligence
29(3), 1-20 2013,
[3] Domingo, R.: Applying data mining to banking.
Http://www.rtdonline.com, accessed 18th November 2015,
[4] Burez, J., & Van den Poel, D. (2007). CRM at a pay-TV
company: Using analytical models to reduce customer
attrition by targeted marketing for subscription services.
Expert Systems with Applications, 32, 277–288.
[5] Omar Adwan, Hossam Faris, Khalid Jaradat, Osama
Harfoushi and Nazeeh Ghatasheh. Predicting Customer
Churn in Telecom Industry using MLP Neural Networks:
Modeling and Analysis. Life Sci J 2014;11(3):75-81
[6] Chiang, D., Wang, Y., Lee, S., & Lin, C. (2003). Goal-
oriented sequential pattern for network banking churn
analysis. Expert Systems with Applications, 25(3),293–302.
[7] Nazari, M. and Alidadi, M.: Measuring credit risk of bank
customers using artificial neural network. Journal of
Management Research 5(2), 17-27, 2013,
[8] Salvatore Madonna, Prof,Greta Cestari, PhD THE
ACCURACY OF BANKRUPTCY PREDICTION
MODELS: A COMPARATIVE ANALYSIS OF
MULTIVARIATE DISCRIMINANT MODELS IN THE
ITALIAN CONTEXT European Scientific Journal
0
500
1000
1500
2000
2500
3000
True No True Yes class prection
prediction
actual
pred.No
pred.Yes
Class recall
0
500
1000
1500
2000
2500
3000
No Yes Error
Prediction
Actual
No
Yes
4. International Journal on Recent and Innovation Trends in Computing and Communication ISSN: 2321-8169
Volume: 5 Issue: 7 293 – 296
_______________________________________________________________________________________________
296
IJRITCC | July 2017, Available @ http://www.ijritcc.org
_______________________________________________________________________________________
December 2015 edition vol.11, No.34 ISSN: 1857 – 7881
(Print) e - ISSN 1857- 7431.
[9] Tsai, C.F. and Wu, J.W.: Using neural network ensembles
for bankruptcy prediction and credit scoring. Expert
Systems with Applications 34(4), 2639-2649, 2008.
[10] Wilson, R.L. and Sharda, R.: Bankruptcy prediction using
neural networks. Decision Support Systems 11(5), 545-557,
1994.
[11] Deep Learning. Yann LeCun, Yoshua Bengio & Geoffrey
Hinton. Nature 521, 436–444 (28 May 2015)
[12] Ivakhnenko, Alexey (1971). "Polynomial theory of
complex systems". IEEE Transactions on Systems, Man
and Cybernetics (4): 364–378.
[13] Mohsen Nazari , Mojtaba Alidadi ―Measuring Credit Risk
of Bank Customers Using Artificial Neural Network‖
Journal of Management Research, ISSN 1941-899X,2013,
Vol. 5, No. 2.
[14] Bar, M.V.: The Computational Intelligence Techniques For
Predictions-Artificial Neural Networks. Annals of
Computational Economics 2(42), 184-190, 2014
[15] Anuj Sharma, Dr.Prabin Kumar Panigrahi. A Neural
Network Based Approach for Predicting Customer Churn
in Cellular Network Services.International Journal of
Computer Applications (0975 – 8887)Volume 27– No.11,
August 2011M. Young, The Technical Writer’s Handbook.
Mill Valley, CA: University Science, 19.