The document outlines a paper on using manifold learning techniques for credit risk analysis. It discusses motivations for dimensionality reduction and bankruptcy prediction. The proposed approach uses Isomap and Supervised Isomap for nonlinear dimensionality reduction to better analyze financial data and predict credit risk. Experimental results are presented on a data set to evaluate the methodology. Conclusions and potential future work are discussed.
Despeckling of Ultrasound Imaging using Median Regularized Coupled PdeIDES Editor
This paper presents an approach for reducing speckle
in ultrasound images using Coupled Partial Differential
Equation (CPDE) which has been obtained by uniting secondorder
and the fourth-order partial differential equations. Using
PDE to reduce the speckle is the noise-smoothing methods
which is getting attention widely, because PDE can keep the
edge well when it reduces the noise. We also introduced a
median regulator to guide energy source to boost the features
in the image and regularize the diffusion. The proposed
method is tested in both simulated and real medical
ultrasound images. The proposed method is compared with
SRAD, Perona Malik diffusion and Non linear coherent
diffusion methods, our method gives better result in terms of
CNR, SSIM and FOM.
Mutual Information for Registration of Monomodal Brain Images using Modified ...IDES Editor
Image registration has great significance in medicine,
with a lot of techniques anticipated in it. This research work
implies an approach for medical image registration that
registers images of the mono modalities for CT or MRI images
using Modified Adaptive Polar Transform (MAPT). The
performance of the Adaptive Polar Transform (APT) with the
proposed technique is examined. The results prove that MAPT
performs better than APT technique. The proposed scheme not
only reduces the source of errors and also reduces the elapsed
time for registration of brain images. An analysis is presented
for the medical image processing on mutual- information-based
registration.
One-Sample Face Recognition Using HMM Model of Fiducial AreasCSCJournals
In most real world applications, multiple image samples of individuals are not easy to collate for direct implementation of recognition or verification systems. Therefore there is a need to perform these tasks even if only one training sample per person is available. This paper describes an effective algorithm for recognition and verification with one sample image per class. It uses two dimensional discrete wavelet transform (2D DWT) to extract features from images and hidden Markov model (HMM) was used for training, recognition and classification. It was tested with a subset of the AT&T database and up to 90% correct classification (Hit) and false acceptance rate (FAR) of 0.02% was achieved.
Research Inventy : International Journal of Engineering and Science is published by the group of young academic and industrial researchers with 12 Issues per year. It is an online as well as print version open access journal that provides rapid publication (monthly) of articles in all areas of the subject such as: civil, mechanical, chemical, electronic and computer engineering as well as production and information technology. The Journal welcomes the submission of manuscripts that meet the general criteria of significance and scientific excellence. Papers will be published by rapid process within 20 days after acceptance and peer review process takes only 7 days. All articles published in Research Inventy will be peer-reviewed.
Despeckling of Ultrasound Imaging using Median Regularized Coupled PdeIDES Editor
This paper presents an approach for reducing speckle
in ultrasound images using Coupled Partial Differential
Equation (CPDE) which has been obtained by uniting secondorder
and the fourth-order partial differential equations. Using
PDE to reduce the speckle is the noise-smoothing methods
which is getting attention widely, because PDE can keep the
edge well when it reduces the noise. We also introduced a
median regulator to guide energy source to boost the features
in the image and regularize the diffusion. The proposed
method is tested in both simulated and real medical
ultrasound images. The proposed method is compared with
SRAD, Perona Malik diffusion and Non linear coherent
diffusion methods, our method gives better result in terms of
CNR, SSIM and FOM.
Mutual Information for Registration of Monomodal Brain Images using Modified ...IDES Editor
Image registration has great significance in medicine,
with a lot of techniques anticipated in it. This research work
implies an approach for medical image registration that
registers images of the mono modalities for CT or MRI images
using Modified Adaptive Polar Transform (MAPT). The
performance of the Adaptive Polar Transform (APT) with the
proposed technique is examined. The results prove that MAPT
performs better than APT technique. The proposed scheme not
only reduces the source of errors and also reduces the elapsed
time for registration of brain images. An analysis is presented
for the medical image processing on mutual- information-based
registration.
One-Sample Face Recognition Using HMM Model of Fiducial AreasCSCJournals
In most real world applications, multiple image samples of individuals are not easy to collate for direct implementation of recognition or verification systems. Therefore there is a need to perform these tasks even if only one training sample per person is available. This paper describes an effective algorithm for recognition and verification with one sample image per class. It uses two dimensional discrete wavelet transform (2D DWT) to extract features from images and hidden Markov model (HMM) was used for training, recognition and classification. It was tested with a subset of the AT&T database and up to 90% correct classification (Hit) and false acceptance rate (FAR) of 0.02% was achieved.
Research Inventy : International Journal of Engineering and Science is published by the group of young academic and industrial researchers with 12 Issues per year. It is an online as well as print version open access journal that provides rapid publication (monthly) of articles in all areas of the subject such as: civil, mechanical, chemical, electronic and computer engineering as well as production and information technology. The Journal welcomes the submission of manuscripts that meet the general criteria of significance and scientific excellence. Papers will be published by rapid process within 20 days after acceptance and peer review process takes only 7 days. All articles published in Research Inventy will be peer-reviewed.
Image Compression based on DCT and BPSO for MRI and Standard ImagesIJERA Editor
Nowadays, digital image compression has become a crucial factor of modern telecommunication systems. Image compression is the process of reducing total bits required to represent an image by reducing redundancies while preserving the image quality as much as possible. Various applications including internet, multimedia, satellite imaging, medical imaging uses image compression in order to store and transmit images in an efficient manner. Selection of compression technique is an application-specific process. In this paper, an improved compression technique based on Butterfly-Particle Swarm Optimization (BPSO) is proposed. BPSO is an intelligence-based iterative algorithm utilized for finding optimal solution from a set of possible values. The dominant factors of BPSO over other optimization techniques are higher convergence rate, searching ability and overall performance. The proposed technique divides the input image into 88 blocks. Discrete Cosine Transform (DCT) is applied to each block to obtain the coefficients. Then, the threshold values are obtained from BPSO. Based on this threshold, values of the coefficients are modified. Finally, quantization followed by the Huffman encoding is used to encode the image. Experimental results show the effectiveness of the proposed method over the existing method.
Fuzzy clustering Approach in segmentation of T1-T2 brain MRIIDES Editor
Segmentation is a difficult and challenging
problem in the magnetic resonance images, and it
considered as important in computer vision and artificial
intelligence. Many researchers have applied various
techniques however fuzzy c-means (FCM) based
algorithms is more effective compared to other methods.
In this paper, we present a novel FCM algorithm for
weighted bias (also called intensity in-homogeneities)
estimation and segmentation of MRI. Normally, the
intensity inhomogeneities are attributed to imperfections
in the radio-frequency coils or to the problems associated
with the image acquisition. Our algorithm is formulated
by modifying the objective function of the standard FCM
and it has the advantage that it can be applied at an early
stage in an automated data analysis. Further this paper
proposes a center knowledge method in order to reduce
the running time of proposed algorithm. The proposed
method can deal with the intensity in-homogeneities and
image noise effectively. We have compared our results
with other reported methods. The results using real MRI
data show that our method provides better results
compared to standard FCM based algorithms and other
modified FCM-based techniques.
Uniform and non uniform single image deblurring based on sparse representatio...ijma
Considering the sparseness property of images, a sparse representation based iterative deblurring method
is presented for single image deblurring under uniform and non-uniform motion blur. The approach taken
is based on sparse and redundant representations over adaptively training dictionaries from single
blurred-noisy image itself. Further, the K-SVD algorithm is used to obtain a dictionary that describes the
image contents effectively. Comprehensive experimental evaluation demonstrate that the proposed
framework integrating the sparseness property of images, adaptive dictionary training and iterative
deblurring scheme together significantly improves the deblurring performance and is comparable with the
state-of-the art deblurring algorithms and seeks a powerful solution to an ill-conditioned inverse problem.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
An Information Maximization approach of ICA for Gender ClassificationIDES Editor
In this paper, a novel and successful method for
gender classification from human faces using dimensionality
reduction technique is proposed. Independent Component
Analysis (ICA) is one of such techniques. In the current
scheme, a thrust is given on the different algorithms and
architectures of ICA. An information maximization ICA is
discussed with its two architecture and compared with the two
architectures of fast ICA. Support Vector Machine (SVM) is
used as a classifier for the separation of male and female
classes. All experiments are done on FERET database. Results
are obtained for the different combinations of train and test
database sizes. For larger
training set SVM is performing with an accuracy of 98%. The
accuracy values are varied for change in size of testing set and
the proposed system performs with an average accuracy of
96%. An improvement in performance is achieved using class
discriminability which performs with 100% accuracy.
Gesture Recognition using Principle Component Analysis & Viola-Jones AlgorithmIJMER
Gesture recognition pertains to recognizing meaningful expressions of motion by a human,
involving the hands, arms, face, head, and/or body. It is of utmost importance in designing an intelligent
and efficient human–computer interface. The applications of gesture recognition are manifold, ranging
from sign language through medical rehabilitation to virtual reality. In this paper, we provide a survey on
gesture recognition with particular emphasis on hand gestures and facial expressions. Applications
involving wavelet transform and principal component analysis for face and hand gesture recognition on
digital images
VIDEO SEGMENTATION & SUMMARIZATION USING MODIFIED GENETIC ALGORITHMijcsa
Video summarization of the segmented video is an essential process for video thumbnails, video surveillance and video downloading. Summarization deals with extracting few frames from each scene and creating a summary video which explains all course of action of full video with in short duration of time. The proposed research work discusses about the segmentation and summarization of the frames. A genetic algorithm (GA) for segmentation and summarization is required to view the highlight of an event by selecting few important frames required. The GA is modified to select only key frames for summarization and the comparison of modified GA is done with the GA.
On Training Targets and Objective Functions for Deep-Learning-Based Audio-Vis...Daniel Michelsanti
Audio-visual speech enhancement (AV-SE) is the task of improving speech quality and intelligibility in a noisy environment using audio and visual information from a talker. Recently, deep learning techniques have been adopted to solve the AV-SE task in a supervised manner. In this context, the choice of the target, i.e. the quantity to be estimated, and the objective function, which quantifies the quality of this estimate, to be used for training is critical for the performance. This work is the first that presents an experimental study of a range of different targets and objective functions used to train a deeplearning-based AV-SE system. The results show that the approaches that directly estimate a mask perform the best overall in terms of estimated speech quality and intelligibility, although the model that directly estimates the log magnitude spectrum performs as good in terms of estimated speech quality.
https://ieeexplore.ieee.org/document/8682790
Extended Fuzzy Hyperline Segment Neural Network for Fingerprint RecognitionCSCJournals
In this paper we have proposed Extended Fuzzy Hyperline Segment Neural Network (EFHLSNN) and its learning algorithm which is an extension of Fuzzy Hyperline Segment Neural Network (FHLSNN). The fuzzy set hyperline segment is an n-dimensional hyperline segment defined by two end points with a corresponding extended membership function. The fingerprint feature extraction process is based on FingerCode feature extraction technique. The performance of EFHLSNN is verified using POLY U HRF fingerprint database. The EFHLSNN is found superior compared to FHLSNN in generalization, training and recall time.
A CONCERT EVALUATION OF EXEMPLAR BASED IMAGE INPAINTING ALGORITHMS FOR NATURA...cscpconf
Image inpainting derives from restoration of art works, and has been applied to repair ancient
art works. Inpainting is a technique of restoring a partially damaged or occluded image in an
undetectable way. It fills the damaged part of an image by employing information of the
undamaged part according to some rules to make it look “reasonable” to human eyes. Digital
image inpainting is relatively new area of research, but numerous and different approaches to
tackle the inpainting problem have been proposed since the concept was first introduced. This
paper analyzes and compares the recent exemplar based inpainting algorithms by Minqin Wang
and Hao Guo et al. A number of examples on real images are demonstrated to evaluate the
results of algorithms using Peak Signal to Noise Ratio (PSNR)
Highly Adaptive Image Restoration In Compressive Sensing Applications Using S...IJARIDEA Journal
Abstract— Image Restoration is the operation of taking a degenerate picture and assessing the perfect, unique picture. Intially the range is separated from caught scene and coordinated with the word reference and are stacked together. At last the pictures are reestablished utilizing SDL calculation. The PSNR qualities are observe to be higher than customary condition of all pressure procedures. The point of word reference learning is to finding an edge in which some preparation information concedes an inadequate portrayal. In this strategy the specimens are taken underneath the Nyquist rate. In any case, in specific cases a lexicon that is prepared to fit the information can essentially enhance the sparsity, which has applications in information disintegration.
Keywords— FSIM , Group based Sparse Representation, PSNR, Sparse Dictionary Learning.
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
machine learning in the age of big data: new approaches and business applicat...Armando Vieira
Presentation at University of Lisbon on Machine Learning and big data.
Deep learning algorithms and applications to credit risk analysis, churn detection and recommendation algorithms
Image Compression based on DCT and BPSO for MRI and Standard ImagesIJERA Editor
Nowadays, digital image compression has become a crucial factor of modern telecommunication systems. Image compression is the process of reducing total bits required to represent an image by reducing redundancies while preserving the image quality as much as possible. Various applications including internet, multimedia, satellite imaging, medical imaging uses image compression in order to store and transmit images in an efficient manner. Selection of compression technique is an application-specific process. In this paper, an improved compression technique based on Butterfly-Particle Swarm Optimization (BPSO) is proposed. BPSO is an intelligence-based iterative algorithm utilized for finding optimal solution from a set of possible values. The dominant factors of BPSO over other optimization techniques are higher convergence rate, searching ability and overall performance. The proposed technique divides the input image into 88 blocks. Discrete Cosine Transform (DCT) is applied to each block to obtain the coefficients. Then, the threshold values are obtained from BPSO. Based on this threshold, values of the coefficients are modified. Finally, quantization followed by the Huffman encoding is used to encode the image. Experimental results show the effectiveness of the proposed method over the existing method.
Fuzzy clustering Approach in segmentation of T1-T2 brain MRIIDES Editor
Segmentation is a difficult and challenging
problem in the magnetic resonance images, and it
considered as important in computer vision and artificial
intelligence. Many researchers have applied various
techniques however fuzzy c-means (FCM) based
algorithms is more effective compared to other methods.
In this paper, we present a novel FCM algorithm for
weighted bias (also called intensity in-homogeneities)
estimation and segmentation of MRI. Normally, the
intensity inhomogeneities are attributed to imperfections
in the radio-frequency coils or to the problems associated
with the image acquisition. Our algorithm is formulated
by modifying the objective function of the standard FCM
and it has the advantage that it can be applied at an early
stage in an automated data analysis. Further this paper
proposes a center knowledge method in order to reduce
the running time of proposed algorithm. The proposed
method can deal with the intensity in-homogeneities and
image noise effectively. We have compared our results
with other reported methods. The results using real MRI
data show that our method provides better results
compared to standard FCM based algorithms and other
modified FCM-based techniques.
Uniform and non uniform single image deblurring based on sparse representatio...ijma
Considering the sparseness property of images, a sparse representation based iterative deblurring method
is presented for single image deblurring under uniform and non-uniform motion blur. The approach taken
is based on sparse and redundant representations over adaptively training dictionaries from single
blurred-noisy image itself. Further, the K-SVD algorithm is used to obtain a dictionary that describes the
image contents effectively. Comprehensive experimental evaluation demonstrate that the proposed
framework integrating the sparseness property of images, adaptive dictionary training and iterative
deblurring scheme together significantly improves the deblurring performance and is comparable with the
state-of-the art deblurring algorithms and seeks a powerful solution to an ill-conditioned inverse problem.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
An Information Maximization approach of ICA for Gender ClassificationIDES Editor
In this paper, a novel and successful method for
gender classification from human faces using dimensionality
reduction technique is proposed. Independent Component
Analysis (ICA) is one of such techniques. In the current
scheme, a thrust is given on the different algorithms and
architectures of ICA. An information maximization ICA is
discussed with its two architecture and compared with the two
architectures of fast ICA. Support Vector Machine (SVM) is
used as a classifier for the separation of male and female
classes. All experiments are done on FERET database. Results
are obtained for the different combinations of train and test
database sizes. For larger
training set SVM is performing with an accuracy of 98%. The
accuracy values are varied for change in size of testing set and
the proposed system performs with an average accuracy of
96%. An improvement in performance is achieved using class
discriminability which performs with 100% accuracy.
Gesture Recognition using Principle Component Analysis & Viola-Jones AlgorithmIJMER
Gesture recognition pertains to recognizing meaningful expressions of motion by a human,
involving the hands, arms, face, head, and/or body. It is of utmost importance in designing an intelligent
and efficient human–computer interface. The applications of gesture recognition are manifold, ranging
from sign language through medical rehabilitation to virtual reality. In this paper, we provide a survey on
gesture recognition with particular emphasis on hand gestures and facial expressions. Applications
involving wavelet transform and principal component analysis for face and hand gesture recognition on
digital images
VIDEO SEGMENTATION & SUMMARIZATION USING MODIFIED GENETIC ALGORITHMijcsa
Video summarization of the segmented video is an essential process for video thumbnails, video surveillance and video downloading. Summarization deals with extracting few frames from each scene and creating a summary video which explains all course of action of full video with in short duration of time. The proposed research work discusses about the segmentation and summarization of the frames. A genetic algorithm (GA) for segmentation and summarization is required to view the highlight of an event by selecting few important frames required. The GA is modified to select only key frames for summarization and the comparison of modified GA is done with the GA.
On Training Targets and Objective Functions for Deep-Learning-Based Audio-Vis...Daniel Michelsanti
Audio-visual speech enhancement (AV-SE) is the task of improving speech quality and intelligibility in a noisy environment using audio and visual information from a talker. Recently, deep learning techniques have been adopted to solve the AV-SE task in a supervised manner. In this context, the choice of the target, i.e. the quantity to be estimated, and the objective function, which quantifies the quality of this estimate, to be used for training is critical for the performance. This work is the first that presents an experimental study of a range of different targets and objective functions used to train a deeplearning-based AV-SE system. The results show that the approaches that directly estimate a mask perform the best overall in terms of estimated speech quality and intelligibility, although the model that directly estimates the log magnitude spectrum performs as good in terms of estimated speech quality.
https://ieeexplore.ieee.org/document/8682790
Extended Fuzzy Hyperline Segment Neural Network for Fingerprint RecognitionCSCJournals
In this paper we have proposed Extended Fuzzy Hyperline Segment Neural Network (EFHLSNN) and its learning algorithm which is an extension of Fuzzy Hyperline Segment Neural Network (FHLSNN). The fuzzy set hyperline segment is an n-dimensional hyperline segment defined by two end points with a corresponding extended membership function. The fingerprint feature extraction process is based on FingerCode feature extraction technique. The performance of EFHLSNN is verified using POLY U HRF fingerprint database. The EFHLSNN is found superior compared to FHLSNN in generalization, training and recall time.
A CONCERT EVALUATION OF EXEMPLAR BASED IMAGE INPAINTING ALGORITHMS FOR NATURA...cscpconf
Image inpainting derives from restoration of art works, and has been applied to repair ancient
art works. Inpainting is a technique of restoring a partially damaged or occluded image in an
undetectable way. It fills the damaged part of an image by employing information of the
undamaged part according to some rules to make it look “reasonable” to human eyes. Digital
image inpainting is relatively new area of research, but numerous and different approaches to
tackle the inpainting problem have been proposed since the concept was first introduced. This
paper analyzes and compares the recent exemplar based inpainting algorithms by Minqin Wang
and Hao Guo et al. A number of examples on real images are demonstrated to evaluate the
results of algorithms using Peak Signal to Noise Ratio (PSNR)
Highly Adaptive Image Restoration In Compressive Sensing Applications Using S...IJARIDEA Journal
Abstract— Image Restoration is the operation of taking a degenerate picture and assessing the perfect, unique picture. Intially the range is separated from caught scene and coordinated with the word reference and are stacked together. At last the pictures are reestablished utilizing SDL calculation. The PSNR qualities are observe to be higher than customary condition of all pressure procedures. The point of word reference learning is to finding an edge in which some preparation information concedes an inadequate portrayal. In this strategy the specimens are taken underneath the Nyquist rate. In any case, in specific cases a lexicon that is prepared to fit the information can essentially enhance the sparsity, which has applications in information disintegration.
Keywords— FSIM , Group based Sparse Representation, PSNR, Sparse Dictionary Learning.
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
machine learning in the age of big data: new approaches and business applicat...Armando Vieira
Presentation at University of Lisbon on Machine Learning and big data.
Deep learning algorithms and applications to credit risk analysis, churn detection and recommendation algorithms
Deep Neural Networks (DNN), or simply Deep Learning (DL), took Artificial Intelligence (AI) by storm and have infiltrated into business at an unprecedented rate. Access to vast amounts of data, recently made available by the Big Data revolution, extensive computational power and a new wave of efficient learning algorithms, helped Artificial Neural Networks to achieve state-of-the-art results in almost all AI challenges.
The implications of DL supported AI in business is tremendous, shaking to the foundations many industries. However, incorporating this technology in established business is far from obvious: cultural inertia in organizations, lack of transparency in most DL models and the complexity in training these models are some of the issues that will be addressed.
ABSTRACT: In the field of computer science known as "machine learning," a computer makes predictions about
the tasks it will perform next by examining the data that has been given to it. The computer can access data via
interacting with the environment or by using digitalized training sets. In contrast to static programming
algorithms, which require explicit human guidance, machine learning algorithms may learn from data and
generate predictions on their own. Various supervised and unsupervised strategies, including rule-based
techniques, logic-based techniques, instance-based techniques, and stochastic techniques, have been presented in
order to solve problems. Our paper's main goal is to present a comprehensive comparison of various cutting-edge
supervised machine learning techniques.
Multimodal Biometrics Recognition by Dimensionality Diminution MethodIJERA Editor
Multimodal biometric system utilizes two or more character modalities, e.g., face, ear, and fingerprint,
Signature, plamprint to improve the recognition accuracy of conventional unimodal methods. We propose a new
dimensionality reduction method called Dimension Diminish Projection (DDP) in this paper. DDP can not only
preserve local information by capturing the intra-modal geometry, but also extract between-class relevant
structures for classification effectively. Experimental results show that our proposed method performs better
than other algorithms including PCA, LDA and MFA.
In this paper, an attempt has been made to extract texture
features from facial images using an improved method of
Illumination Invariant Feature Descriptor. The proposed local
ternary Pattern based feature extractor viz., Steady Illumination
Local Ternary Pattern (SIcLTP) has been used to extract texture
features from Indian face database. The similarity matching
between two extracted feature sets has been obtained using Zero
Mean Sum of Squared Differences (ZSSD). The RGB facial images
are first converted into the YIQ colour space to reduce the
redundancy of the RGB images. The result obtained has been
analysed using Receiver Operating Characteristic curve, and is
found to be promising. Finally the results are validated with
standard local binary pattern (LBP) extractor.
Particle Swarm Optimization in the fine-tuning of Fuzzy Software Cost Estimat...Waqas Tariq
Software cost estimation deals with the financial and strategic planning of software projects. Controlling the expensive investment of software development effectively is of paramount importance. The limitation of algorithmic effort prediction models is their inability to cope with uncertainties and imprecision surrounding software projects at the early development stage. More recently, attention has turned to a variety of machine learning methods, and soft computing in particular to predict software development effort. Fuzzy logic is one such technique which can cope with uncertainties. In the present paper, Particle Swarm Optimization Algorithm (PSOA) is presented to fine tune the fuzzy estimate for the development of software projects . The efficacy of the developed models is tested on 10 NASA software projects, 18 NASA projects and COCOMO 81 database on the basis of various criterion for assessment of software cost estimation models. Comparison of all the models is done and it is found that the developed models provide better estimation
Function Point Software Cost Estimates using Neuro-Fuzzy techniqueijceronline
Software estimation accuracy is among the greatest challenges for software developers. As Neurofuzzy based system is able to approximate the non-linear function with more precision so it is used as a soft computing approach to generate model by formulating the relationship based on its training. The approach presented in this paper is independent of the nature and type of estimation. In this paper, Function point is used as algorithmic model and an attempt is being made to validate the soundness of Neuro fuzzy technique using ISBSG and NASA project data.
5 ijaems sept-2015-9-video feature extraction based on modified lle using ada...INFOGAIN PUBLICATION
Locally linear embedding (LLE) is an unsupervised learning algorithm which computes the low dimensional, neighborhood preserving embeddings of high dimensional data. LLE attempts to discover non-linear structure in high dimensional data by exploiting the local symmetries of linear reconstructions. In this paper, video feature extraction is done using modified LLE alongwith adaptive nearest neighbor approach to find the nearest neighbor and the connected components. The proposed feature extraction method is applied to a video. The video feature description gives a new tool for analysis of video.
International Journal of Engineering and Science Invention (IJESI) is an international journal intended for professionals and researchers in all fields of computer science and electronics. IJESI publishes research articles and reviews within the whole field Engineering Science and Technology, new teaching methods, assessment, validation and the impact of new technologies and it will continue to provide information on the latest trends and developments in this ever-expanding subject. The publications of papers are selected through double peer reviewed to ensure originality, relevance, and readability. The articles published in our journal can be accessed online.
ESTIMATING PROJECT DEVELOPMENT EFFORT USING CLUSTERED REGRESSION APPROACHcscpconf
Due to the intangible nature of “software”, accurate and reliable software effort estimation is a challenge in the software Industry. It is unlikely to expect very accurate estimates of software
development effort because of the inherent uncertainty in software development projects and the complex and dynamic interaction of factors that impact software development. Heterogeneity exists in the software engineering datasets because data is made available from diverse sources.
This can be reduced by defining certain relationship between the data values by classifying them into different clusters. This study focuses on how the combination of clustering and
regression techniques can reduce the potential problems in effectiveness of predictive efficiency due to heterogeneity of the data. Using a clustered approach creates the subsets of data having a degree of homogeneity that enhances prediction accuracy. It was also observed in this study that ridge regression performs better than other regression techniques used in the analysis.
Similar to Manifold learning for credit risk assessment (20)
Predicting online user behaviour using deep learning algorithmsArmando Vieira
We propose a robust classifier to predict buying intentions based on user behaviour within a large e-commerce website. In this work we compare traditional machine learning techniques with the most advanced deep learning approaches. We show that both Deep Belief Networks and Stacked Denoising auto-Encoders achieved a substantial improvement by extracting features from high dimensional data during the pre-train phase. They prove also to be more convenient to deal with severe class imbalance.
We propose an algorithm for training Multi Layer Preceptrons for classification problems, that we named Hidden Layer Learning Vector Quantization (H-LVQ). It consists of applying Learning Vector Quantization to the last hidden layer of a MLP and it gave very successful results on problems containing a large number of correlated inputs. It was applied with excellent results on classification of Rurtherford
backscattering spectra and on a benchmark problem of image recognition. It may also be used for efficient feature extraction.
1. Outline
Motivation
Dimensionality Reduction
Proposed approach
Experimental setup
Conclusions and Future Work
Bankrupcy Analysis for Credit Risk
using Manifold Learning
B Ribeiro, A Vieira, J Duarte, C Silva, J Carvalho das Neves,
University of Coimbra, ISEP and ISEG, Portugal
and
Q Liu, A H Sung
New Mexico Tech, USA
November, 2008
ICONIP 2008
2. Outline
Motivation
Dimensionality Reduction
Proposed approach
Experimental setup
Conclusions and Future Work
1 Motivation
2 Dimensionality Reduction
Manifold Learning
Isomap
Supervised Isomap
3 Proposed approach
Overview
Operation
4 Experimental setup
Data set
Evaluation metrics
Results
5 Conclusions and Future Work
ICONIP 2008
3. Outline
Motivation
Dimensionality Reduction
Proposed approach
Experimental setup
Conclusions and Future Work
Credit Risk Analysis
Predicting bankruptcy has been a very important topic in
accounting and finance attracting considerable research both
from academic and business areas
The question of how to determine the credit-worthiness of a
customer or how safe is to grant credit remains a main
concern for banks and investors, particularly, with the recent
financial crisis
ICONIP 2008
4. Outline
Motivation
Dimensionality Reduction
Proposed approach
Experimental setup
Conclusions and Future Work
Importance of Risk (1)
ICONIP 2008
5. Outline
Motivation
Dimensionality Reduction
Proposed approach
Experimental setup
Conclusions and Future Work
Importance of Risk (2)
ICONIP 2008
6. Outline
Motivation
Dimensionality Reduction
Proposed approach
Experimental setup
Conclusions and Future Work
Problem definition
The problem of bankruptcy prediction can be addressed as
follows:
Given a set of financial ratios describing the situation of a
company over a given period, predict the probability that this
company may become bankrupted in a near future, normally during
the following year
ICONIP 2008
7. Outline
Motivation
Manifold Learning
Dimensionality Reduction
Isomap
Proposed approach
Supervised Isomap
Experimental setup
Conclusions and Future Work
Objectives of dimensionality reduction
Nonlinear dimensionality reduction permits severe reduction
on the feature space
A direct consequence of nonlinear dimension reduction is the
visualization of data which can help to reveal the data
structures
Aims at choosing from the available set of features, a smaller
set that more efficiently represents the data
Supervised or unsupervised
Supervised methods use the label of the training examples in
the reduction step and usually perform better
ICONIP 2008
8. Outline
Motivation
Manifold Learning
Dimensionality Reduction
Isomap
Proposed approach
Supervised Isomap
Experimental setup
Conclusions and Future Work
Introduction
Emerging technique that estimates a low-dimensional
structure, embedded in high-dimensional data
The underpinning idea is to invert a generative model for a
given set of observations
Manifold learning can be used as a pre-processing technique
to tackle the curse of dimensionality
ICONIP 2008
9. Outline
Motivation
Manifold Learning
Dimensionality Reduction
Isomap
Proposed approach
Supervised Isomap
Experimental setup
Conclusions and Future Work
Formulation
Given data points x1 , x2 , · · · , xn ∈ IRD , we assume that the
data lies on a d-dimensional M manifold embedded into IRD ,
where d < D
A manifold M can be described by a single coordinate chart
f : M −→ IRd . The manifold learning consists of finding
y1 , · · · yn ∈ IRd , where yi = f (xi ).
ICONIP 2008
10. Outline
Motivation
Manifold Learning
Dimensionality Reduction
Isomap
Proposed approach
Supervised Isomap
Experimental setup
Conclusions and Future Work
Isomap Algorithm
1 Estimates which points are neighbors on the manifold M,
based on the distances dX (i, j) between pairs of points i, j in
the input space X by computing the weighted graph G of
neighborhood relations given by the edges of weight dX (i, j).
2 Estimates the geodesic distances between all pairs of data
points in the manifold M by computing the shortest path
distance on the k’s nearest neighbor graph built on the data
set.
3 Applies classical MDS to the matrix of graph distances
DG = {dG (i, j)}, constructing an embedding of the data in a
d-dimensional Euclidean space Y that best preserves the
manifolds estimated intrinsic geometry
ICONIP 2008
11. Outline
Motivation
Manifold Learning
Dimensionality Reduction
Isomap
Proposed approach
Supervised Isomap
Experimental setup
Conclusions and Future Work
Analysis
Isomap assumes that there is an isometric chart that preserves
distances between points.
If xi and xj are two points in the manifold M embedded into
IRD and the geodesic distance between them is dG (xi , xj ) ,
then there is a chart f : M −→ IRd such that
||f (xi ) − f (xj )|| = dG (xi , xj )
For nearby points in the high-dimensional space the Euclidean
distance is a good approximation of the geodesic distance
whereas for distant points this is not true
ICONIP 2008
12. Outline
Motivation
Manifold Learning
Dimensionality Reduction
Isomap
Proposed approach
Supervised Isomap
Experimental setup
Conclusions and Future Work
Image Processing Example
[J. Tenenbaum, de Silva, & Langford, 2000]
ICONIP 2008
13. Outline
Motivation
Manifold Learning
Dimensionality Reduction
Isomap
Proposed approach
Supervised Isomap
Experimental setup
Conclusions and Future Work
Analysis
A weighted graph with k’s nearest neighbors is built where its
edges are weighted by the Euclidean distances between nearby
data points
Then a shortest path computation algorithm such as,
Dijkstra’s or Floyd’s, will complete the calculus of the
remainder geodesic distances.
MDS is then used to estimate the points whose Euclidean
distance equal the geodesic distances. Given a matrix
D ∈ IRn×n of dissimilarities, MDS constructs a set of points
whose interpoint Euclidean distances match those in D closely.
ICONIP 2008
14. Outline
Motivation
Manifold Learning
Dimensionality Reduction
Isomap
Proposed approach
Supervised Isomap
Experimental setup
Conclusions and Future Work
Supervised version
The training labels are used to refine the distances between
inputs, since both classification and visualization can benefit
when the inter-class dissimilarity is larger than the intra-class
dissimilarity
The mapping function given by Isomap is only implicitly
defined and nonlinear interpolation techniques, such as GRNN
have to be used to learn it
This can also make the algorithm overfit the training set and
can often make the neighborhood graph of the input data
disconnected
ICONIP 2008
15. Outline
Motivation
Manifold Learning
Dimensionality Reduction
Isomap
Proposed approach
Supervised Isomap
Experimental setup
Conclusions and Future Work
Determining distances
The Euclidean distance dij = d(xi , xj ) between two given
observations xi and xj , labeled ci and cj respectively, is
replaced by a dissimilarity measure:
((a − 1)/a)1/2 if ci = cj
D(xi , xj ) = (1)
a1/2 − d0 if ci = cj
2
where a = 1/e −dij /σ with dij set to one of the distance measures
described above, σ is a smoothing parameter (set according to the
data ’density’), do is a constant (0 ≤ d0 ≤ 1) and ci , cj are the
data class labels.
ICONIP 2008
17. Outline
Motivation
Dimensionality Reduction Overview
Proposed approach Overview
Experimental setup
Conclusions and Future Work
Testing instances
When a reduced space is reached, our aim is to learn a
kernel-based model that can be applied for testing new cases
of failed and non-failed firms
For testing, however, Isomap does not provide an explicit
mapping in the embedded mapping. Therefore we can not
generate the test set directly, since we would need to use the
labels
We use a generalized regression neural network (GRNN) to
learn the mapping, before the SVM prediction phase takes
place
ICONIP 2008
18. Outline
Motivation
Data set
Dimensionality Reduction
Evaluation metrics
Proposed approach
Results
Experimental setup
Conclusions and Future Work
Diane database
Financial statements of French companies, initially of 60,000
industrial French companies, for the years of 2002 to 2006,
with at least 10 employees
3,000 were declared bankrupted in 2007 or presented a
restructuring plan
30 financial ratios which allow the description of firms in
terms of the financial strength, liquidity, solvability,
productivity of labor and capital, margins, net profitability and
return on investment
ICONIP 2008
19. Outline
Motivation
Data set
Dimensionality Reduction
Evaluation metrics
Proposed approach
Results
Experimental setup
Conclusions and Future Work
Financial ratios
1. Number of employees 2. Financial Debt/Capital Employed %
3. Capital Employed/Fixed Assets 4. Depreciation of Tangible Assets
5. Working capital/current assets 6. Current ratio
7. Liquidity ratio 8. Stock Turnover days
9. Collection period 10. Credit Period
11. Turnover per Employee 12. Interest/Turnover
13. Debt Period days 14. Financial Debt/Equity
15. Financial Debt/Cashflow 16. Cashflow/Turnover
17. Working Capital/Turnover (days) 18. Net Current Assets/Turnover (days)
19. Working Capital Needs/Turnover 20. Export
21. Value added per employee 22. Total Assets/Turnover
23. Operating Profit Margin 24. Net Profit Margin
25. Added Value Margin 26. Part of Employees
27. Return on Capital Employed 28. Return on Total Assets
29. EBIT Margin 30. EBITDA Margin
ICONIP 2008
20. Outline
Motivation
Data set
Dimensionality Reduction
Evaluation metrics
Proposed approach
Results
Experimental setup
Conclusions and Future Work
Preprocessing
Many cases with missing values, especially for defaults
companies
Default cases sorted out by the number of missing values.
Examples with 10 missing values at most were considered
600 default examples was obtained
To balance the dataset we selected randomly 600 non-default
examples
ICONIP 2008
21. Outline
Motivation
Data set
Dimensionality Reduction
Evaluation metrics
Proposed approach
Results
Experimental setup
Conclusions and Future Work
Preprocessing
For the ratios of the years 2003 and 2006, each missing value
was replaced by the closest available year value
For 2004 and 2005, if values of the next and previous years
were available, each missing value was replaced by their mean,
otherwise it was replaced by the remaining value
In some cases there was no data available for a ratio in any of
the years. In this very few cases the missing data was replaced
by the median value of the ratio in each year
All ratios were logarithmized and then standardized to zero
mean and unity variance
ICONIP 2008
22. Outline
Motivation
Data set
Dimensionality Reduction
Evaluation metrics
Proposed approach
Results
Experimental setup
Conclusions and Future Work
Historical data
Companies are often subjected to fluctuation of the market,
economy cycles and unavoidable contingencies related to its
business activity
Yearly variations of important financial ratios reflecting the
balance sheet, sometimes quite relevant, are common
particularly for small companies
We included information from the past 3 years preceding the
default. The number of inputs is therefore increased from 30
to 90 ratios
More relevant than the ratios themselves, are the variations
that occur over the period range of the analysis.
ICONIP 2008
23. Outline
Motivation
Data set
Dimensionality Reduction
Evaluation metrics
Proposed approach
Results
Experimental setup
Conclusions and Future Work
Contingency table and error measures
Class Positive Class Negative
Assigned Positive tp fp
(True Positives) (False Positives)
Assigned Negative fn tn
(False Negatives) (True Negatives)
tp tp
Recall ( tp+fn ) and Precision ( tp+fp )
fp
Error type I ( fp+tn ) - % of companies classified as bankrupt
when in reality they are healthy
fn
Error type II ( fn+tp ) - % number of samples classified as
healthy when they are observed to be bankrupt
fp+fn
Error Rate - tp+fp+fn+tn )
ICONIP 2008
24. Outline
Motivation
Data set
Dimensionality Reduction
Evaluation metrics
Proposed approach
Results
Experimental setup
Conclusions and Future Work
Trustworthiness
A projection is trustworthy if the set of the k nearest
neighbors of each data point in the low-dimensional space are
also close-by in the original space:
N
2
M(k) = 1 − (r (i, j) − k), (2)
Nk(2N − 3k − 1)
i=1 j∈Uk (i)
where r (i, j) is the rank of the data point j in the ordering
according to the distance from i in the original data space, and
Uk (i) denotes the set of those data points that are among the
k-nearest neighbors of the data point i in the low-dimensional
space but not in the original space.
ICONIP 2008
25. Outline
Motivation
Data set
Dimensionality Reduction
Evaluation metrics
Proposed approach
Results
Experimental setup
Conclusions and Future Work
Visualization
Trustworthiness with S-Isomap
0.95
nldr=3
nldr=5
nldr=10
0.9
Trustworthiness
0.85
0.8
0.75
0.7
3 4 5 7 10 15 20 40 60 80 100 150 200
K Nearest Neighbors
ICONIP 2008
28. Outline
Motivation
Data set
Dimensionality Reduction
Evaluation metrics
Proposed approach
Results
Experimental setup
Conclusions and Future Work
S-Isomap with Euclidean distance - KNN and SVM
SVM Testing Accuracy with S-Isomap
nldr=3
94 nldr=5
nldr=10
92
Accuracy in %
90
88
86
3 4 5 7 10 15 20 40 60 80 100 150 200
K Nearest Neighbors
ICONIP 2008
29. Outline
Motivation
Data set
Dimensionality Reduction
Evaluation metrics
Proposed approach
Results
Experimental setup
Conclusions and Future Work
Discussion of Results
S-Isomap presents better results in testing accuracy than
single KNN and RVM by 2% and 10%
S-isomaps presents comparable results with SVM, however,
with much reduced embedded space (nldr=3) whereas SVM
algorithm is used with all financial ratios
The error of type II, corresponding to a failure of the correct
prediction of bankruptcy is lower for the SVM.
The same happens with a false alarm, i.e., indicating a
bankruptcy for a healthy firm, which corresponds to the error
of type I.
The fact that firms clump nicely in the reduced space not only
enhances financial data visualization but also improves
prediction results as compared with the kernel machines.
ICONIP 2008
30. Outline
Motivation
Dimensionality Reduction
Proposed approach
Experimental setup
Conclusions and Future Work
Conclusions and Future Work
We proposed an approach for bankruptcy analysis and
prediction based on a supervised Isomap algorithm where class
label information is incorporated
Assuming that corporate financial statuses lie in a manifold
we attempt to uncover this embedded structure using
manifold learning
Isomap acts as a preprocessing stage allowing financial data
visualization
Results have shown that comparable testing accuracy can be
obtained even using a 3-dimensional reduced space
Although the results in the finance setting seem promising,
further work is necessary to design a method for avoiding the
interpolation error resulting from the mapping learning stage.
ICONIP 2008
31. Outline
Motivation
Dimensionality Reduction
Proposed approach
Experimental setup
Conclusions and Future Work
Bankrupcy Analysis for Credit Risk
using Manifold Learning
B Ribeiro, A Vieira, J Duarte, C Silva, J Carvalho das Neves,
University of Coimbra, ISEP and ISEG, Portugal
and
Q Liu, A H Sung
New Mexico Tech, USA
November, 2008
ICONIP 2008