This paper demonstrates how cubic Spline (B-Spline) models can be used to recognize 2-dimension nonrigid handwritten isolated characters. Each handwritten character is represented by a set of nonoverlapping uniformly distributed landmarks. The Spline models are constructed by utilizing cubic order of
polynomial to model the shapes under study. The approach is a two-stage process. The first stage is
learning, we construct a mixture of spline class parameters to capture the variations in spline coefficients
using the apparatus Expectation Maximization algorithm. The second stage is recognition, here we use the
Fréchet distance to compute the variations between the spline models and test spline shape for recognition.
We test the approach on a set of handwritten Arabic letters.
MIXTURES OF TRAINED REGRESSION CURVES MODELS FOR HANDWRITTEN ARABIC CHARACTER...gerogepatton
In this paper, we demonstrate how regression curves can be used to recognize 2D non-rigid handwritten shapes. Each shape is represented by a set of non-overlapping uniformly distributed landmarks. The underlying models utilize 2nd order of polynomials to model shapes within a training set. To estimate the regression models, we need to extract the required coefficients which describe the variations for a set of shape class. Hence, a least square method is used to estimate such modes. We proceed then, by training these coefficients using the apparatus Expectation Maximization algorithm. Recognition is carried out by finding the least error landmarks displacement with respect to the model curves. Handwritten isolated Arabic characters are used to evaluate our approach.
MIXTURES OF TRAINED REGRESSION CURVESMODELS FOR HANDRITTEN ARABIC CHARACTER R...ijaia
In this paper, we demonstrate how regression curves can be used to recognize 2D non-rigid handwritten shapes. Each shape is represented by a set of non-overlapping uniformly distributed landmarks. The underlying models utilize 2nd order of polynomials to model shapes within a training set. To estimate the regression models, we need to extract the required coefficients which describe the variations for a set of shape class. Hence, a least square method is used to estimate such modes. We proceed then, by training these coefficients using the apparatus Expectation Maximization algorithm. Recognition is carried out by finding the least error landmarks displacement with respect to the model curves. Handwritten isolated Arabic characters are used to evaluate our approach.
ARABIC HANDWRITTEN CHARACTER RECOGNITION USING STRUCTURAL SHAPE DECOMPOSITIONsipij
This document summarizes a statistical framework for recognizing 2D shapes represented as arrangements of curves or strokes. It presents a hierarchical model with two layers: 1) At the lower level, each curve is represented by a point distribution model describing its shape variability. Curves are assigned stroke labels. 2) At the top level, shapes are represented by the geometric arrangement of curve center points (using another point distribution model) and a string of stroke labels. Recognition involves assigning stroke labels to curves, and recovering stroke and shape deformation parameters using expectation maximization. The method is applied to Arabic handwritten character recognition.
HANDWRITTEN CHARACTER RECOGNITION USING STRUCTURAL SHAPE DECOMPOSITIONcsandit
This paper presents a statistical framework for recognising 2D shapes which are represented as
an arrangement of curves or strokes. The approach is a hierarchical one which mixes geometric
and symbolic information in a three-layer architecture. Each curve primitive is represented
using a point-distribution model which describes how its shape varies over a set of training
data. We assign stroke labels to the primitives and these indicate to which class they belong.
Shapes are decomposed into an arrangement of primitives and the global shape representation
has two components. The first of these is a second point distribution model that is used to
represent the geometric arrangement of the curve centre-points. The second component is a
string of stroke labels that represents the symbolic arrangement of strokes. Hence each shape
can be represented by a set of centre-point deformation parameters and a dictionary of
permissible stroke label configurations. The hierarchy is a two-level architecture in which the
curve models reside at the nonterminal lower level of the tree. The top level represents the curve
arrangements allowed by the dictionary of permissible stroke combinations. The aim in
recognition is to minimise the cross entropy between the probability distributions for geometric
alignment errors and curve label errors. We show how the stroke parameters, shape-alignment
parameters and stroke labels may be recovered by applying the expectation maximization EM
algorithm to the utility measure. We apply the resulting shape-recognition method to Arabic
character recognition.
AUTOMATIC TRAINING DATA SYNTHESIS FOR HANDWRITING RECOGNITION USING THE STRUC...ijaia
The paper presents a novel technique called “Structural Crossing-Over” to synthesize qualified data for training machine learning-based handwriting recognition. The proposed technique can provide a greater variety of patterns of training data than the existing approaches such as elastic distortion and tangentbased affine transformation. A couple of training characters are chosen, then they are analyzed by their similar and different structures, and finally are crossed over to generate the new characters. The experiments are set to compare the performances of tangent-based affine transformation and the proposed approach in terms of the variety of generated characters and percent of recognition errors. The standard
MNIST corpus including 60,000 training characters and 10,000 test characters is employed in the
experiments. The proposed technique uses 1,000 characters to synthesize 60,000 characters, and then uses
these data to train and test the benchmark handwriting recognition system that exploits Histogram of
Gradient: HOG as features and Support Vector Machine: SVM as recognizer. The experimental result
yields 8.06% of errors. It significantly outperforms the tangent-based affine transformation and the
original MNIST training data, which are 11.74% and 16.55%, respectively.
Data entry is a time consuming and erroneous procedure in its nature. In addition, validity
check of submitted information is not easier than retyping it. In a mega-corporation like Kanoon
Farhangi Amoozesh, there are almost no way to control the authenticity of students' educational
background. By the virtue of fast computer architectures, optical character recognition, a.k.a.
OCR, systems have become viable. Unfortunately, general-purpose OCR systems like Google's
Tesseract are not handful because they don't have any a-priori information about what they are
reading. In this paper the authors have taken a in-depth look on what has done in the field of
OCR in the last 60 years. Then, a custom-made system adapted to the problem is presented
which is way more accurate than general purpose OCRs. The developed system reads more than
60 digits per second. As shown in the Results section, the accuracy of the devised method is
reasonable enough to be exposed in public use.
This document presents a novel approach for measuring shape similarity and using it for object recognition. The key steps are:
1) Solving the correspondence problem between two shapes by attaching a descriptor called "shape context" to sample points on each shape. Shape context captures the distribution of remaining points relative to the reference point.
2) Using the point correspondences to estimate an aligning transformation between the shapes. This provides a measure of shape similarity as the matching error between corresponding points plus the magnitude of the transformation.
3) Treating recognition as a nearest neighbor problem to find the most similar stored prototype shape. The approach is demonstrated on various datasets including handwritten digits, silhouettes, and 3D objects
This document discusses using Constrained Local Models (CLM) for facial landmark localization and applications in pose estimation and facial expression classification. It describes:
1) How CLM models are trained using local image descriptors for each landmark region rather than a holistic texture model, and how the Subspace Constrained Mean Shift algorithm is used to fit the CLM to new images.
2) How the trained CLM can be applied to estimate head pose by inferring orientation from localized facial landmarks.
3) How facial expressions are classified based on detected Action Units (facial muscle movements) from CLM-localized landmarks.
Evaluation results are presented for the CLM landmark localization, pose estimation, and expression
MIXTURES OF TRAINED REGRESSION CURVES MODELS FOR HANDWRITTEN ARABIC CHARACTER...gerogepatton
In this paper, we demonstrate how regression curves can be used to recognize 2D non-rigid handwritten shapes. Each shape is represented by a set of non-overlapping uniformly distributed landmarks. The underlying models utilize 2nd order of polynomials to model shapes within a training set. To estimate the regression models, we need to extract the required coefficients which describe the variations for a set of shape class. Hence, a least square method is used to estimate such modes. We proceed then, by training these coefficients using the apparatus Expectation Maximization algorithm. Recognition is carried out by finding the least error landmarks displacement with respect to the model curves. Handwritten isolated Arabic characters are used to evaluate our approach.
MIXTURES OF TRAINED REGRESSION CURVESMODELS FOR HANDRITTEN ARABIC CHARACTER R...ijaia
In this paper, we demonstrate how regression curves can be used to recognize 2D non-rigid handwritten shapes. Each shape is represented by a set of non-overlapping uniformly distributed landmarks. The underlying models utilize 2nd order of polynomials to model shapes within a training set. To estimate the regression models, we need to extract the required coefficients which describe the variations for a set of shape class. Hence, a least square method is used to estimate such modes. We proceed then, by training these coefficients using the apparatus Expectation Maximization algorithm. Recognition is carried out by finding the least error landmarks displacement with respect to the model curves. Handwritten isolated Arabic characters are used to evaluate our approach.
ARABIC HANDWRITTEN CHARACTER RECOGNITION USING STRUCTURAL SHAPE DECOMPOSITIONsipij
This document summarizes a statistical framework for recognizing 2D shapes represented as arrangements of curves or strokes. It presents a hierarchical model with two layers: 1) At the lower level, each curve is represented by a point distribution model describing its shape variability. Curves are assigned stroke labels. 2) At the top level, shapes are represented by the geometric arrangement of curve center points (using another point distribution model) and a string of stroke labels. Recognition involves assigning stroke labels to curves, and recovering stroke and shape deformation parameters using expectation maximization. The method is applied to Arabic handwritten character recognition.
HANDWRITTEN CHARACTER RECOGNITION USING STRUCTURAL SHAPE DECOMPOSITIONcsandit
This paper presents a statistical framework for recognising 2D shapes which are represented as
an arrangement of curves or strokes. The approach is a hierarchical one which mixes geometric
and symbolic information in a three-layer architecture. Each curve primitive is represented
using a point-distribution model which describes how its shape varies over a set of training
data. We assign stroke labels to the primitives and these indicate to which class they belong.
Shapes are decomposed into an arrangement of primitives and the global shape representation
has two components. The first of these is a second point distribution model that is used to
represent the geometric arrangement of the curve centre-points. The second component is a
string of stroke labels that represents the symbolic arrangement of strokes. Hence each shape
can be represented by a set of centre-point deformation parameters and a dictionary of
permissible stroke label configurations. The hierarchy is a two-level architecture in which the
curve models reside at the nonterminal lower level of the tree. The top level represents the curve
arrangements allowed by the dictionary of permissible stroke combinations. The aim in
recognition is to minimise the cross entropy between the probability distributions for geometric
alignment errors and curve label errors. We show how the stroke parameters, shape-alignment
parameters and stroke labels may be recovered by applying the expectation maximization EM
algorithm to the utility measure. We apply the resulting shape-recognition method to Arabic
character recognition.
AUTOMATIC TRAINING DATA SYNTHESIS FOR HANDWRITING RECOGNITION USING THE STRUC...ijaia
The paper presents a novel technique called “Structural Crossing-Over” to synthesize qualified data for training machine learning-based handwriting recognition. The proposed technique can provide a greater variety of patterns of training data than the existing approaches such as elastic distortion and tangentbased affine transformation. A couple of training characters are chosen, then they are analyzed by their similar and different structures, and finally are crossed over to generate the new characters. The experiments are set to compare the performances of tangent-based affine transformation and the proposed approach in terms of the variety of generated characters and percent of recognition errors. The standard
MNIST corpus including 60,000 training characters and 10,000 test characters is employed in the
experiments. The proposed technique uses 1,000 characters to synthesize 60,000 characters, and then uses
these data to train and test the benchmark handwriting recognition system that exploits Histogram of
Gradient: HOG as features and Support Vector Machine: SVM as recognizer. The experimental result
yields 8.06% of errors. It significantly outperforms the tangent-based affine transformation and the
original MNIST training data, which are 11.74% and 16.55%, respectively.
Data entry is a time consuming and erroneous procedure in its nature. In addition, validity
check of submitted information is not easier than retyping it. In a mega-corporation like Kanoon
Farhangi Amoozesh, there are almost no way to control the authenticity of students' educational
background. By the virtue of fast computer architectures, optical character recognition, a.k.a.
OCR, systems have become viable. Unfortunately, general-purpose OCR systems like Google's
Tesseract are not handful because they don't have any a-priori information about what they are
reading. In this paper the authors have taken a in-depth look on what has done in the field of
OCR in the last 60 years. Then, a custom-made system adapted to the problem is presented
which is way more accurate than general purpose OCRs. The developed system reads more than
60 digits per second. As shown in the Results section, the accuracy of the devised method is
reasonable enough to be exposed in public use.
This document presents a novel approach for measuring shape similarity and using it for object recognition. The key steps are:
1) Solving the correspondence problem between two shapes by attaching a descriptor called "shape context" to sample points on each shape. Shape context captures the distribution of remaining points relative to the reference point.
2) Using the point correspondences to estimate an aligning transformation between the shapes. This provides a measure of shape similarity as the matching error between corresponding points plus the magnitude of the transformation.
3) Treating recognition as a nearest neighbor problem to find the most similar stored prototype shape. The approach is demonstrated on various datasets including handwritten digits, silhouettes, and 3D objects
This document discusses using Constrained Local Models (CLM) for facial landmark localization and applications in pose estimation and facial expression classification. It describes:
1) How CLM models are trained using local image descriptors for each landmark region rather than a holistic texture model, and how the Subspace Constrained Mean Shift algorithm is used to fit the CLM to new images.
2) How the trained CLM can be applied to estimate head pose by inferring orientation from localized facial landmarks.
3) How facial expressions are classified based on detected Action Units (facial muscle movements) from CLM-localized landmarks.
Evaluation results are presented for the CLM landmark localization, pose estimation, and expression
A NOVEL FEATURE SET FOR RECOGNITION OF SIMILAR SHAPED HANDWRITTEN HINDI CHARA...cscpconf
This document describes a study that uses machine learning algorithms to recognize similar shaped handwritten Hindi characters. It extracts 85 features from each character image based on geometry. Four machine learning algorithms (Bayesian Network, RBFN, MLP, C4.5 Decision Tree) are trained on datasets containing samples of a target character pair and evaluated based on precision, misclassification rate, and model build time. Feature selection techniques are also used to reduce the feature set dimensionality before classification. Experimental results show that different algorithms perform best depending on the feature set and number of samples used for training.
Proposal of a similarity measure for unified modeling language class diagram ...IJECEIAES
The unified modeling language (UML) represents an essential tool for modeling and visualizing software systems. UML diagrams provide a graphical representation of a system's components. Comparing and processing these diagrams, for instance, can be complicated, especially as software projects grow in size and complexity. In such contexts, deep learning techniques have emerged as a promising solution for solving complex problems. One of these crucial problems is the measurement of similarity between images, making it possible to compare and calculate the differences between two given diagrams. The present work intends to build a method for calculating the degree of similarity between two UML class diagrams. With a goal to provide teachers a helpful tool for assessing students' UML class diagrams.
CHECKING BEHAVIOURAL COMPATIBILITY IN SERVICE COMPOSITION WITH GRAPH TRANSFOR...csandit
The success of Service Oriented Architecture (SOA) largely depended on the success of
automatic service composition. Dynamic service selection process should ensure full
compatibility between the services involved in the composition. This compatibility must be both
on static proprieties, called interface compatibility which can be easily proved and especially
on behavioural compatibility that needs composability checking of basic services. In this paper,
we propose (1) a formalism for modelling composite services using an extension of the Business
Process (BP) modelling approach proposed by Benatallah et al. and (2) a formal verification
approach of service composition. This approach uses the Graph Transformation (GT)
methodology as a formal verification tool. It allows behavioural compatibility verification of
two given services modelled by their BPs, used as the source graph in the GT operation. The
idea consists of (1) trying to dynamically generate a graph grammar R (a set of transformation
rules) whose application generates the composite service, if it exists, in this case (2) the next
step consist in checking the deadlock free in the resulting composite service. To this end we
propose an algorithm that we have implemented using the AGG, an algebraic graph
transformation API environment under eclipse IDE.
Statistical learning theory provides a framework for machine learning that draws from statistics and functional analysis. It deals with finding predictive functions based on data. Supervised learning involves learning from input-output pairs in a training data set to infer the function that maps inputs to outputs. The goals of learning are understanding relationships and enabling prediction. Statistical learning in R can perform regression for continuous outputs or classification for discrete outputs. It is commonly used in fields like computer vision, speech recognition, and bioinformatics.
FRACTAL ANALYSIS OF GOOD PROGRAMMING STYLEcscpconf
This paper studies a new, quantitative approach using fractal geometry to analyze basic tenets of good programming style. Experiments on C source of the GNU/Linux Core Utilities, a
collection of 114 programs or approximately 70,000 lines of code, show systematic changes in style are correlated with statistically significant changes in fractal dimension (P≤0.0009). The data further show positive but weak correlation between lines of code and fractal dimension (r=0.0878). These results suggest the fractal dimension is a reliable metric of changes that
affect good style, the knowledge of which may be useful for maintaining a code base.
This paper studies a new, quantitative approach using fractal geometry to analyse basic tenets
of good programming style. Experiments on C source of the GNU/Linux Core Utilities, a
collection of 114 programs or approximately 70,000 lines of code, show systematic changes in
style are correlated with statistically significant changes in fractal dimension (P≤0.0009). The
data further show positive but weak correlation between lines of code and fractal dimension
(r=0.0878). These results suggest the fractal dimension is a reliable metric of changes that
affect good style, the knowledge of which may be useful for maintaining a code base.
Object recognition is the challenging problem in the real world application. Object recognition can be achieved through the shape matching. Shape matching is preceded by i) detecting the edges of the objects from the images. ii) Finding the correspondence between the shapes. iii) Measuring the dissimilarity between the shapes using the correspondence. iv) Classifying the object into classes by using this dissimilarity measures. In order to solve the correspondence problem, we attach a descriptor, the shape context, to each point. The shape context at a reference point captures the distribution of the remaining points relative to it, thus offering a globally discriminative characterization. Corresponding points on two similar shapes will have similar shape contexts. The one to one correspondence is achieved through the cost based bipartite graph matching. The cost matrix is reduced through Hungarian algorithm. The dissimilarity between the two shapes is computed using Canberra distance. The nearest neighbor classifier is used to classify the objects with the matching error. The results are obtained using the MATLAB for MINIST hand written digits.
Recognition of Persian handwritten characters has been considered as a significant field of research for
the last few years under pattern analysing technique. In this paper, a new approach for robust handwritten
Persian numerals recognition using strong feature set and a classifier fusion method is scrutinized to
increase the recognition percentage. For implementing the classifier fusion technique, we have considered
k nearest neighbour (KNN), linear classifier (LC) and support vector machine (SVM) classifiers. The
innovation of this tactic is to attain better precision with few features using classifier fusion method. For
evaluation of the proposed method we considered a Persian numerals database with 20,000 handwritten
samples. Spending 15,000 samples for training stage, we verified our technique on other 5,000 samples,
and the correct recognition ratio achievedapproximately 99.90%. Additional, we got 99.97% exactness
using four-fold cross validation procedure on 20,000 databases.
Visual representation and organization of the knowledge have been utilized in different ways in tutoring
systems to upgrade their usefulness. This paper concentrates on the usage of various graphical formalisms,
for example, the conceptual graph, ontology, and concept map in tutoring systems. The paper addresses
what is way of the utilization of every formalism and the offering of the potential outcomes to assist the
student in education systems.
GRAPHICAL REPRESENTATION IN TUTORING SYSTEMSijcsit
Visual representation and organization of the knowledge have been utilized in different ways in tutoring systems to upgrade their usefulness. This paper concentrates on the usage of various graphical formalisms, for example, the conceptual graph, ontology, and concept map in tutoring systems. The paper addresses what is way of the utilization of every formalism and the offering of the potential outcomes to assist the student in education systems.
This document discusses different graphical representations that can be used in tutoring systems, including conceptual graphs, ontologies, and concept maps. It examines how each representation has been utilized in tutoring systems, including representing the domain or learning material, diagnosing student errors, and assessing student knowledge through visual concept mapping exercises. Concept maps, ontologies, and conceptual graphs each have their own characteristics that make them suitable for different uses in tutoring systems.
On comprehensive analysis of learning algorithms on pedestrian detection usin...UniversitasGadjahMada
Despite the surge of deep learning, deploying the deep learning-based pedestrian detection into the real system faces hurdles, mainly due to the huge resource usages. The classical feature-based detection system still becomes feasible option. There have been many efforts to improve the performance of pedestrian detection system. Among many feature set, Histogram of Oriented Gradient seems to be very effective for person detection. In this research, various machine learning algorithms are investigated for person detection. Different machine learning algorithms are evaluated to obtain the optimal accuracy and speed of the system.
Face detection using the 3 x3 block rank patterns of gradient magnitude imagessipij
Face detection locates faces prior to various face-
related applications. The objective of face detecti
on is to
determine whether or not there are any faces in an
image and, if any, the location of each face is det
ected.
Face detection in real images is challenging due to
large variability of illumination and face appeara
nces.
This paper proposes a face detection algorithm usin
g the 3×3 block rank patterns of gradient magnitude
images and a geometrical face model. First, the ill
umination-corrected image of the face region is obt
ained
using the brightness plane that is produced using t
he locally minimum brightness of each block. Next,
the
illumination-corrected image is histogram equalized
, the face region is divided into nine (3×3) blocks
, and
two directional (horizontal and vertical) gradient
magnitude images are computed, from which the 3×3
block rank patterns are obtained. For face detectio
n, using the FERET and GT databases three types of
the
3×3 block rank patterns are a priori determined as
templates based on the distribution of the sum of t
he
gradient magnitudes of each block in the face candi
date region that is also composed of nine (3×3) blo
cks.
The 3×3 block rank patterns roughly classify whethe
r the detected face candidate region contains a fac
e or
not. Finally, facial features are detected and used
to validate the face model. The face candidate is
validated as a face if it is matched with the geome
trical face model. The proposed algorithm is tested
on the
Caltech database images and real images. Experiment
al results with a number of test images show the
effectiveness of the proposed algorithm.
1) The document proposes a penalized likelihood method using a penalty function like SCAD to perform simultaneous variable selection and parameter estimation in structural equation models (SEMs).
2) The method considers a general SEM where latent variables are linearly regressed on themselves with a coefficient matrix, avoiding the need to specify outcome and explanatory latent variables. Selecting nonzero coefficients in the matrix identifies the structure of the latent variable model.
3) Under regularity conditions, the consistency and oracle properties of the proposed penalized maximum likelihood estimators are established. An expectation-conditional maximization algorithm is developed for computation.
The document summarizes research on nonlinear correlation coefficients on manifolds. It defines a new nonlinear correlation coefficient called SEVP, proves some of its basic properties including that it ranges from 0 to 1. It discusses how to measure nonlinear correlation between variables on a manifold and reviews common dimensionality reduction methods for manifolds. The goal is to preserve nonlinear structure as much as possible by projecting onto the orthogonal complement of tangent spaces. An optimization problem is formulated to find the linear space with the largest angle to all tangent spaces, transforming it into an eigenvalue problem to solve.
LITERATURE SURVEY ON SPARSE REPRESENTATION FOR NEURAL NETWORK BASED FACE DETE...csijjournal
Face detection and recognition is a challenging problem in the field of image processing. In this paper, we reviewed some of the recent research works on face recognition. Issues with the previous face recognition
techniques are , time required is more for face recognition , recognition rate and database required to store the data . To overcome these problems sparse representation based classifier technique can be used.
Face Alignment Using Active Shape Model And Support Vector MachineCSCJournals
The document proposes improvements to the classical Active Shape Model (ASM) algorithm for face alignment. The improvements include: 1) Combining Sobel filtering and 2D profiles to build texture models, 2) Applying Canny edge detection for enhancement, 3) Using Support Vector Machines (SVM) to classify landmarks for more accurate localization, and 4) Automatically adjusting 2D profile lengths based on image size. Experimental results on two face databases show the proposed ASM-SVM method achieves better alignment accuracy than the classical ASM and other methods.
International Journal of Artificial Intelligence & Applications (IJAIA)gerogepatton
The International Journal of Artificial Intelligence & Applications (IJAIA) is a bi monthly open access peer-reviewed journal that publishes articles which contribute new results in all areas of the Artificial Intelligence & Applications (IJAIA). It is an international journal intended for professionals and researchers in all fields of AI for researchers, programmers, and software and hardware manufacturers. The journal also aims to publish new attempts in the form of special issues on emerging areas in Artificial Intelligence and applications.
PhotoQR: A Novel ID Card with an Encoded Viewgerogepatton
There is an increasing interest in developing techniques to identify and assess data to allow an easy and
continuous access to resources, services or places that require thorough ID control. Usually, in order to
give access to these resources, different kinds of documents are mandatory. In order to avoid forgeries
without the need of extra credentials, a new system –named photoQR, is here proposed. This system is
based on a ID card having two objects: one person’s picture (pre-processed via blur and/or swirl
techniques) and one QR code containing embedded data related to the picture. The idea is that the picture
and the QR code can assess each other by a proper hash value in the QR. The QR without the picture
cannot be assessed and vice versa. An open source prototype of the photoQR system has been implemented
in Python and can be used both in offline and real-time environments, which effectively combines security
concepts and image processing algorithms to obtain data assessment.
More Related Content
Similar to LEARNING SPLINE MODELS WITH THE EM ALGORITHM FOR SHAPE RECOGNITION
A NOVEL FEATURE SET FOR RECOGNITION OF SIMILAR SHAPED HANDWRITTEN HINDI CHARA...cscpconf
This document describes a study that uses machine learning algorithms to recognize similar shaped handwritten Hindi characters. It extracts 85 features from each character image based on geometry. Four machine learning algorithms (Bayesian Network, RBFN, MLP, C4.5 Decision Tree) are trained on datasets containing samples of a target character pair and evaluated based on precision, misclassification rate, and model build time. Feature selection techniques are also used to reduce the feature set dimensionality before classification. Experimental results show that different algorithms perform best depending on the feature set and number of samples used for training.
Proposal of a similarity measure for unified modeling language class diagram ...IJECEIAES
The unified modeling language (UML) represents an essential tool for modeling and visualizing software systems. UML diagrams provide a graphical representation of a system's components. Comparing and processing these diagrams, for instance, can be complicated, especially as software projects grow in size and complexity. In such contexts, deep learning techniques have emerged as a promising solution for solving complex problems. One of these crucial problems is the measurement of similarity between images, making it possible to compare and calculate the differences between two given diagrams. The present work intends to build a method for calculating the degree of similarity between two UML class diagrams. With a goal to provide teachers a helpful tool for assessing students' UML class diagrams.
CHECKING BEHAVIOURAL COMPATIBILITY IN SERVICE COMPOSITION WITH GRAPH TRANSFOR...csandit
The success of Service Oriented Architecture (SOA) largely depended on the success of
automatic service composition. Dynamic service selection process should ensure full
compatibility between the services involved in the composition. This compatibility must be both
on static proprieties, called interface compatibility which can be easily proved and especially
on behavioural compatibility that needs composability checking of basic services. In this paper,
we propose (1) a formalism for modelling composite services using an extension of the Business
Process (BP) modelling approach proposed by Benatallah et al. and (2) a formal verification
approach of service composition. This approach uses the Graph Transformation (GT)
methodology as a formal verification tool. It allows behavioural compatibility verification of
two given services modelled by their BPs, used as the source graph in the GT operation. The
idea consists of (1) trying to dynamically generate a graph grammar R (a set of transformation
rules) whose application generates the composite service, if it exists, in this case (2) the next
step consist in checking the deadlock free in the resulting composite service. To this end we
propose an algorithm that we have implemented using the AGG, an algebraic graph
transformation API environment under eclipse IDE.
Statistical learning theory provides a framework for machine learning that draws from statistics and functional analysis. It deals with finding predictive functions based on data. Supervised learning involves learning from input-output pairs in a training data set to infer the function that maps inputs to outputs. The goals of learning are understanding relationships and enabling prediction. Statistical learning in R can perform regression for continuous outputs or classification for discrete outputs. It is commonly used in fields like computer vision, speech recognition, and bioinformatics.
FRACTAL ANALYSIS OF GOOD PROGRAMMING STYLEcscpconf
This paper studies a new, quantitative approach using fractal geometry to analyze basic tenets of good programming style. Experiments on C source of the GNU/Linux Core Utilities, a
collection of 114 programs or approximately 70,000 lines of code, show systematic changes in style are correlated with statistically significant changes in fractal dimension (P≤0.0009). The data further show positive but weak correlation between lines of code and fractal dimension (r=0.0878). These results suggest the fractal dimension is a reliable metric of changes that
affect good style, the knowledge of which may be useful for maintaining a code base.
This paper studies a new, quantitative approach using fractal geometry to analyse basic tenets
of good programming style. Experiments on C source of the GNU/Linux Core Utilities, a
collection of 114 programs or approximately 70,000 lines of code, show systematic changes in
style are correlated with statistically significant changes in fractal dimension (P≤0.0009). The
data further show positive but weak correlation between lines of code and fractal dimension
(r=0.0878). These results suggest the fractal dimension is a reliable metric of changes that
affect good style, the knowledge of which may be useful for maintaining a code base.
Object recognition is the challenging problem in the real world application. Object recognition can be achieved through the shape matching. Shape matching is preceded by i) detecting the edges of the objects from the images. ii) Finding the correspondence between the shapes. iii) Measuring the dissimilarity between the shapes using the correspondence. iv) Classifying the object into classes by using this dissimilarity measures. In order to solve the correspondence problem, we attach a descriptor, the shape context, to each point. The shape context at a reference point captures the distribution of the remaining points relative to it, thus offering a globally discriminative characterization. Corresponding points on two similar shapes will have similar shape contexts. The one to one correspondence is achieved through the cost based bipartite graph matching. The cost matrix is reduced through Hungarian algorithm. The dissimilarity between the two shapes is computed using Canberra distance. The nearest neighbor classifier is used to classify the objects with the matching error. The results are obtained using the MATLAB for MINIST hand written digits.
Recognition of Persian handwritten characters has been considered as a significant field of research for
the last few years under pattern analysing technique. In this paper, a new approach for robust handwritten
Persian numerals recognition using strong feature set and a classifier fusion method is scrutinized to
increase the recognition percentage. For implementing the classifier fusion technique, we have considered
k nearest neighbour (KNN), linear classifier (LC) and support vector machine (SVM) classifiers. The
innovation of this tactic is to attain better precision with few features using classifier fusion method. For
evaluation of the proposed method we considered a Persian numerals database with 20,000 handwritten
samples. Spending 15,000 samples for training stage, we verified our technique on other 5,000 samples,
and the correct recognition ratio achievedapproximately 99.90%. Additional, we got 99.97% exactness
using four-fold cross validation procedure on 20,000 databases.
Visual representation and organization of the knowledge have been utilized in different ways in tutoring
systems to upgrade their usefulness. This paper concentrates on the usage of various graphical formalisms,
for example, the conceptual graph, ontology, and concept map in tutoring systems. The paper addresses
what is way of the utilization of every formalism and the offering of the potential outcomes to assist the
student in education systems.
GRAPHICAL REPRESENTATION IN TUTORING SYSTEMSijcsit
Visual representation and organization of the knowledge have been utilized in different ways in tutoring systems to upgrade their usefulness. This paper concentrates on the usage of various graphical formalisms, for example, the conceptual graph, ontology, and concept map in tutoring systems. The paper addresses what is way of the utilization of every formalism and the offering of the potential outcomes to assist the student in education systems.
This document discusses different graphical representations that can be used in tutoring systems, including conceptual graphs, ontologies, and concept maps. It examines how each representation has been utilized in tutoring systems, including representing the domain or learning material, diagnosing student errors, and assessing student knowledge through visual concept mapping exercises. Concept maps, ontologies, and conceptual graphs each have their own characteristics that make them suitable for different uses in tutoring systems.
On comprehensive analysis of learning algorithms on pedestrian detection usin...UniversitasGadjahMada
Despite the surge of deep learning, deploying the deep learning-based pedestrian detection into the real system faces hurdles, mainly due to the huge resource usages. The classical feature-based detection system still becomes feasible option. There have been many efforts to improve the performance of pedestrian detection system. Among many feature set, Histogram of Oriented Gradient seems to be very effective for person detection. In this research, various machine learning algorithms are investigated for person detection. Different machine learning algorithms are evaluated to obtain the optimal accuracy and speed of the system.
Face detection using the 3 x3 block rank patterns of gradient magnitude imagessipij
Face detection locates faces prior to various face-
related applications. The objective of face detecti
on is to
determine whether or not there are any faces in an
image and, if any, the location of each face is det
ected.
Face detection in real images is challenging due to
large variability of illumination and face appeara
nces.
This paper proposes a face detection algorithm usin
g the 3×3 block rank patterns of gradient magnitude
images and a geometrical face model. First, the ill
umination-corrected image of the face region is obt
ained
using the brightness plane that is produced using t
he locally minimum brightness of each block. Next,
the
illumination-corrected image is histogram equalized
, the face region is divided into nine (3×3) blocks
, and
two directional (horizontal and vertical) gradient
magnitude images are computed, from which the 3×3
block rank patterns are obtained. For face detectio
n, using the FERET and GT databases three types of
the
3×3 block rank patterns are a priori determined as
templates based on the distribution of the sum of t
he
gradient magnitudes of each block in the face candi
date region that is also composed of nine (3×3) blo
cks.
The 3×3 block rank patterns roughly classify whethe
r the detected face candidate region contains a fac
e or
not. Finally, facial features are detected and used
to validate the face model. The face candidate is
validated as a face if it is matched with the geome
trical face model. The proposed algorithm is tested
on the
Caltech database images and real images. Experiment
al results with a number of test images show the
effectiveness of the proposed algorithm.
1) The document proposes a penalized likelihood method using a penalty function like SCAD to perform simultaneous variable selection and parameter estimation in structural equation models (SEMs).
2) The method considers a general SEM where latent variables are linearly regressed on themselves with a coefficient matrix, avoiding the need to specify outcome and explanatory latent variables. Selecting nonzero coefficients in the matrix identifies the structure of the latent variable model.
3) Under regularity conditions, the consistency and oracle properties of the proposed penalized maximum likelihood estimators are established. An expectation-conditional maximization algorithm is developed for computation.
The document summarizes research on nonlinear correlation coefficients on manifolds. It defines a new nonlinear correlation coefficient called SEVP, proves some of its basic properties including that it ranges from 0 to 1. It discusses how to measure nonlinear correlation between variables on a manifold and reviews common dimensionality reduction methods for manifolds. The goal is to preserve nonlinear structure as much as possible by projecting onto the orthogonal complement of tangent spaces. An optimization problem is formulated to find the linear space with the largest angle to all tangent spaces, transforming it into an eigenvalue problem to solve.
LITERATURE SURVEY ON SPARSE REPRESENTATION FOR NEURAL NETWORK BASED FACE DETE...csijjournal
Face detection and recognition is a challenging problem in the field of image processing. In this paper, we reviewed some of the recent research works on face recognition. Issues with the previous face recognition
techniques are , time required is more for face recognition , recognition rate and database required to store the data . To overcome these problems sparse representation based classifier technique can be used.
Face Alignment Using Active Shape Model And Support Vector MachineCSCJournals
The document proposes improvements to the classical Active Shape Model (ASM) algorithm for face alignment. The improvements include: 1) Combining Sobel filtering and 2D profiles to build texture models, 2) Applying Canny edge detection for enhancement, 3) Using Support Vector Machines (SVM) to classify landmarks for more accurate localization, and 4) Automatically adjusting 2D profile lengths based on image size. Experimental results on two face databases show the proposed ASM-SVM method achieves better alignment accuracy than the classical ASM and other methods.
Similar to LEARNING SPLINE MODELS WITH THE EM ALGORITHM FOR SHAPE RECOGNITION (20)
International Journal of Artificial Intelligence & Applications (IJAIA)gerogepatton
The International Journal of Artificial Intelligence & Applications (IJAIA) is a bi monthly open access peer-reviewed journal that publishes articles which contribute new results in all areas of the Artificial Intelligence & Applications (IJAIA). It is an international journal intended for professionals and researchers in all fields of AI for researchers, programmers, and software and hardware manufacturers. The journal also aims to publish new attempts in the form of special issues on emerging areas in Artificial Intelligence and applications.
PhotoQR: A Novel ID Card with an Encoded Viewgerogepatton
There is an increasing interest in developing techniques to identify and assess data to allow an easy and
continuous access to resources, services or places that require thorough ID control. Usually, in order to
give access to these resources, different kinds of documents are mandatory. In order to avoid forgeries
without the need of extra credentials, a new system –named photoQR, is here proposed. This system is
based on a ID card having two objects: one person’s picture (pre-processed via blur and/or swirl
techniques) and one QR code containing embedded data related to the picture. The idea is that the picture
and the QR code can assess each other by a proper hash value in the QR. The QR without the picture
cannot be assessed and vice versa. An open source prototype of the photoQR system has been implemented
in Python and can be used both in offline and real-time environments, which effectively combines security
concepts and image processing algorithms to obtain data assessment.
12th International Conference of Artificial Intelligence and Fuzzy Logic (AI ...gerogepatton
12th International Conference of Artificial Intelligence and Fuzzy Logic (AI & FL 2024) provides a
forum for researchers who address this issue and to present their work in a peer-reviewed forum. Authors
are solicited to contribute to the conference by submitting articles that illustrate research results, projects,
surveying works and industrial experiences that describe significant advances in the following areas, but
are not limited to these topics only.
International Journal of Artificial Intelligence & Applications (IJAIA)gerogepatton
The International Journal of Artificial Intelligence & Applications (IJAIA) is a bi monthly open access peer-reviewed journal that publishes articles which contribute new results in all areas of the Artificial Intelligence & Applications (IJAIA). It is an international journal intended for professionals and researchers in all fields of AI for researchers, programmers, and software and hardware manufacturers. The journal also aims to publish new attempts in the form of special issues on emerging areas in Artificial Intelligence and applications.
International Conference on NLP, Artificial Intelligence, Machine Learning an...gerogepatton
International Conference on NLP, Artificial Intelligence, Machine Learning and Applications (NLAIM 2024) offers a premier global platform for exchanging insights and findings in the theory, methodology, and applications of NLP, Artificial Intelligence, Machine Learning, and their applications. The conference seeks substantial contributions across all key domains of NLP, Artificial Intelligence, Machine Learning, and their practical applications, aiming to foster both theoretical advancements and real-world implementations. With a focus on facilitating collaboration between researchers and practitioners from academia and industry, the conference serves as a nexus for sharing the latest developments in the field.
DEEP LEARNING FOR SMART GRID INTRUSION DETECTION: A HYBRID CNN-LSTM-BASED MODELgerogepatton
As digital technology becomes more deeply embedded in power systems, protecting the communication
networks of Smart Grids (SG) has emerged as a critical concern. Distributed Network Protocol 3 (DNP3)
represents a multi-tiered application layer protocol extensively utilized in Supervisory Control and Data
Acquisition (SCADA)-based smart grids to facilitate real-time data gathering and control functionalities.
Robust Intrusion Detection Systems (IDS) are necessary for early threat detection and mitigation because
of the interconnection of these networks, which makes them vulnerable to a variety of cyberattacks. To
solve this issue, this paper develops a hybrid Deep Learning (DL) model specifically designed for intrusion
detection in smart grids. The proposed approach is a combination of the Convolutional Neural Network
(CNN) and the Long-Short-Term Memory algorithms (LSTM). We employed a recent intrusion detection
dataset (DNP3), which focuses on unauthorized commands and Denial of Service (DoS) cyberattacks, to
train and test our model. The results of our experiments show that our CNN-LSTM method is much better
at finding smart grid intrusions than other deep learning algorithms used for classification. In addition,
our proposed approach improves accuracy, precision, recall, and F1 score, achieving a high detection
accuracy rate of 99.50%.
10th International Conference on Artificial Intelligence and Applications (AI...gerogepatton
10th International Conference on Artificial Intelligence and Applications (AI 2024) will provide an excellent international forum for sharing knowledge and results in theory, methodology and applications of Artificial Intelligence and its applications. The Conference looks for significant contributions to all major fields of the Artificial Intelligence, Soft Computing in theoretical and practical aspects. The aim of the Conference is to provide a platform to the researchers and practitioners from both academia as well as industry to meet and share cutting-edge development in the field.
International Journal of Artificial Intelligence & Applications (IJAIA)gerogepatton
The International Journal of Artificial Intelligence & Applications (IJAIA) is a bi monthly open access peer-reviewed journal that publishes articles which contribute new results in all areas of the Artificial Intelligence & Applications (IJAIA). It is an international journal intended for professionals and researchers in all fields of AI for researchers, programmers, and software and hardware manufacturers. The journal also aims to publish new attempts in the form of special issues on emerging areas in Artificial Intelligence and applications.
Immunizing Image Classifiers Against Localized Adversary Attacksgerogepatton
This paper addresses the vulnerability of deep learning models, particularly convolutional neural networks
(CNN)s, to adversarial attacks and presents a proactive training technique designed to counter them. We
introduce a novel volumization algorithm, which transforms 2D images into 3D volumetric representations.
When combined with 3D convolution and deep curriculum learning optimization (CLO), itsignificantly improves
the immunity of models against localized universal attacks by up to 40%. We evaluate our proposed approach
using contemporary CNN architectures and the modified Canadian Institute for Advanced Research (CIFAR-10
and CIFAR-100) and ImageNet Large Scale Visual Recognition Challenge (ILSVRC12) datasets, showcasing
accuracy improvements over previous techniques. The results indicate that the combination of the volumetric
input and curriculum learning holds significant promise for mitigating adversarial attacks without necessitating
adversary training.
May 2024 - Top 10 Read Articles in Artificial Intelligence and Applications (...gerogepatton
The International Journal of Artificial Intelligence & Applications (IJAIA) is a bi monthly open access peer-reviewed journal that publishes articles which contribute new results in all areas of the Artificial Intelligence & Applications (IJAIA). It is an international journal intended for professionals and researchers in all fields of AI for researchers, programmers, and software and hardware manufacturers. The journal also aims to publish new attempts in the form of special issues on emerging areas in Artificial Intelligence and applications.
3rd International Conference on Artificial Intelligence Advances (AIAD 2024)gerogepatton
3rd International Conference on Artificial Intelligence Advances (AIAD 2024) will act as a major forum for the presentation of innovative ideas, approaches, developments, and research projects in the area advanced Artificial Intelligence. It will also serve to facilitate the exchange of information between researchers and industry professionals to discuss the latest issues and advancement in the research area. Core areas of AI and advanced multi-disciplinary and its applications will be covered during the conferences.
International Journal of Artificial Intelligence & Applications (IJAIA)gerogepatton
The International Journal of Artificial Intelligence & Applications (IJAIA) is a bi monthly open access peer-reviewed journal that publishes articles which contribute new results in all areas of the Artificial Intelligence & Applications (IJAIA). It is an international journal intended for professionals and researchers in all fields of AI for researchers, programmers, and software and hardware manufacturers. The journal also aims to publish new attempts in the form of special issues on emerging areas in Artificial Intelligence and applications.
Information Extraction from Product Labels: A Machine Vision Approachgerogepatton
This research tackles the challenge of manual data extraction from product labels by employing a blend of
computer vision and Natural Language Processing (NLP). We introduce an enhanced model that combines
Convolutional Neural Networks (CNN) and Recurrent Neural Networks (RNN) in a Convolutional
Recurrent Neural Network (CRNN) for reliable text recognition. Our model is further refined by
incorporating the Tesseract OCR engine, enhancing its applicability in Optical Character Recognition
(OCR) tasks. The methodology is augmented by NLP techniques and extended through the Open Food
Facts API (Application Programming Interface) for database population and text-only label prediction.
The CRNN model is trained on encoded labels and evaluated for accuracy on a dedicated test set.
Importantly, our approach enables visually impaired individuals to access essential information on
product labels, such as directions and ingredients. Overall, the study highlights the efficacy of deep
learning and OCR in automating label extraction and recognition.
10th International Conference on Artificial Intelligence and Applications (AI...gerogepatton
10th International Conference on Artificial Intelligence and Applications (AI 2024) will provide an excellent international forum for sharing knowledge and results in theory, methodology and applications of Artificial Intelligence and its applications. The Conference looks for significant contributions to all major fields of the Artificial Intelligence, Soft Computing in theoretical and practical aspects. The aim of the Conference is to provide a platform to the researchers and practitioners from both academia as well as industry to meet and share cutting-edge development in the field.
International Journal of Artificial Intelligence & Applications (IJAIA)gerogepatton
The International Journal of Artificial Intelligence & Applications (IJAIA) is a bi monthly open access peer-reviewed journal that publishes articles which contribute new results in all areas of the Artificial Intelligence & Applications (IJAIA). It is an international journal intended for professionals and researchers in all fields of AI for researchers, programmers, and software and hardware manufacturers. The journal also aims to publish new attempts in the form of special issues on emerging areas in Artificial Intelligence and applications.
Research on Fuzzy C- Clustering Recursive Genetic Algorithm based on Cloud Co...gerogepatton
Aiming at the problems of poor local search ability and precocious convergence of fuzzy C-cluster
recursive genetic algorithm (FOLD++), a new fuzzy C-cluster recursive genetic algorithm based on
Bayesian function adaptation search (TS) was proposed by incorporating the idea of Bayesian function
adaptation search into fuzzy C-cluster recursive genetic algorithm. The new algorithm combines the
advantages of FOLD++ and TS. In the early stage of optimization, fuzzy C-cluster recursive genetic
algorithm is used to get a good initial value, and the individual extreme value pbest is put into Bayesian
function adaptation table. In the late stage of optimization, when the searching ability of fuzzy C-cluster
recursive genetic is weakened, the short term memory function of Bayesian function adaptation table in
Bayesian function adaptation search algorithm is utilized. Make it jump out of the local optimal solution,
and allow bad solutions to be accepted during the search. The improved algorithm is applied to function
optimization, and the simulation results show that the calculation accuracy and stability of the algorithm
are improved, and the effectiveness of the improved algorithm is verified
International Journal of Artificial Intelligence & Applications (IJAIA)gerogepatton
The International Journal of Artificial Intelligence & Applications (IJAIA) is a bi monthly open access peer-reviewed journal that publishes articles which contribute new results in all areas of the Artificial Intelligence & Applications (IJAIA). It is an international journal intended for professionals and researchers in all fields of AI for researchers, programmers, and software and hardware manufacturers. The journal also aims to publish new attempts in the form of special issues on emerging areas in Artificial Intelligence and applications.
10th International Conference on Artificial Intelligence and Soft Computing (...gerogepatton
10th International Conference on Artificial Intelligence and Soft Computing (AIS 2024) will
provide an excellent international forum for sharing knowledge and results in theory, methodology, and
applications of Artificial Intelligence, Soft Computing. The Conference looks for significant
contributions to all major fields of the Artificial Intelligence, Soft Computing in theoretical and practical
aspects. The aim of the Conference is to provide a platform to the researchers and practitioners from
both academia as well as industry to meet and share cutting-edge development in the field.
International Journal of Artificial Intelligence & Applications (IJAIA)gerogepatton
Employee attrition refers to the decrease in staff numbers within an organization due to various reasons.
As it has a negative impact on long-term growth objectives and workplace productivity, firms have
recognized it as a significant concern. To address this issue, organizations are increasingly turning to
machine-learning approaches to forecast employee attrition rates. This topic has gained significant
attention from researchers, especially in recent times. Several studies have applied various machinelearning methods to predict employee attrition, producing different resultsdepending on the employed
methods, factors, and datasets. However, there has been no comprehensive comparative review of multiple
studies applying machine-learning models to predict employee attrition to date. Therefore, this study aims
to fill this gap by providing an overview of research conducted on applying machine learning to predict
employee attrition from 2019 to February 2024. A literature review of relevant studies was conducted,
summarized, and classified. Most studies agree on conducting comparative experiments with multiple
predictive models to determine the most effective one.From this literature survey, the RF algorithm and
XGB ensemble method are repeatedly the best-performing, outperforming many other algorithms.
Additionally, the application of deep learning to employee attrition prediction issues also shows promise.
While there are discrepancies in the datasets used in previous studies, it is notable that the dataset
provided by IBM is the most widely utilized. This study serves as a concise review for new researchers,
facilitating their understanding of the primary techniques employed in predicting employee attrition and
highlighting recent research trends in this field. Furthermore, it provides organizations with insight into
the prominent factors affecting employee attrition, as identified by studies, enabling them to implement
solutions aimed at reducing attrition rates.
10th International Conference on Artificial Intelligence and Applications (AI...gerogepatton
10th International Conference on Artificial Intelligence and Applications (AIFU 2024) is a forum for presenting new advances and research results in the fields of Artificial Intelligence. The conference will bring together leading researchers, engineers and scientists in the domain of interest from around the world. The scope of the conference covers all theoretical and practical aspects of the Artificial Intelligence.
Orchestrating the Future: Navigating Today's Data Workflow Challenges with Ai...Kaxil Naik
Navigating today's data landscape isn't just about managing workflows; it's about strategically propelling your business forward. Apache Airflow has stood out as the benchmark in this arena, driving data orchestration forward since its early days. As we dive into the complexities of our current data-rich environment, where the sheer volume of information and its timely, accurate processing are crucial for AI and ML applications, the role of Airflow has never been more critical.
In my journey as the Senior Engineering Director and a pivotal member of Apache Airflow's Project Management Committee (PMC), I've witnessed Airflow transform data handling, making agility and insight the norm in an ever-evolving digital space. At Astronomer, our collaboration with leading AI & ML teams worldwide has not only tested but also proven Airflow's mettle in delivering data reliably and efficiently—data that now powers not just insights but core business functions.
This session is a deep dive into the essence of Airflow's success. We'll trace its evolution from a budding project to the backbone of data orchestration it is today, constantly adapting to meet the next wave of data challenges, including those brought on by Generative AI. It's this forward-thinking adaptability that keeps Airflow at the forefront of innovation, ready for whatever comes next.
The ever-growing demands of AI and ML applications have ushered in an era where sophisticated data management isn't a luxury—it's a necessity. Airflow's innate flexibility and scalability are what makes it indispensable in managing the intricate workflows of today, especially those involving Large Language Models (LLMs).
This talk isn't just a rundown of Airflow's features; it's about harnessing these capabilities to turn your data workflows into a strategic asset. Together, we'll explore how Airflow remains at the cutting edge of data orchestration, ensuring your organization is not just keeping pace but setting the pace in a data-driven future.
Session in https://budapestdata.hu/2024/04/kaxil-naik-astronomer-io/ | https://dataml24.sessionize.com/session/667627
4th Modern Marketing Reckoner by MMA Global India & Group M: 60+ experts on W...Social Samosa
The Modern Marketing Reckoner (MMR) is a comprehensive resource packed with POVs from 60+ industry leaders on how AI is transforming the 4 key pillars of marketing – product, place, price and promotions.
Beyond the Basics of A/B Tests: Highly Innovative Experimentation Tactics You...Aggregage
This webinar will explore cutting-edge, less familiar but powerful experimentation methodologies which address well-known limitations of standard A/B Testing. Designed for data and product leaders, this session aims to inspire the embrace of innovative approaches and provide insights into the frontiers of experimentation!
"Financial Odyssey: Navigating Past Performance Through Diverse Analytical Lens"sameer shah
Embark on a captivating financial journey with 'Financial Odyssey,' our hackathon project. Delve deep into the past performance of two companies as we employ an array of financial statement analysis techniques. From ratio analysis to trend analysis, uncover insights crucial for informed decision-making in the dynamic world of finance."
End-to-end pipeline agility - Berlin Buzzwords 2024Lars Albertsson
We describe how we achieve high change agility in data engineering by eliminating the fear of breaking downstream data pipelines through end-to-end pipeline testing, and by using schema metaprogramming to safely eliminate boilerplate involved in changes that affect whole pipelines.
A quick poll on agility in changing pipelines from end to end indicated a huge span in capabilities. For the question "How long time does it take for all downstream pipelines to be adapted to an upstream change," the median response was 6 months, but some respondents could do it in less than a day. When quantitative data engineering differences between the best and worst are measured, the span is often 100x-1000x, sometimes even more.
A long time ago, we suffered at Spotify from fear of changing pipelines due to not knowing what the impact might be downstream. We made plans for a technical solution to test pipelines end-to-end to mitigate that fear, but the effort failed for cultural reasons. We eventually solved this challenge, but in a different context. In this presentation we will describe how we test full pipelines effectively by manipulating workflow orchestration, which enables us to make changes in pipelines without fear of breaking downstream.
Making schema changes that affect many jobs also involves a lot of toil and boilerplate. Using schema-on-read mitigates some of it, but has drawbacks since it makes it more difficult to detect errors early. We will describe how we have rejected this tradeoff by applying schema metaprogramming, eliminating boilerplate but keeping the protection of static typing, thereby further improving agility to quickly modify data pipelines without fear.
Open Source Contributions to Postgres: The Basics POSETTE 2024ElizabethGarrettChri
Postgres is the most advanced open-source database in the world and it's supported by a community, not a single company. So how does this work? How does code actually get into Postgres? I recently had a patch submitted and committed and I want to share what I learned in that process. I’ll give you an overview of Postgres versions and how the underlying project codebase functions. I’ll also show you the process for submitting a patch and getting that tested and committed.
Build applications with generative AI on Google CloudMárton Kodok
We will explore Vertex AI - Model Garden powered experiences, we are going to learn more about the integration of these generative AI APIs. We are going to see in action what the Gemini family of generative models are for developers to build and deploy AI-driven applications. Vertex AI includes a suite of foundation models, these are referred to as the PaLM and Gemini family of generative ai models, and they come in different versions. We are going to cover how to use via API to: - execute prompts in text and chat - cover multimodal use cases with image prompts. - finetune and distill to improve knowledge domains - run function calls with foundation models to optimize them for specific tasks. At the end of the session, developers will understand how to innovate with generative AI and develop apps using the generative ai industry trends.
LEARNING SPLINE MODELS WITH THE EM ALGORITHM FOR SHAPE RECOGNITION
1. International Journal of Artificial Intelligence and Applications (IJAIA), Vol.14, No.6, November 2023
DOI:10.5121/ijaia.2023.14604 37
LEARNING SPLINE MODELS WITH THE EM
ALGORITHM FOR SHAPE RECOGNITION
Abdullah A. Al-Shaher, Yusef S. AlKhawari
Department of Computer Science and Information Systems, College of Business Studies,
Public Authority for Applied Education and Training, Kuwait
ABSTRACT
This paper demonstrates how cubic Spline (B-Spline) models can be used to recognize 2-dimension non-
rigid handwritten isolated characters. Each handwritten character is represented by a set of non-
overlapping uniformly distributed landmarks. The Spline models are constructed by utilizing cubic order of
polynomial to model the shapes under study. The approach is a two-stage process. The first stage is
learning, we construct a mixture of spline class parameters to capture the variations in spline coefficients
using the apparatus Expectation Maximization algorithm. The second stage is recognition, here we use the
Fréchet distance to compute the variations between the spline models and test spline shape for recognition.
We test the approach on a set of handwritten Arabic letters.
KEYWORDS
Spline Models, Cubic order of Spline Curves, handwritten Arabic Characters, Shape Recognition,
Recognition by Alignment. Expectation Maximisation Algorithm, Unsupervised Learning, Fréchet distance.
1. INTRODUCTION
Shape matching has been the focus of many researchers for the past seven decades [1] and
attracted many scholars in the field of pattern recognition [2], artificial intelligence [3], signal
processing [4], image analysis [5], and computer vision [6]. The difficulties arise when the shape
under study exhibits a high degree in shape variation: as in handwritten characters [7], digits [8],
face detection [9], and gesture authentication [10]. For a single data, shape variation is limited and
cannot be captured ultimately because single data does not provide sufficient information and
knowledge about the data; therefore, multiple existence of data provide better understanding of
shape analysis and manifested by mixture models [11]. Because of the existence of multivariate
data under study, there is always the requirement to estimate the parameters that describe the data
that is encapsulated within a mixture of shapes. Extensive efforts have been invested in analyzing
alphabet of English [12], Latin [13], Chinese [14], Hebrew [15], and Hindi [16]; however, very
little work has been done in analyzing handwritten Arabic characters and least considerable
results achieved [17]. The subject of investigating Arabic letters shape analysis is complicated.
Such difficulties arise from the nature structure of Arabic alphabet and words. Arabic script is
cursive, continues, and similar without diacritics. For example: the word م عل has different
meaning when secondaries are attached to the letters.
The literature demonstrates many statistical and structural approaches with various algorithms to
model shape variations using supervised and unsupervised learning [18] algorithms. In precise,
2. International Journal of Artificial Intelligence and Applications (IJAIA), Vol.14, No.6, November 2023
38
comes the powerful Expectation Maximization algorithm (EM) of Dempster [19] that has been
used widely for such cases. The EM algorithm revolves around two step procedures. The
expectation E step revolves around estimating the parameters of a log likelihood function and
pass it to the Maximization M step. In a maximization (M) step, the algorithm computes
parameters maximizing the expected log-likelihood found on the E step, the process is iterative
one until all parameters come to stability. For instance, Jojic and Frey [20] have used the EM
algorithm to fit mixture models to the appearance manifolds for faces. Bishop and Winn [21]
have used a mixture of principal components analyzers to learn and synthesize variations in facial
appearance. Vasconcelos and Lippman [22] have used the EM algorithm to learn queries for
content-based image retrieval. Finally, several authors have used the EM algorithm to track
multiple moving objects [23]. Revov et al. [24] has developed a generative model which can be
used for handwritten character recognition. Their method employs the EM algorithm to model the
distribution of sample points.
Spline models have been used widely in the signal processing community [25]. Spline models can
describe more complex shapes that contain circles, closed contours, ellipses, and curved shapes
and combine this description in mathematical and statistical framework. For example, Splines are
used to model Functional data analysis [26], Computer Aided Design [27], automobiles [28],
smoothed surfaces [29], curly shapes [30], and noisy materials such as clothing design [31].
Safraz [32] has investigated the issue of Arabic shape recognition by utilizing edge detection of
machine printed letters to bring the shape under landmarks representation, then applying cubic
polynomials on them. In deepest investigation, Safraz [33] addressed the subject of extracting the
curve estimation by finding the optimal set of coefficients of Bezier function and dividing the
curve into multiple segments ending with a smooth curve fitting.
In this research paper, we demonstrate how Spline models are used to recognize 2D handwritten
shapes by fitting 3rd order of polynomial function to a set of landmarks points extracted for the
shape under analysis. We then, train such Spline models to capture the optimal characteristics of
the shapes in the training sets of shapes. Handwritten isolated Arabic characters are used and
tested in this investigation.
2. SHAPE MODELLING
A spline curve is a mathematical representation to construct an interface that allows sequence of
points to design and control the shape under study.
The parametric form of shape coordinates is called control points. The core concept of pattern
recognition methods and approaches aim at constructing a classifier for a set of training patterns.
Such a classifier determines which shape-class the input pattern belongs to; therefore, is known as
object training or shape learning [34][35]. In real world applications such as object recognition, a
certain previous knowledge is required beforehand. The utilization of such knowledge into a
training set is the key element that will allow to increase of performance of shape classification.
Following the concept of constructing a training set of patterns undergo classification. The
landmark patterns are collected as the object in question undergoes representative changes in
shape. To be more formal, each landmark pattern consists of L labelled points. Suppose that there
are T landmark patterns. The tth training pattern is represented using the long vector of landmark
co-ordinates:
(11)
3. International Journal of Artificial Intelligence and Applications (IJAIA), Vol.14, No.6, November 2023
39
where the subscripts of the co-ordinates are the landmark labels. A spline curve is a curve that
passes through these tth training pattern co-ordinates called interpolating curves. To establish
interface, we approach the polynomial in the form of function of landmarks:
(12) +
…..
The degree of the polynomial corresponds with the highest coefficient that is nonzero. The aim of
using cubic spline is to provide a smooth curve that passes through the chosen landmarks for
shape construction as well as support inflection points. Equation (12) can be revised in the form
of cubic polynomials.
(13)
Equation (13) can be rewritten in the form of solving linear equations to solve for the spline
coefficients obtained by the curve segment of 4 coordinates. The new mathematical
formulation to solve for required coefficients, we present a matrix of linear equations.
(14)
Where is the matrix of linear equations
(15)
The solution is
(16)
Similar procedure is applied to the remaining coordinate segments. The resulting set of spline
coefficients presented by
(17)
And in unification form of the set spline coefficients to be
) (18)
Suppose that each shape class have T training spline mean coefficients . The mean vector
of is represented by
4. International Journal of Artificial Intelligence and Applications (IJAIA), Vol.14, No.6, November 2023
40
(19)
The covariance matrix is then constructed.
(20)
3. LEARNING SPLINE MODEL
In machine learning approaches [21], supervised learning depends on data labelling which the
outcomes models are mapping algorithms between data and its counterparts. On the other hand,
unsupervised learning data patterns assumed from the unlabeled input data, hence, the goal of
unsupervised learning is to find the structure and variations from the input data. In another format,
Unsupervised learning does not need any supervision architecture. Instead, it finds data variations
and variable modes from the data by itself. To reproduce more complex variations in shape either
a non-linear deformation or a series of local piecewise linear deformations must be employed. In
this paper we adopt an approach based on mixtures of spline model distributions. Our reasons for
adopting this approach are twofold. First, we would like to be able to model more complex
deformations by using multiple modes of shape deformation. This arises when a set of training
patterns contains examples from different shape classes. The second reason is that shape
variations cannot be captured in a single mode of shape deformation, hence a mixture is
constructed. We commence by assuming that the individual examples in the training set are
conditionally independent of one another. The following approach is based on fitting a Gaussian
mixture model to the set of training examples of spline coefficients. We further assume that the
training data can be represented by a set of spline-classes Ω.
Each spline coefficient model ꞷ has its own mean coefficients and covariance matrix .
With these ingredients we establish the likelihood function for the set of the curve patterns in
(21)
Where the term is the probability for drawing curve pattern αt from the curve-
class ω. Associating the above likelihood function with the Expectation Maximization algorithm,
the likelihood function can be written to be iterative process of two steps. The process revolves
around estimating the expected log-likelihood function iteratively in
(22)
Where the quantity and are the estimated mean curve vector and variance
covariance matrix both at iteration (n) of the algorithm. The quantity is the
a posteriori probability that the training pattern spline coefficients belong to the spline-class
and
5. International Journal of Artificial Intelligence and Applications (IJAIA), Vol.14, No.6, November 2023
41
ω at iteration n of the algorithm. The term is the probability of
distribution of spline-pattern αt belonging to spline-class ω at iteration ( n + 1 ) of the algorithm;
thus, the probability density to associate curve- patterns for ( t = 1 … T ) to class spline-class
ω are estimated by the updated construction of the mean-vector , and
covariance matrix at iteration n+1 of the algorithm. According to the EM algorithm, it
revolves around estimation of the expected log-likelihood function within two iterative processes.
In the M or maximization step of the algorithm, our aim is to maximize the curve
mean-vector , and covariance matrix , while, in the E or expectation step, the aim
is to estimate the distribution of spline-patterns at iteration n along with the mixing proportion
parameters for curve-class ω.
In the E, or Expectation step of the algorithm, the a posteriori spline-class probability is updated
by applying the Bayes factorization rule to the spline-class distribution density at iteration ( n+1 ).
The new estimate is computed by
(23)
Where the revised spline-class ω mixing proportions at iteration ( n + 1 )is computed in
(24)
With that at hand, the distributed spline-pattern t to the class-spline ω is Gaussian distribution
and is classified according to
(29)
In the M, or Maximization step, our aim is to maximize the curve-class ω parameters. The
updated spline mean-vector estimate is computed by
(30)
6. International Journal of Artificial Intelligence and Applications (IJAIA), Vol.14, No.6, November 2023
42
And the new estimate of the spline-class covariance matrix is weighted by
(31)
Both E, and M steps are iteratively converged, the outcome of the learning stage is a set of spline-
class ω parameters such as , hence the complete set of all spline-class Ω are
computed and ready to be used for recognition.
4.SPLINE MATCHING
With the ingredients produced by the previous learning stage, the Spline model parameters
belongs to the set of spline-classes Ω can be used for the purpose of recognition.
To establish a mathematical setting, each control point in the test spline is mapped to its
counterparts in the spline model. Hence, we use the Fréchet distance [36].
Fréchet distance evaluates the similarity between two spline curves. Let's denote is the spline
mean and data spline data curve. Fréchet distance is defined as the minimum cordlength
sufficient to map two control points in both splines. The mapping continues until reaching the end
of the spline curves. The distance is then the end of point traced forward along and one
traveling forward along , although the rate of travel for either point may not
necessarily be uniform.
4.1. Spline Projection
Assume that the class-mean spline and the data spline curve are projected into 2
dimension coordinate space by estimating the y axes over
the spline coefficients. the resulting vector of shape representation for the class-mean spline
denoted by
(32)
And test-data spline projection is denoted by
(33)
4.2. Fréchet Matrix
The matrix of Fréchet distances is then computed for each pair of the control points of the class-
mean spline and the test-data spline in
7. International Journal of Artificial Intelligence and Applications (IJAIA), Vol.14, No.6, November 2023
43
( 34)
where
(35)
is the Euclidean distance between any pair of coordinates space. The matrix constructed is square
non-symmetric matrix. We consider the diagonal cells of the matrix that explain the distance
between control points pair for both objects under study, we choose the maximum value as a top
limit for the final distance. The algorithm proceeds by iterating through the diagonal left to right,
and for each column, calculates the distances and stores them if they are smaller than the
reference maximum. The same process applies to rows.
(36)
These values are compared with the other class-mean spline in the set , the class-mean spline
that has the minimum values is selected data spline is then grouped for that class.
5. EXPERIMENTS
We have evaluated our approach with sets of isolated Arabic handwritten characters. Here, we
have used 22 shape-classes for different writers, each with 40 training patterns. In total, we have
tested the approach with 880 handwritten Arabic character shape patterns for testing and 1760
patterns for testing our approach. Figure 1 shows some training patterns used in this research
paper. Figure 2 shows single shapes and their landmarks representation of being uniformly
distributed along the skeleton of the shape under investigation.
Figure 1: Sample handwritten character training sets.
8. International Journal of Artificial Intelligence and Applications (IJAIA), Vol.14, No.6, November 2023
44
Figure 2: Training patterns and extracted landmarks.
In figure 3, sample cubic spline coefficients resulted from the training stage for spline-class ꞷ.
Here, each raw represents the set of coefficients for the handwritten character ( Seen ) and for
each segment of the shape.
Figure 3: Spline coefficients for sample class-shape
In figure 4, illustrates the result from the training stage. First raw shows sample training sets, raw
2 shows the training stage Expectation Maximization initialization shape, raw 3, shows the
projection of the spline training coefficients producing the landmarks, and raw 4 show the spline-
mean for class shape representation. Figure 5 shows the convergence rate as a function per
9. International Journal of Artificial Intelligence and Applications (IJAIA), Vol.14, No.6, November 2023
45
iteration number. The graph shows how associated distributed probabilities for the set of spline-
classes Ω converged in a few iterations.
Figure 4: a) sample sets b) EM initialization c) Mean Shape landmarks
d) Spline Mean Shape
Figure 5: Convergence Rate as a function per iteration no.
10. International Journal of Artificial Intelligence and Applications (IJAIA), Vol.14, No.6, November 2023
46
Figure 6 shows the Fréchet distance table between the spline-mean shape and a testing shape
pattern.
The graph shows the similarity between the spline curves which considers the location and
ordering landmarks to the sample testing pattern landmarks. It is apparent from the table that the
similarity is shown in the diagonal elements highlighted by green color. The diagonal distance
between the landmarks for both the spline-mean shape and it's corresponding in the testing
pattern. The similarity excluded by the minimum distance.
Figure 6: Fréchet distance
To take the investigation further, we demonstrate how the approach behaves under the presence of
noise. In figure 7, we show how recognition rate is achieved when point position displacement
error is applied. Test shape coordinates are being moved away from their original position. The
figure shows the recognition rate fails to recognize shapes to their correct class ω in a few
iterations and it fails completely when coordinates are moved away, yet, increasing variance
significantly.
Figure 7: Recognition rate as a function per iteration no with point position error.
11. International Journal of Artificial Intelligence and Applications (IJAIA), Vol.14, No.6, November 2023
47
Table 1 shows recognition rates per spline-classes ω. Row 1 represents the shape name in Arabic
alphabet. Row 2 represents the character shape. Row 3 lists the size of the testing patterns. Rows
4,5 show the correct and wrong registration of shapes to their spline-class ω.
Row 6 shows the recognition rate per spline-class ω. In total, we have achieved 96% recognition
rate for such an approach.
Shape Name
Sample Shape Test Size Correct False
Recognition Rate
Ain_1 80 78 2 97.5%
Baa 80 78 2 97.5%
Dal 80 79 1 98.8%
Faa 80 73 7 91.3%
Haa_1 80 79 1 98.8%
Ttah 80 76 4 95%
Hhah_1 80 71 9 88.8%
Ain_2 80 79 1 98.8%
Meem 80 76 4 95%
Seen 80 74 6 92.5%
Yaa 80 77 3 96.3%
12. International Journal of Artificial Intelligence and Applications (IJAIA), Vol.14, No.6, November 2023
48
Haa_2 80 78 2 97.5%
Waw 80 76 4 95%
Ain_2 80 79 1 98.8%
Hhah_2 80 74 6 92.5%
Kaf_1 80 79 1 98.8%
Lam_1 80 80 0 100%
Lam_2 80 73 7 91.3%
Raa 80 79 1 98.8%
Ssad 80 73 7 91.3%
Kaf_2 80 78 2 97.5%
Noon 80 79 1 98.8%
Total 22 classes 1760 1688 72 96%
6. CONCLUSION
In this research paper, we have shown how cubic spline 3rd degree polynomials can be used for
recognizing handwritten Arabic characters. Each shape under testing is represented by a set of
uniformly distributed landmarks along the skeleton. A cubic piecewise B-Spline is implemented
on the shape to extract the coefficients needed for training set. A mixture of spline coefficients
has been trained in a two-level hierarchy Expectation Maximization setting to capture the modes
of spline parameters describing the pattern-sets. Recognition is then applied on a different set of
testing patterns using the Fréchet distance mechanism. We have shown a remarkable recognition
rate of 96%.
13. International Journal of Artificial Intelligence and Applications (IJAIA), Vol.14, No.6, November 2023
49
7. FUTURE WORK
In this research, there are some shortcomings to the approach. The first shortcoming is proving
cubic b-spline 3rd order of polynomials is best representation for handwritten Arabic characters.
The second is, the extracted parameters from the training stage have not been utilized for the
purpose of recognition stage encapsulated within a statistical framework: for example, spline
coefficients are not utilized for spline alignment. Secondly, there is a need to provide a
hierarchical framework for spline alignment.
REFERENCES
[1] J. Wood, "Invariant pattern recognition: A review," Pattern Recognition, vol. 29, no. 1, 1996, 1-17.
[2] Anil, K. Jain,Robert P.W. Duin, and Jianchang Mao: “Statistical Pattern Recognition: A Review”,
IEEE Pattern Analysis and Machine Intelligence, vol 22, No. 1, PP 4-37, 2000.
[3] P.F. Baldi and K. Hornik, "Learning in linear neural networks: A survey," IEEE Transactions on
Neural Networks, vol. 6, no. 4, 1995, 837-858.
[4] T.Y. Kong and A. Rosenfeld, "Digital topology: introduction and survey," Computer Vision,
Graphics, and Image Processing, vol. 48, no. 3, pp. 357-393, 1989.
[5] T.R. Reed and J.M.H. Dubuf, "A review of recent texture segmentation and feature extraction
techniques," CVGIP - Image Understanding, vol. 57, no. 3, pp. 359-372, 1993.
[6] S. Sarkar and K.L. Boyer, "Perceptual organization in computer vision - a review and a proposal for a
classificatory structure", IEEE Transactions on Systems Man and Cybernetics, vol. 23, no. 2, 1993,
382-399.
[7] Lorigo L. M., Govindaraju V., “Offline Arabic Handwriting Recognition: A Survey”, IEEE Trans
Pattern Anal Mach Intelligence, Vol: 28(5): PP 712-24, 2006.
[8] Ishani Patel, Virag Jagtap, Ompriya Kale, “A Survey on Feature Extraction Methods for Handwritten
Digits Recognition”, International Journal of Computer Applications, Vol 107, No 12, PP 1117,
2014.
[9] Muhammad Sharif, Farah Naz, Mussarat Yasmin, Muhammad Alyas Shahid, Amjad Rehman, “Face
Recognition: A Survey”, Journal of Engineering Science and Technology Review, Vol 10, No 2, PP
166-177, 2017.
[10] Rafiqul Zaman Khan, Noor Adnan Ibraheem, “HAND GESTURE RECOGNITION: A
LITERATURE REVIEW “,International Journal of Artificial Intelligence & Applications, Vol.3,
No.4, pp 161-174, 2012.
[11] Bruce George Lindsay,Mary L. Lesperance, “A review of semiparametric mixture models”, Journal of
Statistical Planning and Inference, Vol: 47 , No: 1 PP 29-39 , 1995.
[12] Patel, Shital P. Thakkar, "Handwritten Character Recognition in English: A Survey", International
Journal of Advanced Research in Computer and Communication Engineering, Vol. 4, Issue 2, 2015.
[13] Edyta LUKASIK, Malgorzata CHARYTANOWICZ, Marek MILOSZ, Michail TOKOVAROV,
Monika KACZOROWSKA, Dariusz CZERWINSKI, Tomasz ZIENTARSKI, "Recognition of
handwritten Latin characters with diacritics using CNN", ARTIFICIAL AND COMPUTATIONAL
INTELLIGENCE, Vol. 69. No 1, 2021.
[14] Dai Ruwei,Liu Chenglin,Xiao Baihua, "Chinese character recognition: history, status and prospects",
Frontiers of Computer Science in Chinavolume 1, pages126–136 (2007).
[15] Tabita L. Tobing, Sule Y. Yayilgan, Sony George, " Isolated Handwritten Character Recognition of
Ancient Hebrew Manuscripts ", SOCIETY FOR IMAGING SCIENCE AND TECHNOLOGY, PP
35-39, 2022.
[16] Madhuri Yadav,Ravindra Kumar, Purwar,Mamta Mittal, "Handwritten Hindi character recognition: a
review", IET Image Processing, Volume12,Issue11, November 2018, Pages 1919-1933.
[17] Pervez Ahmed and Yousef Al-Ohali, "Arabic Character Recognition: Progress and Challenges", J.
King Saud Univ., Vol. 12, Comp. & Info. Sci., pp. 85-116,2000.
[18] Christopher M. Bishop, “Neural Networks for Pattern Recognition”, Clarendon Press, 1995, ISSN
0198538642.
[19] A. Dempster, N. Laird, D. Rubin, "Maximum likelihood from incomplete data via the em algorithm",
J. Roy. Statist. Soc. Ser. 39 (1977) 1–38
14. International Journal of Artificial Intelligence and Applications (IJAIA), Vol.14, No.6, November 2023
50
[20] J. Brendan, N. Jojic, Estimating mixture models of images and inferring spatial transformations using
the em algorithm, IEEE Computer Vision Pattern Recognition 2 (1999) 416–422.
[21] C. Bishop, J. Winn, Non-linear Bayesian image modelling, Proceedings of Sixth European
[22] N. Vasconcelos, A. Lippman, A probabilistic architecture for content-based image retrieval,
Proceedings of International Conference on Computer Vision and Pattern Recognition, 2000, pp.
216–221.
[23] B. North, A. Blake, Using expectation-maximization to learn dynamical models from visual data,
Image Vision Computing. 17 (8) (1999) 611–616.
[24] M. Revow, C. Williams, G.E. Hinton, Using generative models for handwritten digit recognition,
IEEE Trans. Pattern Anal. Mach. Intell. 20 (2) (1996) 592–606
[25] M. Svoboda,L. Matiu-Iovan,P. Andea, B-spline interpolation technique for digital signal processing ,
International Conference on Information and Digital Technologies, 2015
[26] A.M. Aguilera, M.C. Aguilera-Morillo, Comparative study of different B-spline approaches for
functional data. Mathematical and Computer Modelling 58 (2013) 1568–1579.
[27] E. Dimas, D. Briassoulis, 3D geometric modeling based on NURBS: a review, Journal of Advances in
Engineering Software, (1999)741-751.
[28] A. Galvez, A. Iglesias, A new iterative mutually coupled hybrid GA-PSO approach for curve fitting in
manufacturing, Journal of Applied Soft Computing, (2012),
[29] I. Zeid, Mastering CAD/CAM, McGraw Hill International Edition, New York, 2005.
[30] M. Sevaux, Y. Mineur, A curve-fitting genetic algorithm for a styling application, European Journal
of Operational Research, 179 (2007) 895-905.
[31] R. Liutkevicius, A. Davidsonas, Surface Reconstruction from Partially Structured Noisy Cloud of
Points Using B-Splines, Journal of Information Technology and Control, 42 (2013).
[32] Muhammed Safraz, Mutaza Khan, “Automatic outline capture of Arabic fonts”, Journal of
information Sciences, Vol 140, No. 3-4, PP 269-281,2002
[33] Muhammad Safraz, “Optimal Curve Fitting to Digital Data”, Journal of WSCG, Vol 11, No. 1-3,
2003.
[34] AlShaher Abdullah, Hancock Edwin, Learning mixtures of Point Distribution models with the EM
algorithm. Pattern Recognition, 36(2003) 2805-2818.
[35] Christopher M. Bishop, Pattern Recognition and Machine Learning, Springer Science, and Business
Media, 2006.
[36] Thomas Heiter and Heikki Manila. Computing Discrete Fréchet Distance. Technical Report CD-TR
94/64, Christian Doppler Laboratory for Expert Systems, Technical University Vienna, Austria, 1994.
AUTHORS
ABDULLAH A. AL-SHAHER is currently enrolled as associate professor faculty
member in the Computer Science and Information Systems Department at the college of
Business Studies in the Public Authority for Applied Education and Training, Kuwait. He
received his Ph.D. in Computer Science from the University of York, York, United
Kingdom in 2005. In 1995 he was awarded master's degree (honour list) in Computer
Science from University of Denver, Denver, Colorado, USA, and a bachelor's degree from
HustonTillotson University in 1988 Austin, Texas, USA. His main interest is Shape
Recognition, Statistical Patten Recognition, Computer Vision, and Artificial Intelligence.
YOUSEF S. ALKHAWARI is currently enrolled as a lecturer faculty member in the
Computer Science and Information Systems Department at the College of Business Studies
in the Public Authority for Applied Education and Training (PAAET), Kuwait. He
received his master's degree in computer science from California State University of
Sacramento (CSUS), Sacramento, California, USA in 1995. In 1991 he earned his
bachelor's degree in computer engineering from University of the Pacific (UOP), Stockton,
California, USA. His main interest is Knowledge Based Systems and Artificial Intelligence
and machine learning.