SlideShare a Scribd company logo
1 of 8
Download to read offline
IOSR Journal of Computer Engineering (IOSR-JCE)
e-ISSN: 2278-0661,p-ISSN: 2278-8727, Volume 17, Issue 3, Ver. VII (May – Jun. 2015), PP 39-46
www.iosrjournals.org
DOI: 10.9790/0661-17373946 www.iosrjournals.org 39 | Page
Semi-Supervised Discriminant Analysis Based On Data Structure
Xuesong Yin*
, Rongrong Jiang, Lifei Jiang
Department of Computer Science & Technology, Zhejiang Radio & TV University, Hangzhou, 310030, China
Abstract: Dimensionality reduction is a key data-analytic technique for mining high-dimensional data. In this
paper, we consider a general problem of learning from pairwise constraints in the form of must-link and cannot-
link. As one kind of side information, the must-link constraints imply that a pair of instances belongs to the same
class, while the cannot-link constraints compel them to be different classes. Given must-link and cannot-link
information, the goal of this paper is learn a smooth and discriminative subspace. Specifically, in order to
generate such a subspace, we use pairwise constraints to present an optimization problem, in which a least
squares formulation that integrates both global and local structures is considered as a regularization term for
dimensionality reduction. Experimental results on benchmark data sets show the effectiveness of the proposed
algorithm.
Keywords: Dimensionality reduction; Pairwise constraints; High-dimensional data; Discriminant analysis.
I. Introduction
Applications in various domains such as face recognition, web mining and image retrieval often lead to
very high-dimensional data. Mining and understanding such high-dimensional data is a contemporary challenge
due to the curse of dimensionality. However, the current researches [3,4,5,6,13] show that reducing the
dimensionality of data has been a feasible method for more precisely charactering such data. Therefore,
dimensionality reduction (DR for short) makes sense in many practical problems [2,3,5] since it allows one to
represent the data in a lower dimensional space. Based on the availability of the side information,
dimensionality reduction algorithms fall into three categories: supervised DR [7,12,28,35], semi-supervised DR
[8,16,18,21,22] and unsupervised DR [10,13,14,15,34]. In this paper, we focus on the case of semi-supervised
DR. With few constraints or class label information, existing semi-supervised DR algorithms appeal to
projecting the observed data onto a low-dimensional manifold, where the margin between data form different
classes is maximized. Most algorithms in this category, such as Locality Sensitive Discriminant Analysis [20],
Locality Sensitive Semi-supervised Feature Selection [36], and Semi-supervised Discriminant Analysis [26],
greatly take the intrinsic geometric structure into account, i.e. local structure of the data, indicating that nearby
instances are likely to have the same label. However, global structure of the data, implying that instances on the
same structure (typically referred to as a cluster or a manifold) are like to have the same label [1], is ignored in
those algorithms. In fact, when mining the data, global structure is the same significance as local structure
[1,10,11,29]. Therefore, it is necessary to integrate both global and local structures into the process of DR. The
key challenge is how we can incorporate relatively few pairwise constraints into DR such that the represented
data can still capture the available class information.
To this end, we propose a semi-supervised discriminant analysis algorithm with integrating both global
and local structures (SGLS), which naturally address DR under semi-supervised settings. Given pairwise
constraints, the key idea in SGLS is to integrate both global and local structures in semi-supervised discriminant
framework so that both discriminant and geometric structures of the data can be accurately captured. The SGLS
algorithm has the same flavor as not only supervised DR algorithms, which try to adjust the distance among
instances to improve the separability of the data, but unsupervised DR algorithms as well, which make nearby
instances in the original space close to each other in the embedding space.
The remainder of the paper is organized as follows. In Section 2, related work on the existing
algorithms of semi-supervised DR is discussed. In Section 3, we introduce the general framework of our
proposed SGLS algorithm. Section 4 discusses the experimental results on a number of real-world data sets.
Finally, we draw conclusions in Section 5.
II. Related Work
Unsupervised DR has attracted considerable interest in the last decade [9,19,23,25,29]. Recently, a few
researchers [5,9,13,17] have considered the case that discovers the local geometrical structure of the data
manifold when the data lie in a low-dimensional space of the high-dimensional ambient space. Algorithms
exploited along this line are used to estimate geometrical and topological properties of the embedding space,
*Corresponding author: Xuesong Yin, yinxs@nuaa.edu.cn
Semi-supervised Discriminant Analysis based on Data Structure
DOI: 10.9790/0661-17373946 www.iosrjournals.org 40 | Page
such as Locality Preserving Projections (LPP) [34], Local Discriminant Embedding (LDE) [35], Marginal Fisher
Analysis (MFA) [37] and Transductive Component Analysis (TCA) [27]. Different with typical unsupervised
DR algorithms, more lately, K-means was applied in the dimension-reduced space for avoiding the curse of
dimensionality in several algorithms [25,31] which perform clustering and dimensionality reduction
simultaneously.
Linear Discriminant Analysis (LDA) [7,12], capturing the global geometric structure of the data by
simultaneously maximizing the between-class distance and minimizing the within-class distance, is a well-
known approach for supervised DR. It has been used widely in various applications, such as face recognition
[5]. However, a major shortcoming of LDA is that it fails to discover the local geometrical structure of the data
manifold.
Semi-supervised DR can be considered as a new issue in DR area, which learns from a combination of
both the side information and unlabeled data. In general, the side information can be expressed in diverse forms,
such as class labels and pairwise constraints. Pairwise constraints include both must-link form in which the
instances belong to the same class and cannot-link form in which the instances belong to different classes. Some
semi-supervised DR algorithms [20,26,27,30,36,37] just applied partly labeled instances to maximize the margin
between instances form different classes for improving performance. However, in many practical data mining
applications, it is fairly expensive to label instances. In contrast, it could be easier for an expert or a user to
specify whether some pairs of instances belong to the same class or not. Moreover, the pairwise constraints can
be derived from labeled data but not vice versa. Therefore, making use of pairwise constraints to reduce the
dimensionality of the data has been an important topic [8,16,33].
III. The SGLS Algorithm
3.1 The Integrating Global and Local structures Framework
Given a set of instances 1, , nx x in
N
R , we can use a k-nearest neighbor graph to model the relationship
between nearby data points. Specifically, we put an edge between nodes i and j if ix and jx are close, i.e. ix and
jx are among k nearest neighbors of each other. Let the corresponding weight matrix be S, defined by
2
2
exp( ) ( ) ( )
0
i j
i k j j k i
ij
x x
x N x or x N xS t
otherwise
 
    


(1)
Where ( )k iN x denotes the set of k nearest neighbors of ix . In general, if a weighted value ijS is large, ix and jx
are close to each other in the original space. Thus, in order to preserve the locality structure, the embedding
space obtained should be as smooth as possible. To learn an appropriate representation y ( 1[ , , ]T
ny y y  )
with the locality structure, it is common to minimize the following objective function [13,34]:
2
,
2 T
i j ij
i j
y y S y Ly  (2)
where D is a diagonal matrix whose diagonal entries are column (or row, since S is symmetric) sum of S, that is
1
n
ii ijj
D S
  . L D S  is the Laplacian matrix.
On the other hand, we expect that the projections{ }T
ia x generated by a projection vector
d
a R , which span
the final embedding space, approach the sufficiently smooth embeddings{ }iy as much as possible. This
expectation can be satisfied by the least squares formulation which can incorporate the local structure
information via a regularization term defined as in Eq. (2). Mathematically, a projection vector a can be acquired
by solving the following optimization problem [27]:
2
1min ( , ) T T
J a y X a y y Ly   (3)
where 1 0  is the trade-off parameter. Here we assume that the data matrix X has been centered.
Taking the derivative of J(a,y) with respect to y and setting it equals to zero, we get
Semi-supervised Discriminant Analysis based on Data Structure
DOI: 10.9790/0661-17373946 www.iosrjournals.org 41 | Page
1
( , )
2( ) 0TJ a y
y X a Ly
y


   

(4)
It follows that
1
1* ( ) T
y I L X a 
  (5)
where I is the identity matrix of size n.
Applying Eq. (5) to Eq. (3), we can eliminate y and obtain the following optimization problem with regard to
a:
1 1
1
1 1
min ( ) ( , *) * ( )( ) *
( ) ( )
T
T T
T T
J a J a y y L I L y
a X I L L X a
a XGX a
 
 
  
 

(6)
where G is a positive semi-definite matrix of size n by n and has the following expression:
1
1 1( ) ( )G I L L 
  (7)
We can observe from Eq. (6) that minimizing J will bring the smooth subspace where both global and local
structures of data are calculated with regard to a.
3.2 Discriminat Analysis With Pairwise Constraints
Suppose we are given two sets of pairwise constraints including must-link (M) and cannot-link (C). In
the previous work [32,38,40,41], such constraints were used for learning an adaptive metric between the
prototypes of instances. However, recent research has shown that the distance metric learned on high-
dimensional space is relatively unreliable [33,39]. Instead of using constraint-guided metric learning, in this
paper our goal is to use pairwise constraints to find a projection vector such that in the embedding space the
distance between any pair of instances involved in the cannot-link constraints is maximized while that between
any pair of instances involved must-link constraints is minimized.
If the projection vector a is acquired, we can introduce such a transformation:
T
i iy a x (8)
Under this transformation, we calculate the following sum of the squared distances of the point pairs in M:
( , )
( ) ( )
i j
T T T T T
w i j i j
x x M
T
w
d a x a x a x a x
a S a

  


(9)
where wS is the covariance matrix of the point pairs in M:
( , )
( )( )
i j
T
w i j i j
x x M
S x x x x

   (10)
Correspondingly, for the point pairs in C, we have
T
b bd a S a (11)
where bS has the following expression form:
( , )
( )( )
i j
T
b i j i j
x x C
S x x x x

   (12)
Similar to LDA that seeks directions on which the data points of different classes are far from each other
while requiring data points of the same class to be close to each other, we try to maximize bd and minimize wd .
Thus, we can optimize the following problem to obtain the projection direction:
max
T
b
Ta
w
a S a
a S a
(13)
Semi-supervised Discriminant Analysis based on Data Structure
DOI: 10.9790/0661-17373946 www.iosrjournals.org 42 | Page
3.3 The Optimization Framework
So far, we pursue the projection direction a from two formulations: (1) we introduce a least squares
formulation for finding a, which facilities the integration of global and local structures; (2) we obtain the
projection vector by maximizing the cannot-link covariance and simultaneously minimizing the must-link
covariance. When the first formulation is considered as a regularization term, we arrive at the following
optimization problem:
2 2
max
(1 ) ( )
T
b
Ta
w
a S a
a S a J a  
(14)
where 20 1  controls the smoothness of estimator. The optimal projection vector a that maximizes the
objective function (14) is acquired by the maximum eigenvalue solution to the following generalized eigenvalue
problem.
2 2((1 ) )T
b wS a S XGX a     (15)
In real-world applications, we usually need more projection vectors to span a subspace rather than just one.
Let the matrix charactering such a target subspace be 1[ , , ]dA a a  which is formed by the eigenvectors
associated with the d largest eigenvalues of Eq. (15). Also, it is worth noting that since our method is linear, it
has a direct out-of-sample extension to novel sample x, i.e.
T
A x .
Based on the analysis above, we propose to develop a semi-supervised dimensionality reduction algorithm
which integrates both global and local structures defined in Eq. (14). The corresponding algorithm, called SGLS
is presented in Algorithm 1.
Algorithm 1
Input: X, M, C.
Output: A projection matrix
N d
A R 
 .
1 Construct an unsupervised graph on all n instances to calculate S as in Eq. (2) and L = D – S.
2 Present a least squares formulation together with the local structure information to compute matrix G as in
Eq. (7).
3 Compute wS and bS as in Eqs. (10) and (12).
4 Compute a generalized eigenvalue problem Eq. (15) and let the solved eigenvectors be 1, , da a in an
order of decreasing eigenvalues.
5 Output 1[ , , ]dA a a  .
To get a stable solution of the eigen-problem in Eq. (15), the matrix is required to be non-singular
which is not true when the number of features is larger than the number of instances, i.e. N > n. In such case, we
can apply the Tikhonov regularization technique [24,26] to improve the estimation. Thus, the generalized eigen-
problem in this paper can be rewritten as:
2 2((1 ) )T
b w NS a S XGX I a       (16)
where NI is the identity matrix of size N and 0  is a regularization parameter.
IV. Experiments
In this section, we study the performance of our SGLS algorithm in terms of clustering accuracy on
several real-world data sets from different application domains. We set k = 5 and t = 1 for the construction of the
similarity matrix S defined in Eq. (1).
4.1 Experimental Setup
In order to verify the efficacy of the proposed algorithm, we compare it with the state-of-the-art, semi-
supervised dimensionality reduction algorithms including SSDR [22], SCREEN [33], LADKM [31] and LMDM
[16]. We implemented SGLS in the MATLAB environment. All experiments were conducted on a PENTIUM
DUAL 3G PC with 2.0 GB RAM. We compared all algorithms on seven benchmark data sets, including
Segment and Digits from UCI machine Learning Repository, three document data sets DOC1, DOC2 and DOC3
from the 20-Newgroups data, and two image data sets : USPS handwritten data and Yale Face B (YALEB). The
statistics of all data sets are summarized in Table 1.
Semi-supervised Discriminant Analysis based on Data Structure
DOI: 10.9790/0661-17373946 www.iosrjournals.org 43 | Page
Table 1: Summary of the test datasets used in our experiment
Data sets Instance Dimension Class
Segment 2309 19 10
Digits 6300 16 10
DOC1 4000 4985 6
DOC2 5200 3132 5
DOC3 6500 6466 9
USPS 7122 256 10
YALEB 3110 2500 10
As the unified clustering platform, we use K-means algorithm as the clustering method to compare
several algorithms after dimensionality reduction. For each data set, we ran different algorithms for 20 times and
the comparison was based on the average performance. The parameter in SGLS are always set to 1 10  and
2 0.8  if without extra explanation. Moreover, for the fair comparison, the dimensionality of all the
algorithms is set to K-1 (K is the number of clustering).
To evaluate the clustering performance, in this paper, we employ normalized mutual information (NMI) as the
clustering validation measure, which is widely used to verify the performance of clustering algorithms [8,25,33].
Then, NMI is defined as
( , )
( ) ( )
I X Y
NMI
H X H Y
 (17)
where X and Y denote the random variables of cluster memberships from the ground truth and the output of
clustering algorithm, respectively. ( , )I X Y is the mutual information between X and Y. ( )H X and ( )H Y are
the entropies of X and Y respectively.
4.2 Experimental Results
Semi-supervised Discriminant Analysis based on Data Structure
DOI: 10.9790/0661-17373946 www.iosrjournals.org 44 | Page
Figure 1 Clustering performance on 6 data sets with different number of constraints
Table 2 presents the NMI results of various algorithms on all seven data sets. Also, Figure 1 shows the
clustering performance of standard K-means applied to the projected data by different dimensionality reduction
algorithms with different numbers of pairwise constraints. As can be seen, our SGLS algorithm achieves the best
performance on Digits, DOC1, DOC3, USPS, and YALEB data sets. In fact, the performance of our algorithm is
always comparable to that of the other algorithm on the rest of all seven data sets used in this paper. Specifically,
we can obtain the following main observations from Table 2 and Figure 1:
(1) The SGLS algorithm significantly outperforms not only LMDM and LDAKM on all of the seven data sets
used in the experiment, but SCREEN and SSDR on the six data sets as well. This can be contributed to such
a fact that integrating both global and local structures in SGLS may be beneficial.
(2) It is interesting to note that SGLS slightly underperform both SSDR on Segment and SCREEN on DOC2.
This implies that in certain case such as Segment, global and local structures may capture similar
information so that integrating such both structures seems to hardly help in the dimensionality reduction
process.
(3) It is clear from the presented results that LDAKM performs relatively worse that the other algorithms. This
can be explained that LDAKM does not employ any side information to pursue the projection direction,
which only makes use of the abundant unlabeled instances. Thus, LDAKM has a greatly inferior
performance.
(4) Due to the fact that SCREEN and LMDM just pairwise constraints to learn an optimal subspace, their
performances are worse that those of SSDR and SGLS respectively. In fact, a large number of the unlabeled
data play an important role for semi-supervised dimensionality reduction [8,26,27].
(5) Since SSDR can apply both the pairwise constraints and the unlabeled data to learn a subspace, it can
provide a relatively satisfying clustering performance in comparison with LDAKM, SCREEN and LMDM.
Compared with our SGLS, however, SSDR has still provided the fairly inferior performance. Actually,
SSDR can be considered as the constrained Principal Component Analysis (PCA). Thus, it fails to preserve
local neighborhood structures of the instances in the reduced low-dimensional space.
Table 2: NMI comparisons on seven data sets
Data sets LDAKM SCREEN SSDR LMDM SGLS
Segment 0.687 0.723 0.744 0.712 0.737
Digits 0.714 0.755 0.767 0.750 0.788
DOC1 0.567 0.596 0.604 0.588 0.625
DOC2 0.502 0.561 0.547 0.539 0.556
DOC3 0.746 0.802 0.810 0.797 0.823
USPS 0.636 0.675 0.682 0.669 0.705
YALEB 0.811 0.848 0.837 0.833 0.861
V. Conclusions
In this paper, we propose a novel semi-supervised dimensionality reduction algorithm, called SGLS.
For dimensionality reduction and clustering, SGLS integrates both global and local geometric structures so that
it can be naturally extended to deal with the abundant labeled data, as graph Laplacian is defined on all the
instances. Compared with the state-of-the-art dimensionality reduction algorithms on a collection of benchmark
data sets in terms of clustering accuracy, SGLS can always achieve a clear performance gain. Extensive
experiments exhibit the advantages of the novel semi-supervised clustering method of SGLS+K-means.
Semi-supervised Discriminant Analysis based on Data Structure
DOI: 10.9790/0661-17373946 www.iosrjournals.org 45 | Page
Acknowledgments
The research reported in this paper has been partially supported by Natural Science Foundation of
Zhejiang Province under Grant No. Y1100349, Public Project of Zhejiang Province under Grant No.
2013C33087, Academic Climbing Project of Young Academic Leaders of Zhejiang Province No. pd2013446.
References
[1]. D. Zhou, O. Bousquet, T.N. Lal, J. Weston, and B. Scholkopf. Learning with local and global consistency. In Advances in Neural
Information Processing Systems 16, 2003, pp. 321-328.
[2]. P. Garcia-Laencina, G. Rodriguez-Bermudez, J. Roca-Dorda. Exploring dimensionality reduction of EEG features in motor imagery
task classification, Expert Systems with Applications, 41 (11) (2014)5285-5295.
[3]. Y. Lin, T. Liu, C. Fuh Dimensionality Reduction for Data in Multiple Feature Representations. In Advances in Neural Information
Processing Systems 21, 2008, pp. 961-968.
[4]. X. Zhu, Z. Ghahramani, and J. Lafferty. Semi-Supervised Learning using Gaussian Fields and Harmonic Functions, Proc. Twentieth
Int’l Conf. Machine Learning, 2003, pp. 912-919.
[5]. X. He, S. Yan, Y. Hu, P. Niyogi, H. Zhang, Face recognition using laplacianfaces, IEEE Trans. Pattern Anal. Mach. Intell. 27 (3)
(2005) 328-340.
[6]. S. Roweis, L. Saul. Nonlinear dimensionality reduction by locally linear embedding, Science, 290 (5500) (2000) 2323-2326.
[7]. R. O. Duda, P. E. Hart, and D. G. Stork. Pattern Classification. Wiley-Interscience, Hoboken, NJ, 2nd edition, 2000.
[8]. X. Yin, T. Shu, Q. Huang. Semi-supervised fuzzy clustering with metric learning and entropy regularization. Knowledge-Based
Systems, 2012, 35(11):304-311.
[9]. M. Belkin and P. Niyogi. Laplacian Eigenmaps for Dimensionality Reduction and Data Representation. Neural Computation, 15(6)
(2003) 1373-1396.
[10]. J. Tenenbaum, V. de Silva, and J. Langford. A global geometric framework for nonlinear dimensionality reduction. Science,
290(5500) (2000) 2319-2323.
[11]. F. Tsai. Dimensionality reduction techniques for blog visualization, Expert Systems with Applications, 38(3) (2011) 2766-2773.
[12]. J. Yang, A.F. Frangi, J.Y. Yang, D. Zhang, Z. Jin, KPCA plus LDA: A complete kernel fisher discriminant framework for feature
extraction and recognition, IEEE Trans. Pattern Anal. Mach. Intell. 27 (2) (2005) 230-244.
[13]. M. Belkin, P. Niyogi, Laplacian eigenmaps and spectral techniques for embedding and clustering, in: Advances in Neural
Information Processing Systems 14, 2001, pp. 585-591.
[14]. D. Cai, X. He, J. Han, Document clustering using locality preserving indexing, IEEE Trans. Knowl. Data Eng. 17 (12) (2005) 1624-
1637.
[15]. J. He, L. Ding, L. Jiang, Z. Li, Q. Hu. Intrinsic dimensionality estimation based on manifold assumption, Journal of Visual
Communication and Image Representation, 25( 5)(2014) 740-747.
[16]. S. Xiang, F. Nie, and C. Zhang. Learning a Mahalanobis distance metric for data clustering and classification. Pattern Recognition,
41(12) (2008) 3600 - 3612.
[17]. D. Cai, X. He, J. Han, H.-J. Zhang, Orthogonal laplacianfaces for face recognition, IEEE Trans. Image Process. 15 (11) (2006)
3608-3614.
[18]. Z. Zhang, M. Zhao, T. Chow. Marginal semi-supervised sub-manifold projections with informative constraints for dimensionality
reduction and recognition, Neural Networks, 36(2012) 97-111.
[19]. J. Chen, J. Ye, and Q. Li. Integrating Global and Local Structures: A Least Squares Framework for Dimensionality Reduction, in:
IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2007, pp. 1-8.
[20]. D. Cai, X. He, K. Zhou, J. Han, H. Bao, Locality sensitive discriminant analysis, in: International Joint Conference on Artificial
Intelligence, 2007, pp. 708-713.
[21]. W. Du, K. Inoue, K. Urahama, Dimensionality reduction for semi-supervised face recognition, in: Proceedings of the Fuzzy
Systems and Knowledge Discovery, 2005, pp. 1-10.
[22]. D. Zhang, Z.-H. Zhou, S. Chen, Semi-supervised dimensionality reduction, in: SIAM Conference on Data Mining (SDM), 2007, pp.
629-634.
[23]. X. Yang, H. Fu, H. Zha, J. Barlow, Semi-supervised nonlinear dimensionality reduction, in: Proceedings of the International
Conference on Machine Learning, 2006, pp. 1065-1072.
[24]. M. Belkin, P. Niyogi, V. Sindhwani, Manifold regularization: a geometric framework for learning from labeled and unlabeled
examples, J. Mach. Learning Res. 1 (1) (2006) 1-48.
[25]. J. Ye, Z. Zhao, H. Liu. Adaptive distance metric learning for clustering. In: Proc. of the Conference on Computer Vision and Pattern
Recognition, 2007, pp. 1-7.
[26]. D. Cai, X. He, J. Han, Semi-supervised discriminant analysis, in: Proceedings of the International Conference on Computer Vision,
2007, pp. 1-7.
[27]. W. Liu, D. Tao, and J. Liu. Transductive Component Analysis, in: Proceedings of the International Conference on Data Mining,
2008, pp. 433-442.
[28]. P. Han, A. Jin, A. Kar. Kernel Discriminant Embedding in face recognition, Journal of Visual Communication and Image
Representation, 22(7) (2011) 634-642.
[29]. J. Ye. Least Squares Linear Discriminant Analysis, in: Proceedings of International Conference on Machine Learning, 2007, pp.
1087-1093.
[30]. X. Chen, S. Chen, H. Xue, X. Zhou. A unified dimensionality reduction framework for semi-paired and semi-supervised multi-view
data, Pattern Recognition, 45(5) (2012) 2005-2018.
[31]. C. Ding, T. Li. Adaptive dimension reduction using discriminant analysis and K-means clustering. In: Proceedings of the
International Conference on Machine Learning, 2007, pp. 521-528.
[32]. E. Xing, A. Ng, M. Jordan, S. Russell. Distance metric learning with application to clustering with side-information. In: Advances
in Neural Information Processing Systems 15, 2003, pp. 505-512.
[33]. W. Tang, H. Xiong, S. Zhong, J. Wu. Enhancing semi-supervised clustering: a feature projection perspective. In: Proceedings of the
International Conference on Knowledge Discovery and Data Mining, 2007, pp. 707-716.
[34]. X. He, P. Niyogi. Locality preserving projections, In: Advances in Neural Information Processing Systems 16, 2003.
[35]. H. Chen, H. Chang, and T. Liu. Local discriminant embedding and its variants. In: Proceedings of the Internal Conference on
Computer Vision and Pattern Recognition, 2005, pp.846-853.
[36]. J. Zhao, K. Lu, X. He, Locality sensitive semi-supervised feature selection, Neurocomputing, 71 (10-12), (2008)1842-1849.
Semi-supervised Discriminant Analysis based on Data Structure
DOI: 10.9790/0661-17373946 www.iosrjournals.org 46 | Page
[37]. S. Yan, D. Xu, B. Zhang, H.J. Zhang, Q. Yang, and S. Lin. Graph embedding and extension: A general framework for
dimensionality reduction. IEEE Transactions on PAMI, 29(1) (2007) 40-51.
[38]. B. Kulis, S. Basu, I. Dhillon, and R. Mooney. Semi-supervised graph clustering: a kernel approach. In: Proceedings of the
International Conference on Machine Learning, 2005, pp.59-68.
[39]. D. Meng, Y. Leung, Z. Xu. Passage method for nonlinear dimensionality reduction of data on multi-cluster manifolds, Pattern
Recognition, 46 (8) (2013) 2175-2186.
[40]. M. Bilenko, S. Basu, R.J. Mooney, Integrating constraints and metric learning in semi-supervised clustering, in: Proceedings of
International Conference on Machine Learning, 2004, pp. 81–88.
[41]. D.Y. Yeung, H. Chang, Extending the relevant component analysis algorithm for metric learning using both positive and negative
equivalence constraints, Pattern Recognition 39 (5) (2006) 1007–1010.

More Related Content

What's hot

A Novel Multi- Viewpoint based Similarity Measure for Document Clustering
A Novel Multi- Viewpoint based Similarity Measure for Document ClusteringA Novel Multi- Viewpoint based Similarity Measure for Document Clustering
A Novel Multi- Viewpoint based Similarity Measure for Document ClusteringIJMER
 
Mammogram image segmentation using rough
Mammogram image segmentation using roughMammogram image segmentation using rough
Mammogram image segmentation using rougheSAT Publishing House
 
A Novel Algorithm for Design Tree Classification with PCA
A Novel Algorithm for Design Tree Classification with PCAA Novel Algorithm for Design Tree Classification with PCA
A Novel Algorithm for Design Tree Classification with PCAEditor Jacotech
 
Survey on Unsupervised Learning in Datamining
Survey on Unsupervised Learning in DataminingSurvey on Unsupervised Learning in Datamining
Survey on Unsupervised Learning in DataminingIOSR Journals
 
A STUDY ON SIMILARITY MEASURE FUNCTIONS ON ENGINEERING MATERIALS SELECTION
A STUDY ON SIMILARITY MEASURE FUNCTIONS ON ENGINEERING MATERIALS SELECTION A STUDY ON SIMILARITY MEASURE FUNCTIONS ON ENGINEERING MATERIALS SELECTION
A STUDY ON SIMILARITY MEASURE FUNCTIONS ON ENGINEERING MATERIALS SELECTION cscpconf
 
Sparse representation based classification of mr images of brain for alzheime...
Sparse representation based classification of mr images of brain for alzheime...Sparse representation based classification of mr images of brain for alzheime...
Sparse representation based classification of mr images of brain for alzheime...eSAT Publishing House
 
A Review on Non Linear Dimensionality Reduction Techniques for Face Recognition
A Review on Non Linear Dimensionality Reduction Techniques for Face RecognitionA Review on Non Linear Dimensionality Reduction Techniques for Face Recognition
A Review on Non Linear Dimensionality Reduction Techniques for Face Recognitionrahulmonikasharma
 
MINING TRIADIC ASSOCIATION RULES
MINING TRIADIC ASSOCIATION RULESMINING TRIADIC ASSOCIATION RULES
MINING TRIADIC ASSOCIATION RULEScsandit
 
International Refereed Journal of Engineering and Science (IRJES)
International Refereed Journal of Engineering and Science (IRJES)International Refereed Journal of Engineering and Science (IRJES)
International Refereed Journal of Engineering and Science (IRJES)irjes
 
Mammogram image segmentation using rough clustering
Mammogram image segmentation using rough clusteringMammogram image segmentation using rough clustering
Mammogram image segmentation using rough clusteringeSAT Journals
 
A Combined Approach for Feature Subset Selection and Size Reduction for High ...
A Combined Approach for Feature Subset Selection and Size Reduction for High ...A Combined Approach for Feature Subset Selection and Size Reduction for High ...
A Combined Approach for Feature Subset Selection and Size Reduction for High ...IJERA Editor
 
Multiview Alignment Hashing for Efficient Image Search
Multiview Alignment Hashing for Efficient Image SearchMultiview Alignment Hashing for Efficient Image Search
Multiview Alignment Hashing for Efficient Image Search1crore projects
 
Finding similarities between structured documents as a crucial stage for gene...
Finding similarities between structured documents as a crucial stage for gene...Finding similarities between structured documents as a crucial stage for gene...
Finding similarities between structured documents as a crucial stage for gene...Alexander Decker
 
Compressing the dependent elements of multiset
Compressing the dependent elements of multisetCompressing the dependent elements of multiset
Compressing the dependent elements of multisetIRJET Journal
 
Volume 2-issue-6-1930-1932
Volume 2-issue-6-1930-1932Volume 2-issue-6-1930-1932
Volume 2-issue-6-1930-1932Editor IJARCET
 

What's hot (17)

A Novel Multi- Viewpoint based Similarity Measure for Document Clustering
A Novel Multi- Viewpoint based Similarity Measure for Document ClusteringA Novel Multi- Viewpoint based Similarity Measure for Document Clustering
A Novel Multi- Viewpoint based Similarity Measure for Document Clustering
 
Mammogram image segmentation using rough
Mammogram image segmentation using roughMammogram image segmentation using rough
Mammogram image segmentation using rough
 
A Novel Algorithm for Design Tree Classification with PCA
A Novel Algorithm for Design Tree Classification with PCAA Novel Algorithm for Design Tree Classification with PCA
A Novel Algorithm for Design Tree Classification with PCA
 
Lattice2 tree
Lattice2 treeLattice2 tree
Lattice2 tree
 
Survey on Unsupervised Learning in Datamining
Survey on Unsupervised Learning in DataminingSurvey on Unsupervised Learning in Datamining
Survey on Unsupervised Learning in Datamining
 
A STUDY ON SIMILARITY MEASURE FUNCTIONS ON ENGINEERING MATERIALS SELECTION
A STUDY ON SIMILARITY MEASURE FUNCTIONS ON ENGINEERING MATERIALS SELECTION A STUDY ON SIMILARITY MEASURE FUNCTIONS ON ENGINEERING MATERIALS SELECTION
A STUDY ON SIMILARITY MEASURE FUNCTIONS ON ENGINEERING MATERIALS SELECTION
 
Sparse representation based classification of mr images of brain for alzheime...
Sparse representation based classification of mr images of brain for alzheime...Sparse representation based classification of mr images of brain for alzheime...
Sparse representation based classification of mr images of brain for alzheime...
 
A Review on Non Linear Dimensionality Reduction Techniques for Face Recognition
A Review on Non Linear Dimensionality Reduction Techniques for Face RecognitionA Review on Non Linear Dimensionality Reduction Techniques for Face Recognition
A Review on Non Linear Dimensionality Reduction Techniques for Face Recognition
 
MINING TRIADIC ASSOCIATION RULES
MINING TRIADIC ASSOCIATION RULESMINING TRIADIC ASSOCIATION RULES
MINING TRIADIC ASSOCIATION RULES
 
International Refereed Journal of Engineering and Science (IRJES)
International Refereed Journal of Engineering and Science (IRJES)International Refereed Journal of Engineering and Science (IRJES)
International Refereed Journal of Engineering and Science (IRJES)
 
Mammogram image segmentation using rough clustering
Mammogram image segmentation using rough clusteringMammogram image segmentation using rough clustering
Mammogram image segmentation using rough clustering
 
A Combined Approach for Feature Subset Selection and Size Reduction for High ...
A Combined Approach for Feature Subset Selection and Size Reduction for High ...A Combined Approach for Feature Subset Selection and Size Reduction for High ...
A Combined Approach for Feature Subset Selection and Size Reduction for High ...
 
K044055762
K044055762K044055762
K044055762
 
Multiview Alignment Hashing for Efficient Image Search
Multiview Alignment Hashing for Efficient Image SearchMultiview Alignment Hashing for Efficient Image Search
Multiview Alignment Hashing for Efficient Image Search
 
Finding similarities between structured documents as a crucial stage for gene...
Finding similarities between structured documents as a crucial stage for gene...Finding similarities between structured documents as a crucial stage for gene...
Finding similarities between structured documents as a crucial stage for gene...
 
Compressing the dependent elements of multiset
Compressing the dependent elements of multisetCompressing the dependent elements of multiset
Compressing the dependent elements of multiset
 
Volume 2-issue-6-1930-1932
Volume 2-issue-6-1930-1932Volume 2-issue-6-1930-1932
Volume 2-issue-6-1930-1932
 

Viewers also liked

Viewers also liked (20)

I017525560
I017525560I017525560
I017525560
 
Prevention of Selective Jamming Attacks by Using Packet Hiding Methods
Prevention of Selective Jamming Attacks by Using Packet Hiding MethodsPrevention of Selective Jamming Attacks by Using Packet Hiding Methods
Prevention of Selective Jamming Attacks by Using Packet Hiding Methods
 
D1802051622
D1802051622D1802051622
D1802051622
 
C1803011419
C1803011419C1803011419
C1803011419
 
B1802021823
B1802021823B1802021823
B1802021823
 
Study of the Tribological behavior of as cast Al-4.2%Cu-Al2O3 composite
Study of the Tribological behavior of as cast Al-4.2%Cu-Al2O3 compositeStudy of the Tribological behavior of as cast Al-4.2%Cu-Al2O3 composite
Study of the Tribological behavior of as cast Al-4.2%Cu-Al2O3 composite
 
Objective function to Optimize Bending Stress at Critical Section of Asymmetr...
Objective function to Optimize Bending Stress at Critical Section of Asymmetr...Objective function to Optimize Bending Stress at Critical Section of Asymmetr...
Objective function to Optimize Bending Stress at Critical Section of Asymmetr...
 
A013150107
A013150107A013150107
A013150107
 
A012470128
A012470128A012470128
A012470128
 
D012251726
D012251726D012251726
D012251726
 
I017355767
I017355767I017355767
I017355767
 
I010346366
I010346366I010346366
I010346366
 
R01232147152
R01232147152R01232147152
R01232147152
 
C017510717
C017510717C017510717
C017510717
 
H017365863
H017365863H017365863
H017365863
 
M018147883
M018147883M018147883
M018147883
 
K011117277
K011117277K011117277
K011117277
 
F017663344
F017663344F017663344
F017663344
 
N18030296105
N18030296105N18030296105
N18030296105
 
F017323439
F017323439F017323439
F017323439
 

Similar to E017373946

A Study in Employing Rough Set Based Approach for Clustering on Categorical ...
A Study in Employing Rough Set Based Approach for Clustering  on Categorical ...A Study in Employing Rough Set Based Approach for Clustering  on Categorical ...
A Study in Employing Rough Set Based Approach for Clustering on Categorical ...IOSR Journals
 
Spatio textual similarity join
Spatio textual similarity joinSpatio textual similarity join
Spatio textual similarity joinIJDKP
 
Improved probabilistic distance based locality preserving projections method ...
Improved probabilistic distance based locality preserving projections method ...Improved probabilistic distance based locality preserving projections method ...
Improved probabilistic distance based locality preserving projections method ...IJECEIAES
 
GRAPH BASED LOCAL RECODING FOR DATA ANONYMIZATION
GRAPH BASED LOCAL RECODING FOR DATA ANONYMIZATIONGRAPH BASED LOCAL RECODING FOR DATA ANONYMIZATION
GRAPH BASED LOCAL RECODING FOR DATA ANONYMIZATIONijdms
 
Pak eko 4412ijdms01
Pak eko 4412ijdms01Pak eko 4412ijdms01
Pak eko 4412ijdms01hyuviridvic
 
SVD BASED LATENT SEMANTIC INDEXING WITH USE OF THE GPU COMPUTATIONS
SVD BASED LATENT SEMANTIC INDEXING WITH USE OF THE GPU COMPUTATIONSSVD BASED LATENT SEMANTIC INDEXING WITH USE OF THE GPU COMPUTATIONS
SVD BASED LATENT SEMANTIC INDEXING WITH USE OF THE GPU COMPUTATIONSijscmcj
 
Ba2419551957
Ba2419551957Ba2419551957
Ba2419551957IJMER
 
Influence over the Dimensionality Reduction and Clustering for Air Quality Me...
Influence over the Dimensionality Reduction and Clustering for Air Quality Me...Influence over the Dimensionality Reduction and Clustering for Air Quality Me...
Influence over the Dimensionality Reduction and Clustering for Air Quality Me...IJAEMSJORNAL
 
A frame work for clustering time evolving data
A frame work for clustering time evolving dataA frame work for clustering time evolving data
A frame work for clustering time evolving dataiaemedu
 
Bat-Cluster: A Bat Algorithm-based Automated Graph Clustering Approach
Bat-Cluster: A Bat Algorithm-based Automated Graph Clustering Approach Bat-Cluster: A Bat Algorithm-based Automated Graph Clustering Approach
Bat-Cluster: A Bat Algorithm-based Automated Graph Clustering Approach IJECEIAES
 
Reduct generation for the incremental data using rough set theory
Reduct generation for the incremental data using rough set theoryReduct generation for the incremental data using rough set theory
Reduct generation for the incremental data using rough set theorycsandit
 
Data reduction techniques for high dimensional biological data
Data reduction techniques for high dimensional biological dataData reduction techniques for high dimensional biological data
Data reduction techniques for high dimensional biological dataeSAT Journals
 
IRJET- Customer Segmentation from Massive Customer Transaction Data
IRJET- Customer Segmentation from Massive Customer Transaction DataIRJET- Customer Segmentation from Massive Customer Transaction Data
IRJET- Customer Segmentation from Massive Customer Transaction DataIRJET Journal
 
Juha vesanto esa alhoniemi 2000:clustering of the som
Juha vesanto esa alhoniemi 2000:clustering of the somJuha vesanto esa alhoniemi 2000:clustering of the som
Juha vesanto esa alhoniemi 2000:clustering of the somArchiLab 7
 
A comparative study of clustering and biclustering of microarray data
A comparative study of clustering and biclustering of microarray dataA comparative study of clustering and biclustering of microarray data
A comparative study of clustering and biclustering of microarray dataijcsit
 
Literature Survey On Clustering Techniques
Literature Survey On Clustering TechniquesLiterature Survey On Clustering Techniques
Literature Survey On Clustering TechniquesIOSR Journals
 
Finding Relationships between the Our-NIR Cluster Results
Finding Relationships between the Our-NIR Cluster ResultsFinding Relationships between the Our-NIR Cluster Results
Finding Relationships between the Our-NIR Cluster ResultsCSCJournals
 

Similar to E017373946 (20)

1376846406 14447221
1376846406  144472211376846406  14447221
1376846406 14447221
 
A Study in Employing Rough Set Based Approach for Clustering on Categorical ...
A Study in Employing Rough Set Based Approach for Clustering  on Categorical ...A Study in Employing Rough Set Based Approach for Clustering  on Categorical ...
A Study in Employing Rough Set Based Approach for Clustering on Categorical ...
 
Spatio textual similarity join
Spatio textual similarity joinSpatio textual similarity join
Spatio textual similarity join
 
Improved probabilistic distance based locality preserving projections method ...
Improved probabilistic distance based locality preserving projections method ...Improved probabilistic distance based locality preserving projections method ...
Improved probabilistic distance based locality preserving projections method ...
 
GRAPH BASED LOCAL RECODING FOR DATA ANONYMIZATION
GRAPH BASED LOCAL RECODING FOR DATA ANONYMIZATIONGRAPH BASED LOCAL RECODING FOR DATA ANONYMIZATION
GRAPH BASED LOCAL RECODING FOR DATA ANONYMIZATION
 
Az36311316
Az36311316Az36311316
Az36311316
 
Pak eko 4412ijdms01
Pak eko 4412ijdms01Pak eko 4412ijdms01
Pak eko 4412ijdms01
 
SVD BASED LATENT SEMANTIC INDEXING WITH USE OF THE GPU COMPUTATIONS
SVD BASED LATENT SEMANTIC INDEXING WITH USE OF THE GPU COMPUTATIONSSVD BASED LATENT SEMANTIC INDEXING WITH USE OF THE GPU COMPUTATIONS
SVD BASED LATENT SEMANTIC INDEXING WITH USE OF THE GPU COMPUTATIONS
 
call for papers, research paper publishing, where to publish research paper, ...
call for papers, research paper publishing, where to publish research paper, ...call for papers, research paper publishing, where to publish research paper, ...
call for papers, research paper publishing, where to publish research paper, ...
 
Ba2419551957
Ba2419551957Ba2419551957
Ba2419551957
 
Influence over the Dimensionality Reduction and Clustering for Air Quality Me...
Influence over the Dimensionality Reduction and Clustering for Air Quality Me...Influence over the Dimensionality Reduction and Clustering for Air Quality Me...
Influence over the Dimensionality Reduction and Clustering for Air Quality Me...
 
A frame work for clustering time evolving data
A frame work for clustering time evolving dataA frame work for clustering time evolving data
A frame work for clustering time evolving data
 
Bat-Cluster: A Bat Algorithm-based Automated Graph Clustering Approach
Bat-Cluster: A Bat Algorithm-based Automated Graph Clustering Approach Bat-Cluster: A Bat Algorithm-based Automated Graph Clustering Approach
Bat-Cluster: A Bat Algorithm-based Automated Graph Clustering Approach
 
Reduct generation for the incremental data using rough set theory
Reduct generation for the incremental data using rough set theoryReduct generation for the incremental data using rough set theory
Reduct generation for the incremental data using rough set theory
 
Data reduction techniques for high dimensional biological data
Data reduction techniques for high dimensional biological dataData reduction techniques for high dimensional biological data
Data reduction techniques for high dimensional biological data
 
IRJET- Customer Segmentation from Massive Customer Transaction Data
IRJET- Customer Segmentation from Massive Customer Transaction DataIRJET- Customer Segmentation from Massive Customer Transaction Data
IRJET- Customer Segmentation from Massive Customer Transaction Data
 
Juha vesanto esa alhoniemi 2000:clustering of the som
Juha vesanto esa alhoniemi 2000:clustering of the somJuha vesanto esa alhoniemi 2000:clustering of the som
Juha vesanto esa alhoniemi 2000:clustering of the som
 
A comparative study of clustering and biclustering of microarray data
A comparative study of clustering and biclustering of microarray dataA comparative study of clustering and biclustering of microarray data
A comparative study of clustering and biclustering of microarray data
 
Literature Survey On Clustering Techniques
Literature Survey On Clustering TechniquesLiterature Survey On Clustering Techniques
Literature Survey On Clustering Techniques
 
Finding Relationships between the Our-NIR Cluster Results
Finding Relationships between the Our-NIR Cluster ResultsFinding Relationships between the Our-NIR Cluster Results
Finding Relationships between the Our-NIR Cluster Results
 

More from IOSR Journals (20)

A011140104
A011140104A011140104
A011140104
 
M0111397100
M0111397100M0111397100
M0111397100
 
L011138596
L011138596L011138596
L011138596
 
K011138084
K011138084K011138084
K011138084
 
J011137479
J011137479J011137479
J011137479
 
I011136673
I011136673I011136673
I011136673
 
G011134454
G011134454G011134454
G011134454
 
H011135565
H011135565H011135565
H011135565
 
F011134043
F011134043F011134043
F011134043
 
E011133639
E011133639E011133639
E011133639
 
D011132635
D011132635D011132635
D011132635
 
C011131925
C011131925C011131925
C011131925
 
B011130918
B011130918B011130918
B011130918
 
A011130108
A011130108A011130108
A011130108
 
I011125160
I011125160I011125160
I011125160
 
H011124050
H011124050H011124050
H011124050
 
G011123539
G011123539G011123539
G011123539
 
F011123134
F011123134F011123134
F011123134
 
E011122530
E011122530E011122530
E011122530
 
D011121524
D011121524D011121524
D011121524
 

Recently uploaded

Advanced Test Driven-Development @ php[tek] 2024
Advanced Test Driven-Development @ php[tek] 2024Advanced Test Driven-Development @ php[tek] 2024
Advanced Test Driven-Development @ php[tek] 2024Scott Keck-Warren
 
Beyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
Beyond Boundaries: Leveraging No-Code Solutions for Industry InnovationBeyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
Beyond Boundaries: Leveraging No-Code Solutions for Industry InnovationSafe Software
 
"Federated learning: out of reach no matter how close",Oleksandr Lapshyn
"Federated learning: out of reach no matter how close",Oleksandr Lapshyn"Federated learning: out of reach no matter how close",Oleksandr Lapshyn
"Federated learning: out of reach no matter how close",Oleksandr LapshynFwdays
 
Transcript: New from BookNet Canada for 2024: BNC BiblioShare - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: BNC BiblioShare - Tech Forum 2024Transcript: New from BookNet Canada for 2024: BNC BiblioShare - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: BNC BiblioShare - Tech Forum 2024BookNet Canada
 
Snow Chain-Integrated Tire for a Safe Drive on Winter Roads
Snow Chain-Integrated Tire for a Safe Drive on Winter RoadsSnow Chain-Integrated Tire for a Safe Drive on Winter Roads
Snow Chain-Integrated Tire for a Safe Drive on Winter RoadsHyundai Motor Group
 
Build your next Gen AI Breakthrough - April 2024
Build your next Gen AI Breakthrough - April 2024Build your next Gen AI Breakthrough - April 2024
Build your next Gen AI Breakthrough - April 2024Neo4j
 
Streamlining Python Development: A Guide to a Modern Project Setup
Streamlining Python Development: A Guide to a Modern Project SetupStreamlining Python Development: A Guide to a Modern Project Setup
Streamlining Python Development: A Guide to a Modern Project SetupFlorian Wilhelm
 
Enhancing Worker Digital Experience: A Hands-on Workshop for Partners
Enhancing Worker Digital Experience: A Hands-on Workshop for PartnersEnhancing Worker Digital Experience: A Hands-on Workshop for Partners
Enhancing Worker Digital Experience: A Hands-on Workshop for PartnersThousandEyes
 
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024BookNet Canada
 
Key Features Of Token Development (1).pptx
Key  Features Of Token  Development (1).pptxKey  Features Of Token  Development (1).pptx
Key Features Of Token Development (1).pptxLBM Solutions
 
SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024Scott Keck-Warren
 
Kotlin Multiplatform & Compose Multiplatform - Starter kit for pragmatics
Kotlin Multiplatform & Compose Multiplatform - Starter kit for pragmaticsKotlin Multiplatform & Compose Multiplatform - Starter kit for pragmatics
Kotlin Multiplatform & Compose Multiplatform - Starter kit for pragmaticscarlostorres15106
 
Bluetooth Controlled Car with Arduino.pdf
Bluetooth Controlled Car with Arduino.pdfBluetooth Controlled Car with Arduino.pdf
Bluetooth Controlled Car with Arduino.pdfngoud9212
 
Making_way_through_DLL_hollowing_inspite_of_CFG_by_Debjeet Banerjee.pptx
Making_way_through_DLL_hollowing_inspite_of_CFG_by_Debjeet Banerjee.pptxMaking_way_through_DLL_hollowing_inspite_of_CFG_by_Debjeet Banerjee.pptx
Making_way_through_DLL_hollowing_inspite_of_CFG_by_Debjeet Banerjee.pptxnull - The Open Security Community
 
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024BookNet Canada
 
Unleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubUnleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubKalema Edgar
 
My INSURER PTE LTD - Insurtech Innovation Award 2024
My INSURER PTE LTD - Insurtech Innovation Award 2024My INSURER PTE LTD - Insurtech Innovation Award 2024
My INSURER PTE LTD - Insurtech Innovation Award 2024The Digital Insurer
 
Connect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck PresentationConnect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck PresentationSlibray Presentation
 

Recently uploaded (20)

Advanced Test Driven-Development @ php[tek] 2024
Advanced Test Driven-Development @ php[tek] 2024Advanced Test Driven-Development @ php[tek] 2024
Advanced Test Driven-Development @ php[tek] 2024
 
Beyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
Beyond Boundaries: Leveraging No-Code Solutions for Industry InnovationBeyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
Beyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
 
"Federated learning: out of reach no matter how close",Oleksandr Lapshyn
"Federated learning: out of reach no matter how close",Oleksandr Lapshyn"Federated learning: out of reach no matter how close",Oleksandr Lapshyn
"Federated learning: out of reach no matter how close",Oleksandr Lapshyn
 
Transcript: New from BookNet Canada for 2024: BNC BiblioShare - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: BNC BiblioShare - Tech Forum 2024Transcript: New from BookNet Canada for 2024: BNC BiblioShare - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: BNC BiblioShare - Tech Forum 2024
 
Snow Chain-Integrated Tire for a Safe Drive on Winter Roads
Snow Chain-Integrated Tire for a Safe Drive on Winter RoadsSnow Chain-Integrated Tire for a Safe Drive on Winter Roads
Snow Chain-Integrated Tire for a Safe Drive on Winter Roads
 
Build your next Gen AI Breakthrough - April 2024
Build your next Gen AI Breakthrough - April 2024Build your next Gen AI Breakthrough - April 2024
Build your next Gen AI Breakthrough - April 2024
 
Streamlining Python Development: A Guide to a Modern Project Setup
Streamlining Python Development: A Guide to a Modern Project SetupStreamlining Python Development: A Guide to a Modern Project Setup
Streamlining Python Development: A Guide to a Modern Project Setup
 
Enhancing Worker Digital Experience: A Hands-on Workshop for Partners
Enhancing Worker Digital Experience: A Hands-on Workshop for PartnersEnhancing Worker Digital Experience: A Hands-on Workshop for Partners
Enhancing Worker Digital Experience: A Hands-on Workshop for Partners
 
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
 
DMCC Future of Trade Web3 - Special Edition
DMCC Future of Trade Web3 - Special EditionDMCC Future of Trade Web3 - Special Edition
DMCC Future of Trade Web3 - Special Edition
 
Key Features Of Token Development (1).pptx
Key  Features Of Token  Development (1).pptxKey  Features Of Token  Development (1).pptx
Key Features Of Token Development (1).pptx
 
SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024
 
Kotlin Multiplatform & Compose Multiplatform - Starter kit for pragmatics
Kotlin Multiplatform & Compose Multiplatform - Starter kit for pragmaticsKotlin Multiplatform & Compose Multiplatform - Starter kit for pragmatics
Kotlin Multiplatform & Compose Multiplatform - Starter kit for pragmatics
 
Bluetooth Controlled Car with Arduino.pdf
Bluetooth Controlled Car with Arduino.pdfBluetooth Controlled Car with Arduino.pdf
Bluetooth Controlled Car with Arduino.pdf
 
Making_way_through_DLL_hollowing_inspite_of_CFG_by_Debjeet Banerjee.pptx
Making_way_through_DLL_hollowing_inspite_of_CFG_by_Debjeet Banerjee.pptxMaking_way_through_DLL_hollowing_inspite_of_CFG_by_Debjeet Banerjee.pptx
Making_way_through_DLL_hollowing_inspite_of_CFG_by_Debjeet Banerjee.pptx
 
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
Transcript: #StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
 
Unleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubUnleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding Club
 
My INSURER PTE LTD - Insurtech Innovation Award 2024
My INSURER PTE LTD - Insurtech Innovation Award 2024My INSURER PTE LTD - Insurtech Innovation Award 2024
My INSURER PTE LTD - Insurtech Innovation Award 2024
 
E-Vehicle_Hacking_by_Parul Sharma_null_owasp.pptx
E-Vehicle_Hacking_by_Parul Sharma_null_owasp.pptxE-Vehicle_Hacking_by_Parul Sharma_null_owasp.pptx
E-Vehicle_Hacking_by_Parul Sharma_null_owasp.pptx
 
Connect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck PresentationConnect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck Presentation
 

E017373946

  • 1. IOSR Journal of Computer Engineering (IOSR-JCE) e-ISSN: 2278-0661,p-ISSN: 2278-8727, Volume 17, Issue 3, Ver. VII (May – Jun. 2015), PP 39-46 www.iosrjournals.org DOI: 10.9790/0661-17373946 www.iosrjournals.org 39 | Page Semi-Supervised Discriminant Analysis Based On Data Structure Xuesong Yin* , Rongrong Jiang, Lifei Jiang Department of Computer Science & Technology, Zhejiang Radio & TV University, Hangzhou, 310030, China Abstract: Dimensionality reduction is a key data-analytic technique for mining high-dimensional data. In this paper, we consider a general problem of learning from pairwise constraints in the form of must-link and cannot- link. As one kind of side information, the must-link constraints imply that a pair of instances belongs to the same class, while the cannot-link constraints compel them to be different classes. Given must-link and cannot-link information, the goal of this paper is learn a smooth and discriminative subspace. Specifically, in order to generate such a subspace, we use pairwise constraints to present an optimization problem, in which a least squares formulation that integrates both global and local structures is considered as a regularization term for dimensionality reduction. Experimental results on benchmark data sets show the effectiveness of the proposed algorithm. Keywords: Dimensionality reduction; Pairwise constraints; High-dimensional data; Discriminant analysis. I. Introduction Applications in various domains such as face recognition, web mining and image retrieval often lead to very high-dimensional data. Mining and understanding such high-dimensional data is a contemporary challenge due to the curse of dimensionality. However, the current researches [3,4,5,6,13] show that reducing the dimensionality of data has been a feasible method for more precisely charactering such data. Therefore, dimensionality reduction (DR for short) makes sense in many practical problems [2,3,5] since it allows one to represent the data in a lower dimensional space. Based on the availability of the side information, dimensionality reduction algorithms fall into three categories: supervised DR [7,12,28,35], semi-supervised DR [8,16,18,21,22] and unsupervised DR [10,13,14,15,34]. In this paper, we focus on the case of semi-supervised DR. With few constraints or class label information, existing semi-supervised DR algorithms appeal to projecting the observed data onto a low-dimensional manifold, where the margin between data form different classes is maximized. Most algorithms in this category, such as Locality Sensitive Discriminant Analysis [20], Locality Sensitive Semi-supervised Feature Selection [36], and Semi-supervised Discriminant Analysis [26], greatly take the intrinsic geometric structure into account, i.e. local structure of the data, indicating that nearby instances are likely to have the same label. However, global structure of the data, implying that instances on the same structure (typically referred to as a cluster or a manifold) are like to have the same label [1], is ignored in those algorithms. In fact, when mining the data, global structure is the same significance as local structure [1,10,11,29]. Therefore, it is necessary to integrate both global and local structures into the process of DR. The key challenge is how we can incorporate relatively few pairwise constraints into DR such that the represented data can still capture the available class information. To this end, we propose a semi-supervised discriminant analysis algorithm with integrating both global and local structures (SGLS), which naturally address DR under semi-supervised settings. Given pairwise constraints, the key idea in SGLS is to integrate both global and local structures in semi-supervised discriminant framework so that both discriminant and geometric structures of the data can be accurately captured. The SGLS algorithm has the same flavor as not only supervised DR algorithms, which try to adjust the distance among instances to improve the separability of the data, but unsupervised DR algorithms as well, which make nearby instances in the original space close to each other in the embedding space. The remainder of the paper is organized as follows. In Section 2, related work on the existing algorithms of semi-supervised DR is discussed. In Section 3, we introduce the general framework of our proposed SGLS algorithm. Section 4 discusses the experimental results on a number of real-world data sets. Finally, we draw conclusions in Section 5. II. Related Work Unsupervised DR has attracted considerable interest in the last decade [9,19,23,25,29]. Recently, a few researchers [5,9,13,17] have considered the case that discovers the local geometrical structure of the data manifold when the data lie in a low-dimensional space of the high-dimensional ambient space. Algorithms exploited along this line are used to estimate geometrical and topological properties of the embedding space, *Corresponding author: Xuesong Yin, yinxs@nuaa.edu.cn
  • 2. Semi-supervised Discriminant Analysis based on Data Structure DOI: 10.9790/0661-17373946 www.iosrjournals.org 40 | Page such as Locality Preserving Projections (LPP) [34], Local Discriminant Embedding (LDE) [35], Marginal Fisher Analysis (MFA) [37] and Transductive Component Analysis (TCA) [27]. Different with typical unsupervised DR algorithms, more lately, K-means was applied in the dimension-reduced space for avoiding the curse of dimensionality in several algorithms [25,31] which perform clustering and dimensionality reduction simultaneously. Linear Discriminant Analysis (LDA) [7,12], capturing the global geometric structure of the data by simultaneously maximizing the between-class distance and minimizing the within-class distance, is a well- known approach for supervised DR. It has been used widely in various applications, such as face recognition [5]. However, a major shortcoming of LDA is that it fails to discover the local geometrical structure of the data manifold. Semi-supervised DR can be considered as a new issue in DR area, which learns from a combination of both the side information and unlabeled data. In general, the side information can be expressed in diverse forms, such as class labels and pairwise constraints. Pairwise constraints include both must-link form in which the instances belong to the same class and cannot-link form in which the instances belong to different classes. Some semi-supervised DR algorithms [20,26,27,30,36,37] just applied partly labeled instances to maximize the margin between instances form different classes for improving performance. However, in many practical data mining applications, it is fairly expensive to label instances. In contrast, it could be easier for an expert or a user to specify whether some pairs of instances belong to the same class or not. Moreover, the pairwise constraints can be derived from labeled data but not vice versa. Therefore, making use of pairwise constraints to reduce the dimensionality of the data has been an important topic [8,16,33]. III. The SGLS Algorithm 3.1 The Integrating Global and Local structures Framework Given a set of instances 1, , nx x in N R , we can use a k-nearest neighbor graph to model the relationship between nearby data points. Specifically, we put an edge between nodes i and j if ix and jx are close, i.e. ix and jx are among k nearest neighbors of each other. Let the corresponding weight matrix be S, defined by 2 2 exp( ) ( ) ( ) 0 i j i k j j k i ij x x x N x or x N xS t otherwise          (1) Where ( )k iN x denotes the set of k nearest neighbors of ix . In general, if a weighted value ijS is large, ix and jx are close to each other in the original space. Thus, in order to preserve the locality structure, the embedding space obtained should be as smooth as possible. To learn an appropriate representation y ( 1[ , , ]T ny y y  ) with the locality structure, it is common to minimize the following objective function [13,34]: 2 , 2 T i j ij i j y y S y Ly  (2) where D is a diagonal matrix whose diagonal entries are column (or row, since S is symmetric) sum of S, that is 1 n ii ijj D S   . L D S  is the Laplacian matrix. On the other hand, we expect that the projections{ }T ia x generated by a projection vector d a R , which span the final embedding space, approach the sufficiently smooth embeddings{ }iy as much as possible. This expectation can be satisfied by the least squares formulation which can incorporate the local structure information via a regularization term defined as in Eq. (2). Mathematically, a projection vector a can be acquired by solving the following optimization problem [27]: 2 1min ( , ) T T J a y X a y y Ly   (3) where 1 0  is the trade-off parameter. Here we assume that the data matrix X has been centered. Taking the derivative of J(a,y) with respect to y and setting it equals to zero, we get
  • 3. Semi-supervised Discriminant Analysis based on Data Structure DOI: 10.9790/0661-17373946 www.iosrjournals.org 41 | Page 1 ( , ) 2( ) 0TJ a y y X a Ly y        (4) It follows that 1 1* ( ) T y I L X a    (5) where I is the identity matrix of size n. Applying Eq. (5) to Eq. (3), we can eliminate y and obtain the following optimization problem with regard to a: 1 1 1 1 1 min ( ) ( , *) * ( )( ) * ( ) ( ) T T T T T J a J a y y L I L y a X I L L X a a XGX a           (6) where G is a positive semi-definite matrix of size n by n and has the following expression: 1 1 1( ) ( )G I L L    (7) We can observe from Eq. (6) that minimizing J will bring the smooth subspace where both global and local structures of data are calculated with regard to a. 3.2 Discriminat Analysis With Pairwise Constraints Suppose we are given two sets of pairwise constraints including must-link (M) and cannot-link (C). In the previous work [32,38,40,41], such constraints were used for learning an adaptive metric between the prototypes of instances. However, recent research has shown that the distance metric learned on high- dimensional space is relatively unreliable [33,39]. Instead of using constraint-guided metric learning, in this paper our goal is to use pairwise constraints to find a projection vector such that in the embedding space the distance between any pair of instances involved in the cannot-link constraints is maximized while that between any pair of instances involved must-link constraints is minimized. If the projection vector a is acquired, we can introduce such a transformation: T i iy a x (8) Under this transformation, we calculate the following sum of the squared distances of the point pairs in M: ( , ) ( ) ( ) i j T T T T T w i j i j x x M T w d a x a x a x a x a S a       (9) where wS is the covariance matrix of the point pairs in M: ( , ) ( )( ) i j T w i j i j x x M S x x x x     (10) Correspondingly, for the point pairs in C, we have T b bd a S a (11) where bS has the following expression form: ( , ) ( )( ) i j T b i j i j x x C S x x x x     (12) Similar to LDA that seeks directions on which the data points of different classes are far from each other while requiring data points of the same class to be close to each other, we try to maximize bd and minimize wd . Thus, we can optimize the following problem to obtain the projection direction: max T b Ta w a S a a S a (13)
  • 4. Semi-supervised Discriminant Analysis based on Data Structure DOI: 10.9790/0661-17373946 www.iosrjournals.org 42 | Page 3.3 The Optimization Framework So far, we pursue the projection direction a from two formulations: (1) we introduce a least squares formulation for finding a, which facilities the integration of global and local structures; (2) we obtain the projection vector by maximizing the cannot-link covariance and simultaneously minimizing the must-link covariance. When the first formulation is considered as a regularization term, we arrive at the following optimization problem: 2 2 max (1 ) ( ) T b Ta w a S a a S a J a   (14) where 20 1  controls the smoothness of estimator. The optimal projection vector a that maximizes the objective function (14) is acquired by the maximum eigenvalue solution to the following generalized eigenvalue problem. 2 2((1 ) )T b wS a S XGX a     (15) In real-world applications, we usually need more projection vectors to span a subspace rather than just one. Let the matrix charactering such a target subspace be 1[ , , ]dA a a  which is formed by the eigenvectors associated with the d largest eigenvalues of Eq. (15). Also, it is worth noting that since our method is linear, it has a direct out-of-sample extension to novel sample x, i.e. T A x . Based on the analysis above, we propose to develop a semi-supervised dimensionality reduction algorithm which integrates both global and local structures defined in Eq. (14). The corresponding algorithm, called SGLS is presented in Algorithm 1. Algorithm 1 Input: X, M, C. Output: A projection matrix N d A R   . 1 Construct an unsupervised graph on all n instances to calculate S as in Eq. (2) and L = D – S. 2 Present a least squares formulation together with the local structure information to compute matrix G as in Eq. (7). 3 Compute wS and bS as in Eqs. (10) and (12). 4 Compute a generalized eigenvalue problem Eq. (15) and let the solved eigenvectors be 1, , da a in an order of decreasing eigenvalues. 5 Output 1[ , , ]dA a a  . To get a stable solution of the eigen-problem in Eq. (15), the matrix is required to be non-singular which is not true when the number of features is larger than the number of instances, i.e. N > n. In such case, we can apply the Tikhonov regularization technique [24,26] to improve the estimation. Thus, the generalized eigen- problem in this paper can be rewritten as: 2 2((1 ) )T b w NS a S XGX I a       (16) where NI is the identity matrix of size N and 0  is a regularization parameter. IV. Experiments In this section, we study the performance of our SGLS algorithm in terms of clustering accuracy on several real-world data sets from different application domains. We set k = 5 and t = 1 for the construction of the similarity matrix S defined in Eq. (1). 4.1 Experimental Setup In order to verify the efficacy of the proposed algorithm, we compare it with the state-of-the-art, semi- supervised dimensionality reduction algorithms including SSDR [22], SCREEN [33], LADKM [31] and LMDM [16]. We implemented SGLS in the MATLAB environment. All experiments were conducted on a PENTIUM DUAL 3G PC with 2.0 GB RAM. We compared all algorithms on seven benchmark data sets, including Segment and Digits from UCI machine Learning Repository, three document data sets DOC1, DOC2 and DOC3 from the 20-Newgroups data, and two image data sets : USPS handwritten data and Yale Face B (YALEB). The statistics of all data sets are summarized in Table 1.
  • 5. Semi-supervised Discriminant Analysis based on Data Structure DOI: 10.9790/0661-17373946 www.iosrjournals.org 43 | Page Table 1: Summary of the test datasets used in our experiment Data sets Instance Dimension Class Segment 2309 19 10 Digits 6300 16 10 DOC1 4000 4985 6 DOC2 5200 3132 5 DOC3 6500 6466 9 USPS 7122 256 10 YALEB 3110 2500 10 As the unified clustering platform, we use K-means algorithm as the clustering method to compare several algorithms after dimensionality reduction. For each data set, we ran different algorithms for 20 times and the comparison was based on the average performance. The parameter in SGLS are always set to 1 10  and 2 0.8  if without extra explanation. Moreover, for the fair comparison, the dimensionality of all the algorithms is set to K-1 (K is the number of clustering). To evaluate the clustering performance, in this paper, we employ normalized mutual information (NMI) as the clustering validation measure, which is widely used to verify the performance of clustering algorithms [8,25,33]. Then, NMI is defined as ( , ) ( ) ( ) I X Y NMI H X H Y  (17) where X and Y denote the random variables of cluster memberships from the ground truth and the output of clustering algorithm, respectively. ( , )I X Y is the mutual information between X and Y. ( )H X and ( )H Y are the entropies of X and Y respectively. 4.2 Experimental Results
  • 6. Semi-supervised Discriminant Analysis based on Data Structure DOI: 10.9790/0661-17373946 www.iosrjournals.org 44 | Page Figure 1 Clustering performance on 6 data sets with different number of constraints Table 2 presents the NMI results of various algorithms on all seven data sets. Also, Figure 1 shows the clustering performance of standard K-means applied to the projected data by different dimensionality reduction algorithms with different numbers of pairwise constraints. As can be seen, our SGLS algorithm achieves the best performance on Digits, DOC1, DOC3, USPS, and YALEB data sets. In fact, the performance of our algorithm is always comparable to that of the other algorithm on the rest of all seven data sets used in this paper. Specifically, we can obtain the following main observations from Table 2 and Figure 1: (1) The SGLS algorithm significantly outperforms not only LMDM and LDAKM on all of the seven data sets used in the experiment, but SCREEN and SSDR on the six data sets as well. This can be contributed to such a fact that integrating both global and local structures in SGLS may be beneficial. (2) It is interesting to note that SGLS slightly underperform both SSDR on Segment and SCREEN on DOC2. This implies that in certain case such as Segment, global and local structures may capture similar information so that integrating such both structures seems to hardly help in the dimensionality reduction process. (3) It is clear from the presented results that LDAKM performs relatively worse that the other algorithms. This can be explained that LDAKM does not employ any side information to pursue the projection direction, which only makes use of the abundant unlabeled instances. Thus, LDAKM has a greatly inferior performance. (4) Due to the fact that SCREEN and LMDM just pairwise constraints to learn an optimal subspace, their performances are worse that those of SSDR and SGLS respectively. In fact, a large number of the unlabeled data play an important role for semi-supervised dimensionality reduction [8,26,27]. (5) Since SSDR can apply both the pairwise constraints and the unlabeled data to learn a subspace, it can provide a relatively satisfying clustering performance in comparison with LDAKM, SCREEN and LMDM. Compared with our SGLS, however, SSDR has still provided the fairly inferior performance. Actually, SSDR can be considered as the constrained Principal Component Analysis (PCA). Thus, it fails to preserve local neighborhood structures of the instances in the reduced low-dimensional space. Table 2: NMI comparisons on seven data sets Data sets LDAKM SCREEN SSDR LMDM SGLS Segment 0.687 0.723 0.744 0.712 0.737 Digits 0.714 0.755 0.767 0.750 0.788 DOC1 0.567 0.596 0.604 0.588 0.625 DOC2 0.502 0.561 0.547 0.539 0.556 DOC3 0.746 0.802 0.810 0.797 0.823 USPS 0.636 0.675 0.682 0.669 0.705 YALEB 0.811 0.848 0.837 0.833 0.861 V. Conclusions In this paper, we propose a novel semi-supervised dimensionality reduction algorithm, called SGLS. For dimensionality reduction and clustering, SGLS integrates both global and local geometric structures so that it can be naturally extended to deal with the abundant labeled data, as graph Laplacian is defined on all the instances. Compared with the state-of-the-art dimensionality reduction algorithms on a collection of benchmark data sets in terms of clustering accuracy, SGLS can always achieve a clear performance gain. Extensive experiments exhibit the advantages of the novel semi-supervised clustering method of SGLS+K-means.
  • 7. Semi-supervised Discriminant Analysis based on Data Structure DOI: 10.9790/0661-17373946 www.iosrjournals.org 45 | Page Acknowledgments The research reported in this paper has been partially supported by Natural Science Foundation of Zhejiang Province under Grant No. Y1100349, Public Project of Zhejiang Province under Grant No. 2013C33087, Academic Climbing Project of Young Academic Leaders of Zhejiang Province No. pd2013446. References [1]. D. Zhou, O. Bousquet, T.N. Lal, J. Weston, and B. Scholkopf. Learning with local and global consistency. In Advances in Neural Information Processing Systems 16, 2003, pp. 321-328. [2]. P. Garcia-Laencina, G. Rodriguez-Bermudez, J. Roca-Dorda. Exploring dimensionality reduction of EEG features in motor imagery task classification, Expert Systems with Applications, 41 (11) (2014)5285-5295. [3]. Y. Lin, T. Liu, C. Fuh Dimensionality Reduction for Data in Multiple Feature Representations. In Advances in Neural Information Processing Systems 21, 2008, pp. 961-968. [4]. X. Zhu, Z. Ghahramani, and J. Lafferty. Semi-Supervised Learning using Gaussian Fields and Harmonic Functions, Proc. Twentieth Int’l Conf. Machine Learning, 2003, pp. 912-919. [5]. X. He, S. Yan, Y. Hu, P. Niyogi, H. Zhang, Face recognition using laplacianfaces, IEEE Trans. Pattern Anal. Mach. Intell. 27 (3) (2005) 328-340. [6]. S. Roweis, L. Saul. Nonlinear dimensionality reduction by locally linear embedding, Science, 290 (5500) (2000) 2323-2326. [7]. R. O. Duda, P. E. Hart, and D. G. Stork. Pattern Classification. Wiley-Interscience, Hoboken, NJ, 2nd edition, 2000. [8]. X. Yin, T. Shu, Q. Huang. Semi-supervised fuzzy clustering with metric learning and entropy regularization. Knowledge-Based Systems, 2012, 35(11):304-311. [9]. M. Belkin and P. Niyogi. Laplacian Eigenmaps for Dimensionality Reduction and Data Representation. Neural Computation, 15(6) (2003) 1373-1396. [10]. J. Tenenbaum, V. de Silva, and J. Langford. A global geometric framework for nonlinear dimensionality reduction. Science, 290(5500) (2000) 2319-2323. [11]. F. Tsai. Dimensionality reduction techniques for blog visualization, Expert Systems with Applications, 38(3) (2011) 2766-2773. [12]. J. Yang, A.F. Frangi, J.Y. Yang, D. Zhang, Z. Jin, KPCA plus LDA: A complete kernel fisher discriminant framework for feature extraction and recognition, IEEE Trans. Pattern Anal. Mach. Intell. 27 (2) (2005) 230-244. [13]. M. Belkin, P. Niyogi, Laplacian eigenmaps and spectral techniques for embedding and clustering, in: Advances in Neural Information Processing Systems 14, 2001, pp. 585-591. [14]. D. Cai, X. He, J. Han, Document clustering using locality preserving indexing, IEEE Trans. Knowl. Data Eng. 17 (12) (2005) 1624- 1637. [15]. J. He, L. Ding, L. Jiang, Z. Li, Q. Hu. Intrinsic dimensionality estimation based on manifold assumption, Journal of Visual Communication and Image Representation, 25( 5)(2014) 740-747. [16]. S. Xiang, F. Nie, and C. Zhang. Learning a Mahalanobis distance metric for data clustering and classification. Pattern Recognition, 41(12) (2008) 3600 - 3612. [17]. D. Cai, X. He, J. Han, H.-J. Zhang, Orthogonal laplacianfaces for face recognition, IEEE Trans. Image Process. 15 (11) (2006) 3608-3614. [18]. Z. Zhang, M. Zhao, T. Chow. Marginal semi-supervised sub-manifold projections with informative constraints for dimensionality reduction and recognition, Neural Networks, 36(2012) 97-111. [19]. J. Chen, J. Ye, and Q. Li. Integrating Global and Local Structures: A Least Squares Framework for Dimensionality Reduction, in: IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2007, pp. 1-8. [20]. D. Cai, X. He, K. Zhou, J. Han, H. Bao, Locality sensitive discriminant analysis, in: International Joint Conference on Artificial Intelligence, 2007, pp. 708-713. [21]. W. Du, K. Inoue, K. Urahama, Dimensionality reduction for semi-supervised face recognition, in: Proceedings of the Fuzzy Systems and Knowledge Discovery, 2005, pp. 1-10. [22]. D. Zhang, Z.-H. Zhou, S. Chen, Semi-supervised dimensionality reduction, in: SIAM Conference on Data Mining (SDM), 2007, pp. 629-634. [23]. X. Yang, H. Fu, H. Zha, J. Barlow, Semi-supervised nonlinear dimensionality reduction, in: Proceedings of the International Conference on Machine Learning, 2006, pp. 1065-1072. [24]. M. Belkin, P. Niyogi, V. Sindhwani, Manifold regularization: a geometric framework for learning from labeled and unlabeled examples, J. Mach. Learning Res. 1 (1) (2006) 1-48. [25]. J. Ye, Z. Zhao, H. Liu. Adaptive distance metric learning for clustering. In: Proc. of the Conference on Computer Vision and Pattern Recognition, 2007, pp. 1-7. [26]. D. Cai, X. He, J. Han, Semi-supervised discriminant analysis, in: Proceedings of the International Conference on Computer Vision, 2007, pp. 1-7. [27]. W. Liu, D. Tao, and J. Liu. Transductive Component Analysis, in: Proceedings of the International Conference on Data Mining, 2008, pp. 433-442. [28]. P. Han, A. Jin, A. Kar. Kernel Discriminant Embedding in face recognition, Journal of Visual Communication and Image Representation, 22(7) (2011) 634-642. [29]. J. Ye. Least Squares Linear Discriminant Analysis, in: Proceedings of International Conference on Machine Learning, 2007, pp. 1087-1093. [30]. X. Chen, S. Chen, H. Xue, X. Zhou. A unified dimensionality reduction framework for semi-paired and semi-supervised multi-view data, Pattern Recognition, 45(5) (2012) 2005-2018. [31]. C. Ding, T. Li. Adaptive dimension reduction using discriminant analysis and K-means clustering. In: Proceedings of the International Conference on Machine Learning, 2007, pp. 521-528. [32]. E. Xing, A. Ng, M. Jordan, S. Russell. Distance metric learning with application to clustering with side-information. In: Advances in Neural Information Processing Systems 15, 2003, pp. 505-512. [33]. W. Tang, H. Xiong, S. Zhong, J. Wu. Enhancing semi-supervised clustering: a feature projection perspective. In: Proceedings of the International Conference on Knowledge Discovery and Data Mining, 2007, pp. 707-716. [34]. X. He, P. Niyogi. Locality preserving projections, In: Advances in Neural Information Processing Systems 16, 2003. [35]. H. Chen, H. Chang, and T. Liu. Local discriminant embedding and its variants. In: Proceedings of the Internal Conference on Computer Vision and Pattern Recognition, 2005, pp.846-853. [36]. J. Zhao, K. Lu, X. He, Locality sensitive semi-supervised feature selection, Neurocomputing, 71 (10-12), (2008)1842-1849.
  • 8. Semi-supervised Discriminant Analysis based on Data Structure DOI: 10.9790/0661-17373946 www.iosrjournals.org 46 | Page [37]. S. Yan, D. Xu, B. Zhang, H.J. Zhang, Q. Yang, and S. Lin. Graph embedding and extension: A general framework for dimensionality reduction. IEEE Transactions on PAMI, 29(1) (2007) 40-51. [38]. B. Kulis, S. Basu, I. Dhillon, and R. Mooney. Semi-supervised graph clustering: a kernel approach. In: Proceedings of the International Conference on Machine Learning, 2005, pp.59-68. [39]. D. Meng, Y. Leung, Z. Xu. Passage method for nonlinear dimensionality reduction of data on multi-cluster manifolds, Pattern Recognition, 46 (8) (2013) 2175-2186. [40]. M. Bilenko, S. Basu, R.J. Mooney, Integrating constraints and metric learning in semi-supervised clustering, in: Proceedings of International Conference on Machine Learning, 2004, pp. 81–88. [41]. D.Y. Yeung, H. Chang, Extending the relevant component analysis algorithm for metric learning using both positive and negative equivalence constraints, Pattern Recognition 39 (5) (2006) 1007–1010.