Publications by authors named "Huilin Xiong"

In this paper, we present a multiclass data classifier, denoted by optimal conformal transformation kernel (OCTK), based on learning a specific kernel model, the CTK, and utilize it in two types of image recognition tasks, namely, face recognition and object categorization. We show that the learned CTK can lead to a desirable spatial geometry change in mapping data from the input space to the feature space, so that the local spatial geometry of the heterogeneous regions is magnified to favor a more refined distinguishing, while that of the homogeneous regions is compressed to neglect or suppress the intraclass variations. This nature of the learned CTK is of great benefit in image recognition, since in image recognition we always have to face a challenge that the images to be classified are with a large intraclass diversity and interclass similarity.

View Article and Find Full Text PDF

The Gaussian process (GP) approaches to classification synthesize Bayesian methods and kernel techniques, which are developed for the purpose of small sample analysis. Here we propose a GP model and investigate it for the facial expression recognition in the Japanese female facial expression dataset. By the strategy of leave-one-out cross validation, the accuracy of the GP classifiers reaches 93.

View Article and Find Full Text PDF

Small sample size is one of the biggest challenges in microarray data analysis. With microarray data being dramatically accumulated, integrating data from related studies represents a natural way to increase sample size so that more reliable statistical analysis may be performed. In this paper, we present a simple and effective integration scheme, called Normalised Linear Transform (NLT), to combine data from different microarray platforms.

View Article and Find Full Text PDF

One important application of gene expression analysis is to classify tissue samples according to their gene expression levels. Gene expression data are typically characterized by high dimensionality and small sample size, which makes the classification task quite challenging. In this paper, we present a data-dependent kernel for microarray data classification.

View Article and Find Full Text PDF

Background: The most fundamental task using gene expression data in clinical oncology is to classify tissue samples according to their gene expression levels. Compared with traditional pattern classifications, gene expression-based data classification is typically characterized by high dimensionality and small sample size, which make the task quite challenging.

Results: In this paper, we present a modified K-nearest-neighbor (KNN) scheme, which is based on learning an adaptive distance metric in the data space, for cancer classification using microarray data.

View Article and Find Full Text PDF

In this paper, we present a method of kernel optimization by maximizing a measure of class separability in the empirical feature space, an Euclidean space in which the training data are embedded in such a way that the geometrical structure of the data in the feature space is preserved. Employing a data-dependent kernel, we derive an effective kernel optimization algorithm that maximizes the class separability of the data in the empirical feature space. It is shown that there exists a close relationship between the class separability measure introduced here and the alignment measure defined recently by Cristianini.

View Article and Find Full Text PDF

This paper presents a new self-creating model of a neural network in which a branching mechanism is incorporated with competitive learning. Unlike other self-creating models, the proposed scheme, called branching competitive learning (BCL), adopts a special node-splitting criterion, which is based mainly on the geometrical measurements of the movement of the synaptic vectors in the weight space. Compared with other self-creating and nonself-creating competitive learning models, the BCL network is more efficient to capture the spatial distribution of the input data and, therefore, tends to give better clustering or quantization results.

View Article and Find Full Text PDF