dc.description.abstract | Some classification techniques require class labels to be specified. One such technique is Fisher's Linear Discriminant (FLD), which has been used to develop an algorithm that seeks bimodal projection directions. Although FLD is a supervised classification method, our approach to feature extraction is unsupervised. An artificial two-class problem is created by randomly assigning labels to the data points. FLD then attempts to align the classifier to separate these classes, thereby assisting in the detection of bimodal projection directions.
This method achieves effective dimensionality reduction and naturally imposes a significant degree of weight sharing on the neural network performing the feature extraction.
The objectives of accurate encoding and generating discriminative representations for pattern classification are not always consistent. The feature vectors obtained using bimodal projection directions demonstrate strong discriminative power, but are not designed for encoding. To address this, the feature map is extended to define an embedding in the feature space, ensuring a smooth inverse map exists from the feature space back to the input space. As a result, the transformation becomes lossless. Although the feature vector has a higher dimension than the input vector, the number of weight vectors required remains the same as in the unextended feature map.
The NIST dataset for handwritten digits was used to evaluate the applicability and efficiency of the bimodal projection-based features for handling large, high-dimensional data. The proposed method constrains the number of free parameters (weights) in the network, improving the generalization performance of the resulting neural networks. This has been convincingly demonstrated by the algorithm's performance on the NIST dataset, which includes 344,307 training images and 58,646 test images, each of size 32 × 32. Despite limited computational resources, reasonable results were achieved.
The embedding-based extension of the feature set provides a unified approach to both data classification and data representation. Although it results in feature vectors with dimensions larger than the input space, experiments using popular benchmarks like the Wine and Vowel datasets show that simple classifiers can still achieve good performance. Thus, the increased complexity due to higher dimensionality is offset by the simplicity and efficiency of the classifiers used. | |