AUT Journal of Mathematics and Computing (Oct 2024)
Feature representation via graph-regularized entropy-weighted nonnegative matrix factorization
Abstract
Feature extraction plays a crucial role in dimensionality reduction in machine learning applications. Nonnegative Matrix Factorization (NMF) has emerged as a powerful technique for dimensionality reduction; however, its equal treatment of all features may limit accuracy. To address this challenge, this paper introduces Graph-Regularized Entropy-Weighted Nonnegative Matrix Factorization (GEWNMF) for enhanced feature representation. The proposed method improves feature extraction through two key innovations: optimizable feature weights and graph regularization. GEWNMF uses optimizable weights to prioritize the extraction of crucial features that best describe the underlying data structure. These weights, determined using entropy measures, ensure a diverse selection of features, thereby enhancing the fidelity of the data representation. This adaptive weighting not only improves interpretability but also strengthens the model against noisy or outlier-prone datasets. Furthermore, GEWNMF integrates robust graph regularization techniques to preserve local data relationships. By constructing an adjacency graph that captures these relationships, the method enhances its ability to discern meaningful patterns amid noise and variability. This regularization not only stabilizes the method but also ensures that nearby data points appropriately influence feature extraction. Thus, GEWNMF produces representations that capture both global trends and local nuances, making it applicable across various domains. Extensive experiments on four widely used datasets validate the efficacy of GEWNMF compared to existing methods, demonstrating its superior performance in capturing meaningful data patterns and enhancing interpretability.
Keywords