IEEE Access (Jan 2024)
ARIA, HaRIA, and GeRIA: Novel Metrics for Pre-Model Interpretability
Abstract
This work proposes three novel Pre-Model Interpretability metrics: HaRIA, ARIA, and GeRIA. They aim to assess the potential utilization of features in machine learning models prior to the training phase, by quantifying the Relative Information Availability. These metrics integrate Mutual Information and ANOVA F-values, scaled using Maximum Absolute Scaling. This allows to evaluate the potential of a feature being used in the learning process efficiently and effectively without the computational expense of model training. The metrics are designed to provide a holistic view of feature relevance by capturing both the non-linear dependencies and variance effects among features. Validation of these metrics across multiple datasets demonstrates their capability to approximate the importance assigned by more complex models, as evidenced by their strong correlation with traditional feature importance measures and SHAP values obtained post-model training. The consistency observed in various datasets underscores the potential of RIA metrics to facilitate early-stage model development decisions, offering a cost-effective tool for feature evaluation in scenarios where computational resources are limited or rapid prototyping is necessary. However, some discrepancies, especially with complex models like ANNs, indicate areas for future research and refinement. The introduction of these metrics marks a significant step toward enhancing the efficiency and transparency of AI development by enabling a better understanding of data characteristics and potential model behavior before actual model deployment.
Keywords