A review: Feature selection from the interaction perspective
Abstract
Feature selection plays a vital role in machine learning and data mining by identifying a representative subset of features to improve model performance, reduce computational cost, and prevent overfitting. Although traditional methods aim to eliminate redundant and irrelevant features, research into feature interactions remains relatively underexplored. Feature interaction refers to the joint contribution of multiple features to predictive performance; effectively capturing such interactions can substantially enhance the accuracy of the model. This article reviews the evolution of feature selection methods over the past three decades, highlighting their strengths and limitations. In addition, it investigates efficient strategies for identifying interactive features, taking into account relevance, redundancy, interactivity, and complementarity. To provide a broader and up-to-date perspective, recent advances in interaction-aware, explainability-driven, and deep learning-based feature selection methods are also discussed. Finally, the paper summarizes the open issues in the search for feature subsets and outlines key challenges for future research.