摘要

In this paper, we propose a novel feature selection method which can suppress the input features during the process of model construction automatically. The main idea is to obtain better performance and sparse solutions by introducing Tikhonov regularization terms and measuring the objective function with -norm, based on projection twin support vector machine. Furthermore, to make the problem easy to solve, the exterior penalty theory is adopted to convert the original problem into an unconstrained problem. In contrast with twin support vector machine which needs solve two QPPs, our method only solves two linear equations by using a fast generalized Newton algorithm. In order to improve performance, a recursive algorithm is proposed to generate multiple projection axes for each class. To disclose the feasibility and effectiveness of our method, we conduct some experiments on UCI and Binary Alpha-digits data sets.