Journal on Communications ›› 2019, Vol. 40 ›› Issue (10): 101-108.doi: 10.11959/j.issn.1000-436x.2019154

• Papers • Previous Articles     Next Articles

Feature selection algorithm based on XGBoost

Zhanshan LI1,2,3, Zhaogeng LIU2,3()   

  1. 1 College of Computer Science and Technology,Jilin University,Changchun 130012,China
    2 College of Software,Jilin University,Changchun 130012,China
    3 Key Laboratory of Symbolic Computation and Knowledge Engineering,Ministry of Education,Jilin University,Changchun 130012,China
  • Revised:2019-04-04 Online:2019-10-25 Published:2019-11-07
  • Supported by:
    The National Natural Science Foundation of China(6167226);The Natural Science Foundation of Jilin Province(2018010143JC);Industrial Technology Research and Development Special Project of Jilin Province Development and Reform Commission(2019C053-9)

Abstract:

Feature selection in classification has always been an important but difficult problem.This kind of problem requires that feature selection algorithms can not only help classifiers to improve the classification accuracy,but also reduce the redundant features as much as possible.Therefore,in order to solve feature selection in the classification problems better,a new wrapped feature selection algorithm XGBSFS was proposed.The thought process of building trees in XGBoost was used for reference,and the importance of features from three importance metrics was measured to avoid the limitation of single importance metric.Then the improved sequential floating forward selection (ISFFS) was applied to search the feature subset so that it had high quality.Compared with the experimental results of eight datasets in UCI,the proposed algorithm has good performance.

Key words: feature selection, XGBoost, sequential floating forward selection

CLC Number: 

No Suggested Reading articles found!