Abstract
Feature selection is an important part of data preprocessing. Selecting effective feature subsets can effectively reduce feature redundancy and reduce irrelevant features, and reduce training costs. Based on the theory of feature clusters, this paper proposes a feature selection strategy based on the graph structure. Considering a feature as a node in the graph, using the idea of graph message propagation, integrating the first-order neighbor information of each node, and then selecting the key point of the local maximum score as the selected feature, this can effectively reduce the feature redundancy and reduce features that are not related to the label. Finally, in order to verify the anti-interference of this novel method, the noise dimension was added in the UCI data set, and the comparison test was again performed. The experimental results show that the proposed algorithm can effectively improve the classification accuracy in a specific data set, and the anti-interference is better than other feature selection algorithms.
Z. Zhang and D. Zheng—This work was supported by the Natural Science Foundation of Heilongjiang Province (No. F2017024, No. F2017025).
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Cai, Z., Goebel, R., Salavatipour, M., Lin, G.: Selecting dissimilar genes for multi-class classification, an application in cancer subtyping. BMC Bioinform. 8, 206 (2007). (IF: 3.428)
Cai, Z., Zhang, T., Wan, X.: A computational framework for influenza antigenic cartography. PLoS Comput. Biol. 6(10), e1000949 (2010)
Cai, Z., Xu, L., Shi, Y., Salavatipour, M., Goebel, R., Lin, G.: Using gene clustering to identify discriminatory genes with higher classification accuracy. In: IEEE the 6th Symposium on Bioinformatics and Bioengineering (BIBE 2006) (2006)
Yang, K., Cai, Z., Li, J., Lin, G.: A stable gene selection in microarray data analysis. BMC Bioinform. 7, 228 (2006)
Gilmer, J., Schoenholz, S.S., Riley, P.F., et al.: Neural message passing for quantum chemistry (2017)
Lowe, D.G.: Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vision 60(2), 91–110 (2004)
Caruana, R., De Sa, V.R.: Benefitting from the variables that variable selection discards. J. Mach. Learn. Res. 3, 1245–1264 (2003)
Kira, K., Rendell, L.A.: A practical approach to feature selection. In: Proceedings of the Ninth International Workshop on Machine Learning, pp. 249–256 (1992)
Kononenko, I.: Estimating attributes: analysis and extensions of RELIEF. In: Bergadano, F., De Raedt, L. (eds.) ECML 1994. LNCS, vol. 784, pp. 171–182. Springer, Heidelberg (1994). https://doi.org/10.1007/3-540-57868-4_57
Pal, S.K., Wang, P.P.: Genetic Algorithms for Pattern Recognition. CRC Press Inc., Boca Raton (1996)
Kennedy, J.: Particle swarm optimization. In: Sammut, C., Webb, G.I. (eds.) Encyclopedia of Machine Learning, pp. 760–766. Springer, Boston (2010). https://doi.org/10.1007/978-0-387-30164-8
Chuang, L.Y., Chang, H.W., Tu, C.J., et al.: Improved binary PSO for feature selection using gene expression data. Comput. Biol. Chem. 32(1), 29–38 (2008)
Mirjalili, S., Mirjalili, S.M., Lewis, A.: Grey wolf optimizer. Adv. Eng. Softw. 69(3), 46–61 (2014)
Emary, E., Zawbaa, H.M., Hassanien, A.E.: Binary grey wolf optimization approaches for feature selection. Neurocomputing 172, 371–381 (2016)
Mafarja, M.M., Mirjalili, S.: Hybrid binary ant lion optimizer with rough set and approximate entropy reducts for feature selection. Soft. Comput. 23(15), 6249–6265 (2019)
Al-Tashi, Q., Kadir, S.J.A., Rais, H.M., et al.: Binary optimization using hybrid grey wolf optimization for feature selection. IEEE Access 7, 39496–39508 (2019)
Mafarja, M., Aljarah, I., Faris, H., et al.: Binary grasshopper optimisation algorithm approaches for feature selection problems. Expert Syst. Appl. 117, 267–286 (2019)
Li, W., Chao, X.Q.: Improved particle swarm optimization method for feature selection. J. Front. Comput. Sci. Technol. 13(6), 990–1004 (2019)
Efron, B., Hastie, T., Johnstone, I., et al.: Least angle regression. Ann. Stat. 32(2), 407–451 (2004)
Wen, C., Zhang, A., Quan, S., et al.: BeSS: an R package for best subset selection in linear, logistic and CoxPH models (2017)
Wang, L., Jiang, S.: Novel feature selection method based on feature clustering. Appl. Res. Comput. 32(5), 1305–1308 (2015)
Song, Q., Ni, J., Wang, G.: A fast clustering-based feature subset selection algorithm for high-dimensional data. IEEE Trans. Knowl. Data Eng. 25(1), 1–14 (2013)
Author information
Authors and Affiliations
Corresponding authors
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2019 Springer Nature Switzerland AG
About this paper
Cite this paper
Hu, Z., Zhang, Z., Huang, Z., Zheng, D., Zhang, Z. (2019). Feature Selection Based on Graph Structure. In: Li, Y., Cardei, M., Huang, Y. (eds) Combinatorial Optimization and Applications. COCOA 2019. Lecture Notes in Computer Science(), vol 11949. Springer, Cham. https://doi.org/10.1007/978-3-030-36412-0_23
Download citation
DOI: https://doi.org/10.1007/978-3-030-36412-0_23
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-36411-3
Online ISBN: 978-3-030-36412-0
eBook Packages: Computer ScienceComputer Science (R0)