Non-linear Feature Selection Based on Convolution Neural Networks with Sparse Regularization

Zhang S, Lang Z-Q. Orthogonal least squares based fast feature selection for linear classification. Patt Recog. 2022;3(123):108419.

Article  Google Scholar 

Shang R, Zhang X, Feng J, et al. Sparse and low-dimensional representation with maximum entropy adaptive graph for feature selection. Neurocomputing. 2022;7(485):57–73.

Article  Google Scholar 

Hallajian B, Motameni H, Akbari E. Ensemble feature selection using distance-based supervised and unsupervised methods in binary classification. Exp Syst Appl. 2022;15(200):116794.

Article  Google Scholar 

Li M, Huan J, Yang J. Automatic feature extraction and fusion recognition of motor imagery EEG using multilevel multiscale CNN. Med Biol Eng Comput. 2021;59:2037–50.

Article  Google Scholar 

Chen S, Ding CHQ, Zhou Z, Luo B. Feature selection based on correlation deflation. Neural Comput Appl. 2019;10(31):6383–92.

Article  Google Scholar 

You D, Sun M, Liang S, et al. Online feature selection for multi-source streaming features. Inf Sci. 2022;4(590):267–95.

Article  Google Scholar 

Wei Z, Li Q, Wei J, et al. Neural networks for a class of sparse optimization with \(L_0\)-regularization. Neural Netw. 2022;151:211–21.

Vu V, Lei J. Minimax sparse principal subspace estimation in high dimension. Inst Math Stat. 2013;6(41):2905–47.

MathSciNet  MATH  Google Scholar 

Pang T, Nie F, Han J, et al. Efficient feature selection via \(L_\)-norm constrained sparsed regression. IEEE Trans Knowl Data Eng. 2019;5(31):880–93.

Article  Google Scholar 

Jin X, Miao J, Wang Q, et al. Sparse matrix factorization with \(L_\)-norm for matrix completion. Patt Recog. 2022;127:108655.

Article  Google Scholar 

Huang Y, Jie W, Yu Z, et al. Supervised feature selection through deep neural networks with pairwise connected structure. Knowl Based Syst. 2020;27(204):106202.

Article  Google Scholar 

Tokovarov M. Convolutional neural networks with reusable full-dimension-long layers for feature selection and classification of motor imagery in EEG signals. In: 29th International Conference on Artificial Neural Networks. 2020. p. 79–91.

Wu Y, Lan Y, Zhang L, et al. Feature flow regularization: improving structured sparsity in deep neural networks. Neural Netw. 2023;161:598–613.

Article  Google Scholar 

Nie F, Huang H, Cai X, Ding C.: Efficient and robust feature selection via joint \(L_\)-norm minimization. In: Proceedings of the 23rd International Conference on Neural Information Processing Systems. 2020. p. 1813–21.

Wang Z, Nie F, Lai T, et al. Discriminative feature selection via a structured sparse subspace learning Module. In: Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence (IJCAI-2020). pp. 3009–15.

Zhang H, Wang J, Sun Z, et al. Feature selection for neural networks using group Lasso regularization. IEEE Trans Knowl Data Eng. 2020;4(32):659–73.

Article  Google Scholar 

Cai X, Nie F, Huang H. Exact top-k feature selection via l2,0-norm constraint. In: Proceedings of the Twenty-Third International Joint Conference on Artificial Intelligence. 2013. p. 1240–6.

Scardapance S, Comminiello D, Hussain A, et al. Group sparse regularization for deep neural networks. Neurocomputing. 2017;7(241):81–9.

Article  Google Scholar 

Rui T, Zou J, Zhou Y, et al. Convolutional neural network feature maps selection based on LDA. Multimed Tools Appl. 2018;77:10635–49.

Article  Google Scholar 

Xie X, Zhang H, Wang J, et al. Learning optimized structure of neural networks by hidden node pruning with \(L_1\) regularization. IEEE Trans Cybern. 2020;3(50):1333–46.

Article  Google Scholar 

Li Y, Yu C, Wasserman W. Deep feature selection: theory and application to identify enhancers and promoters. J Comput Biol. 2016;5(23):322–36.

Article  Google Scholar 

Yamada Y, Lindenbaum O, Negahban S, et al. Feature selection using stochastic gates. In: Proceedings of the 37th International Conference on Machine Learning, 119. 2020. p. 10648–59.

Roffo G, Melzi S, Castellani U, et al. Infinite feature selection: a graph-based feature filtering approach. IEEE Trans Patt Anal Mach Intell. 2021;12(43):4396–410.

Article  Google Scholar 

Zuo Z, Li J, Xu H, et al. Curvature-based feature selection with application in classifying electronic health records. Technol Forecast Soc Change. 2021;173:121–7.

Article  Google Scholar 

Guo X, Yu K, Cao F, et al. Error-aware Markov blanket learning for causal feature selection. Inf Sci. 2022;589:849–77.

Article  Google Scholar 

Saadatmand H, Akbarzadeh-T M-R. Set-based integer-coded fuzzy granular evolutionary algorithms for high-dimensional feature selection. Appl Soft Comput. 2023;142:110240.

留言 (0)

沒有登入
gif