Facebook Instagram Twitter RSS Feed PodBean Back to top on side

Study on Unsupervised Feature Selection Method Based on Extended Entropy

In: Computing and Informatics, vol. 38, no. 1
Z. Sun - F. Li - H. Huang

Details:

Year, pages: 2019, 223 - 239
Language: eng
Keywords:
Unsupervised feature selection, extended entropy, information loss, correlation value
Document type: article
About article:
Feature selection techniques are designed to find the relevant feature subset of the original features that can facilitate clustering, classification and retrieval. It is an important research topic in pattern recognition and machine learning. Feature selection is mainly partitioned into two classes, i.e. supervised and unsupervised methods. Currently research mostly concentrates on supervised ones. Few efficient unsupervised feature selection methods have been developed because no label information is available. On the other hand, it is difficult to evaluate the selected features. An unsupervised feature selection method based on extended entropy is proposed here. The information loss based on extended entropy is used to measure the correlation between features. The method assures that the selected features have both big individual information and little redundancy information with the selected features. At last, the efficiency of the proposed method is illustrated with some practical datasets.
How to cite:
ISO 690:
Sun, Z., Li, F., Huang, H. 2019. Study on Unsupervised Feature Selection Method Based on Extended Entropy. In Computing and Informatics, vol. 38, no.1, pp. 223-239. 1335-9150. DOI: https://doi.org/10.31577/cai_2019_1_223

APA:
Sun, Z., Li, F., Huang, H. (2019). Study on Unsupervised Feature Selection Method Based on Extended Entropy. Computing and Informatics, 38(1), 223-239. 1335-9150. DOI: https://doi.org/10.31577/cai_2019_1_223
About edition:
Publisher: Ústav informatiky SAV
Published: 4. 6. 2019