Study on Unsupervised Feature Selection Method Based on Extended Entropy
Keywords:Unsupervised feature selection, extended entropy, information loss, correlation value
AbstractFeature selection techniques are designed to find the relevant feature subset of the original features that can facilitate clustering, classification and retrieval. It is an important research topic in pattern recognition and machine learning. Feature selection is mainly partitioned into two classes, i.e. supervised and unsupervised methods. Currently research mostly concentrates on supervised ones. Few efficient unsupervised feature selection methods have been developed because no label information is available. On the other hand, it is difficult to evaluate the selected features. An unsupervised feature selection method based on extended entropy is proposed here. The information loss based on extended entropy is used to measure the correlation between features. The method assures that the selected features have both big individual information and little redundancy information with the selected features. At last, the efficiency of the proposed method is illustrated with some practical datasets.
Download data is not yet available.
How to Cite
Sun, Z., Li, F., & Huang, H. (2019). Study on Unsupervised Feature Selection Method Based on Extended Entropy. COMPUTING AND INFORMATICS, 38(1), 223–239. Retrieved from https://www.cai.sk/ojs/index.php/cai/article/view/2019_1_223