A Cheap Feature Selection Approach for the K-Means Algorithm
The increase in the number of features that need to be analyzed in a wide variety of areas, such as genome sequencing, computer vision, or sensor networks, represents a challenge for the <inline-formula> <tex-math notation="LaTeX">K </tex-math></inline-formula>-mean...
Saved in:
| Published in | IEEE transaction on neural networks and learning systems Vol. 32; no. 5; pp. 2195 - 2208 |
|---|---|
| Main Authors | , , |
| Format | Journal Article |
| Language | English |
| Published |
United States
IEEE
01.05.2021
The Institute of Electrical and Electronics Engineers, Inc. (IEEE) |
| Subjects | |
| Online Access | Get full text |
| ISSN | 2162-237X 2162-2388 2162-2388 |
| DOI | 10.1109/TNNLS.2020.3002576 |
Cover
| Summary: | The increase in the number of features that need to be analyzed in a wide variety of areas, such as genome sequencing, computer vision, or sensor networks, represents a challenge for the <inline-formula> <tex-math notation="LaTeX">K </tex-math></inline-formula>-means algorithm. In this regard, different dimensionality reduction approaches for the <inline-formula> <tex-math notation="LaTeX">K </tex-math></inline-formula>-means algorithm have been designed recently, leading to algorithms that have proved to generate competitive clusterings. Unfortunately, most of these techniques tend to have fairly high computational costs and/or might not be easy to parallelize. In this article, we propose a fully parallelizable feature selection technique intended for the <inline-formula> <tex-math notation="LaTeX">K </tex-math></inline-formula>-means algorithm. The proposal is based on a novel feature relevance measure that is closely related to the <inline-formula> <tex-math notation="LaTeX">K </tex-math></inline-formula>-means error of a given clustering. Given a disjoint partition of the features, the technique consists of obtaining a clustering for each subset of features and selecting the <inline-formula> <tex-math notation="LaTeX">m </tex-math></inline-formula> features with the highest relevance measure. The computational cost of this approach is just <inline-formula> <tex-math notation="LaTeX">\mathcal {O}(m\cdot \max \{n\cdot K,\log m\}) </tex-math></inline-formula> per subset of features. We additionally provide a theoretical analysis on the quality of the obtained solution via our proposal and empirically analyze its performance with respect to well-known feature selection and feature extraction techniques. Such an analysis shows that our proposal consistently obtains the results with lower <inline-formula> <tex-math notation="LaTeX">K </tex-math></inline-formula>-means error than all the considered feature selection techniques: Laplacian scores, maximum variance, multicluster feature selection, and random selection while also requiring similar or lower computational times than these approaches. Moreover, when compared with feature extraction techniques, such as random projections, the proposed approach also shows a noticeable improvement in both error and computational time. |
|---|---|
| Bibliography: | ObjectType-Article-1 SourceType-Scholarly Journals-1 ObjectType-Feature-2 content type line 14 content type line 23 |
| ISSN: | 2162-237X 2162-2388 2162-2388 |
| DOI: | 10.1109/TNNLS.2020.3002576 |