Interactive Media Systems, TU Wien

Robust and sparse k-means clustering in high dimension

By Peter Filzmoser, Sarka Brodinova, Thomas Ortner, Christian Breiteneder, and Maia Rohm


We introduce a robust k-means-based clustering method for high-dimensional data where not only outliers but also a large number of noise variables are very likely to be present [4]. Although Kondo et al. [2] already addressed such an application scenario, our approach goes even further. Firstly, the introduced method is designed to identify clusters, informative variables, and outliers simultaneously. Secondly, the proposed clustering technique additionally aims at optimizing required parameters, e.g. the number of clusters. This is a great advantage over most existing methods. Moreover, the robustness aspect is achieved through a robust initialization [3] and a proposed weighting function using the Local Outlier Factor [1]. The weighting function provides a valuable source of information about the outlyingness of each observation for a subsequent outlier detection. In order to reveal both clusters and informative variables properly, the approach uses a lasso-type penalty [5]. The method has thoroughly been tested on simulated as well as on real high-dimensional datasets. The conducted experiments demonstrated a great ability of the clustering method to identify clusters, outliers, and informative variables.


P. Filzmoser, S. Brodinova, T. Ortner, C. Breiteneder, M. Rohm: "Robust and sparse k-means clustering in high dimension"; Talk: Seminarvortrag an der JKU Linz, Linz (invited); 01-23-2020 - 01-24-2020.


Click into the text area and press Ctrl+A/Ctrl+C or ⌘+A/⌘+C to copy the BibTeX into your clipboard… or download the BibTeX.