TY - GEN
T1 - AutoClustering
T2 - 18th IEEE International Conference on Data Mining Workshops, ICDMW 2018
AU - Kimura, Masaomi
N1 - Publisher Copyright:
© 2018 IEEE.
PY - 2019/2/7
Y1 - 2019/2/7
N2 - Since a clustering process can be regarded as a map of data to cluster labels, it should be natural to employ a deep learning technique, especially a feed-forward neural network, to realize the clustering method. In this study, we discussed a novel clustering method realized only by a feed-forward neural network. Unlike self-organizing maps and growing neural gas networks, the proposed method is compatible with deep learning neural networks. The proposed method has three parts: A map of records to clusters (encoder), a map of clusters to their exemplars (decoder), and a loss function to measure positional closeness between the records and the exemplars. In order to accelerate clustering performance, we proposed an improved activation function at the encoder, which migrates a soft-max function to a max function continuously. Though most of the clustering methods require the number of clusters in advance, the proposed method naturally provides the number of clusters as the number of unique one-hot vectors obtained as a result. We also discussed the existence of local minima of the loss function and their relationship to clusters.
AB - Since a clustering process can be regarded as a map of data to cluster labels, it should be natural to employ a deep learning technique, especially a feed-forward neural network, to realize the clustering method. In this study, we discussed a novel clustering method realized only by a feed-forward neural network. Unlike self-organizing maps and growing neural gas networks, the proposed method is compatible with deep learning neural networks. The proposed method has three parts: A map of records to clusters (encoder), a map of clusters to their exemplars (decoder), and a loss function to measure positional closeness between the records and the exemplars. In order to accelerate clustering performance, we proposed an improved activation function at the encoder, which migrates a soft-max function to a max function continuously. Though most of the clustering methods require the number of clusters in advance, the proposed method naturally provides the number of clusters as the number of unique one-hot vectors obtained as a result. We also discussed the existence of local minima of the loss function and their relationship to clusters.
KW - Autoencoder
KW - Clustering
KW - Feed forward neural network
UR - http://www.scopus.com/inward/record.url?scp=85062848726&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85062848726&partnerID=8YFLogxK
U2 - 10.1109/ICDMW.2018.00102
DO - 10.1109/ICDMW.2018.00102
M3 - Conference contribution
AN - SCOPUS:85062848726
T3 - IEEE International Conference on Data Mining Workshops, ICDMW
SP - 659
EP - 666
BT - Proceedings - 18th IEEE International Conference on Data Mining Workshops, ICDMW 2018
A2 - Li, Zhenhui
A2 - Yu, Jeffrey
A2 - Tong, Hanghang
A2 - Zhu, Feida
PB - IEEE Computer Society
Y2 - 17 November 2018 through 20 November 2018
ER -