A framework to uncover multiple alternative clusterings

Verfasser / Beitragende:
[Xuan Dang, James Bailey]
Ort, Verlag, Jahr:
2015
Enthalten in:
Machine Learning, 98/1-2(2015-01-01), 7-30
Format:
Artikel (online)
ID: 605478139
LEADER caa a22 4500
001 605478139
003 CHVBK
005 20210128100404.0
007 cr unu---uuuuu
008 210128e20150101xx s 000 0 eng
024 7 0 |a 10.1007/s10994-013-5338-7  |2 doi 
035 |a (NATIONALLICENCE)springer-10.1007/s10994-013-5338-7 
245 0 2 |a A framework to uncover multiple alternative clusterings  |h [Elektronische Daten]  |c [Xuan Dang, James Bailey] 
520 3 |a Clustering is often referred to as unsupervised learning which aims at uncovering hidden structures from data. Unfortunately, though widely being used as one of the principal tools to understand the data, most conventional clustering techniques are limited in achieving this goal since they only attempt to find a single clustering solution from the data. For many real-world applications, especially those being described in high dimensional data, it is common to see that the data can be grouped into different yet meaningful ways. This gives rise to the recently emerging research area of mining alternative clusterings. In this paper, we propose a framework named MACL that is capable of discovering multiple alternative clusterings from a given dataset. MACL seeks alternative clusterings in sequence and a novel solution is found by conditioning on all previously known clusterings. The framework takes a mathematically appealing approach by combining the maximum likelihood framework and mutual information. Consequently, its resultant clustering quality is achieved by the likelihood maximization over the data whereas the dissimilarity is ensured by the minimization over the information sharing amongst alternatives. We test the proposed algorithm on both synthetic and real-world datasets and the experimental results demonstrate its potential in discovering multiple alternative clusterings from data. 
540 |a The Author(s), 2013 
690 7 |a Unsupervised learning  |2 nationallicence 
690 7 |a Alternative clustering  |2 nationallicence 
690 7 |a Expectation maximization  |2 nationallicence 
690 7 |a Mutual information  |2 nationallicence 
700 1 |a Dang  |D Xuan  |u Department of Computer Science, Aarhus University, 8200, Aarhus N, Denmark  |4 aut 
700 1 |a Bailey  |D James  |u Department of Computing and Information Systems, The University of Melbourne, 3010, Melbourne, VIC, Australia  |4 aut 
773 0 |t Machine Learning  |d Springer US; http://www.springer-ny.com  |g 98/1-2(2015-01-01), 7-30  |x 0885-6125  |q 98:1-2<7  |1 2015  |2 98  |o 10994 
856 4 0 |u https://doi.org/10.1007/s10994-013-5338-7  |q text/html  |z Onlinezugriff via DOI 
898 |a BK010053  |b XK010053  |c XK010000 
900 7 |a Metadata rights reserved  |b Springer special CC-BY-NC licence  |2 nationallicence 
908 |D 1  |a research-article  |2 jats 
949 |B NATIONALLICENCE  |F NATIONALLICENCE  |b NL-springer 
950 |B NATIONALLICENCE  |P 856  |E 40  |u https://doi.org/10.1007/s10994-013-5338-7  |q text/html  |z Onlinezugriff via DOI 
950 |B NATIONALLICENCE  |P 700  |E 1-  |a Dang  |D Xuan  |u Department of Computer Science, Aarhus University, 8200, Aarhus N, Denmark  |4 aut 
950 |B NATIONALLICENCE  |P 700  |E 1-  |a Bailey  |D James  |u Department of Computing and Information Systems, The University of Melbourne, 3010, Melbourne, VIC, Australia  |4 aut 
950 |B NATIONALLICENCE  |P 773  |E 0-  |t Machine Learning  |d Springer US; http://www.springer-ny.com  |g 98/1-2(2015-01-01), 7-30  |x 0885-6125  |q 98:1-2<7  |1 2015  |2 98  |o 10994