TY - JOUR
T1 - Context-sensitive intra-class clustering
AU - Yu, Yingwei
AU - Gutierrez-Osuna, Ricardo
AU - Choe, Yoonsuck
N1 - KAUST Repository Item: Exported on 2020-10-01
Acknowledged KAUST grant number(s): KUSC1-016-04
Acknowledgements: This publication is based in part on work supported by Award No. KUSC1-016-04, made by King Abdullah University of Science and Technology (KAUST).
This publication acknowledges KAUST support, but has no KAUST affiliated authors.
PY - 2014/2
Y1 - 2014/2
N2 - This paper describes a new semi-supervised learning algorithm for intra-class clustering (ICC). ICC partitions each class into sub-classes in order to minimize overlap across clusters from different classes. This is achieved by allowing partitioning of a certain class to be assisted by data points from other classes in a context-dependent fashion. The result is that overlap across sub-classes (both within- and across class) is greatly reduced. ICC is particularly useful when combined with algorithms that assume that each class has a unimodal Gaussian distribution (e.g., Linear Discriminant Analysis (LDA), quadratic classifiers), an assumption that is not always true in many real-world situations. ICC can help partition non-Gaussian, multimodal distributions to overcome such a problem. In this sense, ICC works as a preprocessor. Experiments with our ICC algorithm on synthetic data sets and real-world data sets indicated that it can significantly improve the performance of LDA and quadratic classifiers. We expect our approach to be applicable to a broader class of pattern recognition problems where class-conditional densities are significantly non-Gaussian or multi-modal. © 2013 Elsevier Ltd. All rights reserved.
AB - This paper describes a new semi-supervised learning algorithm for intra-class clustering (ICC). ICC partitions each class into sub-classes in order to minimize overlap across clusters from different classes. This is achieved by allowing partitioning of a certain class to be assisted by data points from other classes in a context-dependent fashion. The result is that overlap across sub-classes (both within- and across class) is greatly reduced. ICC is particularly useful when combined with algorithms that assume that each class has a unimodal Gaussian distribution (e.g., Linear Discriminant Analysis (LDA), quadratic classifiers), an assumption that is not always true in many real-world situations. ICC can help partition non-Gaussian, multimodal distributions to overcome such a problem. In this sense, ICC works as a preprocessor. Experiments with our ICC algorithm on synthetic data sets and real-world data sets indicated that it can significantly improve the performance of LDA and quadratic classifiers. We expect our approach to be applicable to a broader class of pattern recognition problems where class-conditional densities are significantly non-Gaussian or multi-modal. © 2013 Elsevier Ltd. All rights reserved.
UR - http://hdl.handle.net/10754/597845
UR - https://linkinghub.elsevier.com/retrieve/pii/S016786551300189X
UR - http://www.scopus.com/inward/record.url?scp=84891633909&partnerID=8YFLogxK
U2 - 10.1016/j.patrec.2013.04.031
DO - 10.1016/j.patrec.2013.04.031
M3 - Article
SN - 0167-8655
VL - 37
SP - 85
EP - 93
JO - Pattern Recognition Letters
JF - Pattern Recognition Letters
IS - 1
ER -