TY - JOUR
T1 - Stochastic learning of multi-instance dictionary for earth mover’s distance-based histogram comparison
AU - Fan, Jihong
AU - Liang, Ru-Ze
N1 - KAUST Repository Item: Exported on 2020-10-01
Acknowledgements: The work was funded by Science and Technology project under Grant No. 12531826 of Education Department, Heilongjiang, China.
PY - 2016/9/17
Y1 - 2016/9/17
N2 - Dictionary plays an important role in multi-instance data representation. It maps bags of instances to histograms. Earth mover’s distance (EMD) is the most effective histogram distance metric for the application of multi-instance retrieval. However, up to now, there is no existing multi-instance dictionary learning methods designed for EMD-based histogram comparison. To fill this gap, we develop the first EMD-optimal dictionary learning method using stochastic optimization method. In the stochastic learning framework, we have one triplet of bags, including one basic bag, one positive bag, and one negative bag. These bags are mapped to histograms using a multi-instance dictionary. We argue that the EMD between the basic histogram and the positive histogram should be smaller than that between the basic histogram and the negative histogram. Base on this condition, we design a hinge loss. By minimizing this hinge loss and some regularization terms of the dictionary, we update the dictionary instances. The experiments over multi-instance retrieval applications shows its effectiveness when compared to other dictionary learning methods over the problems of medical image retrieval and natural language relation classification. © 2016 The Natural Computing Applications Forum
AB - Dictionary plays an important role in multi-instance data representation. It maps bags of instances to histograms. Earth mover’s distance (EMD) is the most effective histogram distance metric for the application of multi-instance retrieval. However, up to now, there is no existing multi-instance dictionary learning methods designed for EMD-based histogram comparison. To fill this gap, we develop the first EMD-optimal dictionary learning method using stochastic optimization method. In the stochastic learning framework, we have one triplet of bags, including one basic bag, one positive bag, and one negative bag. These bags are mapped to histograms using a multi-instance dictionary. We argue that the EMD between the basic histogram and the positive histogram should be smaller than that between the basic histogram and the negative histogram. Base on this condition, we design a hinge loss. By minimizing this hinge loss and some regularization terms of the dictionary, we update the dictionary instances. The experiments over multi-instance retrieval applications shows its effectiveness when compared to other dictionary learning methods over the problems of medical image retrieval and natural language relation classification. © 2016 The Natural Computing Applications Forum
UR - http://hdl.handle.net/10754/622254
UR - http://link.springer.com/article/10.1007%2Fs00521-016-2603-2
UR - http://www.scopus.com/inward/record.url?scp=84988384110&partnerID=8YFLogxK
U2 - 10.1007/s00521-016-2603-2
DO - 10.1007/s00521-016-2603-2
M3 - Article
SN - 0941-0643
VL - 29
SP - 733
EP - 743
JO - Neural Computing and Applications
JF - Neural Computing and Applications
IS - 10
ER -