Dissimilarity Measure Machines

Abstract : This paper presents a dissimilarity-based discriminative framework for learning from data coming in the form of probability distributions. Departing from the use of positive kernel-based methods, we build upon embeddings based on dissimilarities tailored for distribution. We enable this by extending \citet{balcan2008theory}'s theory of learning with similarity functions to the case of distribution-shaped data. Then, we show that several learning guarantees of the dissimilarity still hold when estimated from empirical distributions. Algorithmically, the proposed approach consists in building features from pairwise dissimilarities and in learning a linear decision function in this new feature space. Our experimental results show that this dissimilarity-based approach works better than the so-called support measure machines or the sliced Wasserstein kernel, and that among several dissimilarities including Kullback-Leibler divergence and Maximum Mean Discrepancy, the entropy-regularized Wasserstein distance provides the best compromise between computational efficiency and accuracy.
Complete list of metadatas

https://hal.archives-ouvertes.fr/hal-01717940
Contributor : Alain Rakotomamonjy <>
Submitted on : Wednesday, November 7, 2018 - 11:08:29 PM
Last modification on : Friday, September 13, 2019 - 9:50:02 AM
Long-term archiving on : Friday, February 8, 2019 - 4:12:15 PM

Files

WDMM.pdf
Files produced by the author(s)

Identifiers

  • HAL Id : hal-01717940, version 2
  • ARXIV : 1803.00250

Citation

Alain Rakotomamonjy, Abraham Traoré, Maxime Berar, Rémi Flamary, Nicolas Courty, et al.. Dissimilarity Measure Machines. 2018. ⟨hal-01717940v2⟩

Share

Metrics

Record views

319

Files downloads

458