An analytic distance metric for Gaussian mixture models with application in image retrieval
dc.contributor.author | Sfikas, G. | en |
dc.contributor.author | Constantinopoulos, C. | en |
dc.contributor.author | Likas, A. | en |
dc.contributor.author | Galatsanos, N. P. | en |
dc.date.accessioned | 2015-11-24T17:01:08Z | |
dc.date.available | 2015-11-24T17:01:08Z | |
dc.identifier.issn | 0302-9743 | - |
dc.identifier.uri | https://olympias.lib.uoi.gr/jspui/handle/123456789/10875 | |
dc.rights | Default Licence | - |
dc.title | An analytic distance metric for Gaussian mixture models with application in image retrieval | en |
heal.abstract | In this paper we propose a new distance metric for probability density functions (PDF). The main advantage of this metric is that unlike the popular Kullback-Liebler (KL) divergence it can be computed in closed form when the PDFs are modeled as Gaussian Mixtures (GM). The application in mind for this metric is histogram based image retrieval. We experimentally show that in an image retrieval scenario the proposed metric provides as good results as the KL divergence at a fraction of the computational cost. This metric is also compared to a Bhattacharyya-based distance metric that can be computed in closed form for GMs and is found to produce better results. | en |
heal.access | campus | - |
heal.fullTextAvailability | TRUE | - |
heal.journalName | Artificial Neural Networks: Formal Models and Their Applications - Icann 2005, Pt 2, Proceedings | en |
heal.journalType | peer reviewed | - |
heal.language | en | - |
heal.publicationDate | 2005 | - |
heal.recordProvider | Πανεπιστήμιο Ιωαννίνων. Σχολή Θετικών Επιστημών. Τμήμα Μηχανικών Ηλεκτρονικών Υπολογιστών και Πληροφορικής | el |
heal.type | journalArticle | - |
heal.type.el | Άρθρο Περιοδικού | el |
heal.type.en | Journal article | en |
Αρχεία
Φάκελος/Πακέτο αδειών
1 - 1 of 1
Φόρτωση...
- Ονομα:
- license.txt
- Μέγεθος:
- 1.74 KB
- Μορφότυπο:
- Item-specific license agreed upon to submission
- Περιγραφή: