Vector-Quantization by density matching in the minimum Kullback-Leibler divergence sense |
|
Abstract | Representation of a large set of high-dimensional data is a fundamental problem in many applications such as communications and biomedical systems. The problem has been tackled by encoding the data with a compact set of code-vectors called processing elements. In this study, we propose a vector quantization technique that encodes the information in the data using concepts derived from information theoretic learning. The algorithm minimizes a cost function based on the Kullback-Liebler divergence to match the distribution of the processing elements with the distribution of the data. The performance of this algorithm is demonstrated on synthetic data as well as on an edge-image of a face. Comparisons are provided with some of the existing algorithms such as LBG and SOM. |
Type | Conference paper [With referee] |
Conference | IEEE International Conference on Neural Networks - Conference Proceedings |
Year | 2004 Vol. 1 pp. 105--109 |
ISBN / ISSN | 10987576 |
Electronic version(s) | [pdf] |
BibTeX data | [bibtex] |
IMM Group(s) | Intelligent Signal Processing |