Vector-Quantization by density matching in the minimum Kullback-Leibler divergence sense



AbstractRepresentation of a large set of high-dimensional data is a fundamental problem in many applications such as communications and biomedical systems. The problem has been tackled by encoding the data with a compact set of code-vectors called processing elements. In this study, we propose a vector quantization technique that encodes the information in the data using concepts derived from information theoretic learning. The algorithm minimizes a cost function based on the Kullback-Liebler divergence to match the distribution of the processing elements with the distribution of the data. The performance of this algorithm is demonstrated on synthetic data as well as on an edge-image of a face. Comparisons are provided with some of the existing algorithms such as LBG and SOM.
TypeConference paper [With referee]
ConferenceIEEE International Conference on Neural Networks - Conference Proceedings
Year2004    Vol. 1    pp. 105--109
ISBN / ISSN10987576
Electronic version(s)[pdf]
BibTeX data [bibtex]
IMM Group(s)Intelligent Signal Processing