Avoiding Prototype Proliferation in Incremental Vector Quantization of Large Heterogeneous Datasets

Satizabal, Hector F. (HEC, Lausanne, Switzerland) ; Perez-Uribe, Andres (School of Management and Engineering Vaud, HES-SO // University of Applied Sciences Western Switzerland) ; Tomassini, Marco (HEC, Lausanne, Switzerland)

Vector quantization of large datasets can be carried out by means of an incremental modelling approach where the modelling task is transformed into an incremental task by partitioning or sampling the data, and the resulting datasets are processed by means of an incremental learner. Growing Neural Gas is an incremental vector quantization algorithm with the capabilities of topology-preserving and distribution-matching. Distribution matching can produce overpopulation of prototypes in zones with high density of data. In order to tackle this drawback, we introduce some modifications to the original Growing Neural Gas algorithm by adding three new parameters, one of them controlling the distribution of the codebook and the other two controlling the quantization error and the amount of units in the network. The resulting learning algorithm is capable of efficiently quantizing large datasets presenting high and low density regions while solving the prototype proliferation problem.


Note: SATIZABAL, Hector F. est chercheur à la HES-SO, HEIG-VD depuis 2011.


Keywords:
Faculty:
Ingénierie et Architecture
School:
HEIG-VD
Institute:
ReDS - Reconfigurable & embedded Digital Systems
Publisher:
Berlin, Heidelberg, Springer
Date:
2009-06
Berlin, Heidelberg
Springer
Pagination:
18 p.
Published in:
Studies in Computational Intelligence
DOI:
ISSN:
1860-949X
ISBN:
978-3-642-04511-0
External resources:
Appears in Collection:

Note: The status of this file is: restricted


 Record created 2021-01-26, last modified 2021-01-26

Fulltext:
Download fulltext
PDF

Rate this document:

Rate this document:
1
2
3
 
(Not yet reviewed)