Parameter learning for alpha-integration
SCIE
SCOPUS
- Title
- Parameter learning for alpha-integration
- Authors
- Choi, H; Choi, S; Choe, Y
- Date Issued
- 2013-06
- Publisher
- MIT Press Journals
- Abstract
- In pattern recognition, data integration is an important issue, and when properly done, it can lead to improved performance. Also, data integration can be used to help model and understand multimodal processing in the brain. Amari proposed alpha-integration as a principled way of blending multiple positive measures (e.g., stochastic models in the form of probability distributions), enabling an optimal integration in the sense of minimizing the alpha-divergence. It also encompasses existing integration methods as its special case, for example, a weighted average and an exponential mixture. The parameter a determines integration characteristics, and the weight vector w assigns the degree of importance to each measure. In most work, however, alpha and w are given in advance rather than learned. In this letter, we present a parameter learning algorithm for learning a and w from data when multiple integrated target values are available. Numerical experiments on synthetic as well as real-world data demonstrate the effectiveness of the proposed method.
- Keywords
- DIVERGENCE; MODELS
- URI
- https://oasis.postech.ac.kr/handle/2014.oak/15480
- DOI
- 10.1162/NECO_a_00445
- ISSN
- 0899-7667
- Article Type
- Article
- Citation
- Neural Computation, vol. 25, no. 6, page. 1585 - 1604, 2013-06
- Files in This Item:
- There are no files associated with this item.
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.