Algorithms for Better Representation and Faster Learning in Radial Basis Function Networks

Part of Advances in Neural Information Processing Systems 2 (NIPS 1989)

Bibtex Metadata Paper

Authors

Avijit Saha, James Keeler

Abstract

in

learning, using radial basis functions

In this paper we present upper bounds for the learning rates for hybrid models that employ a combination of both self-organized and supervised to build receptive field representations the hidden units. The learning performance in such networks with nearest neighbor heuristic can be improved upon by multiplying the individual receptive field widths by a suitable overlap factor. We present results indicat!ng optimal values for such overlap factors. We also present a new algorithm for determining receptive field centers. This method negotiates more hidden units in the regions of the input space as a function of the output and is conducive to better learning when the number of patterns (hidden units) is small.