Volume 23, Issue 4 pp. 92-104
Article
Full Access

A new information criterion combined with cross-validation method to estimate generalization capability

Yasuhiro Wada

Yasuhiro Wada

ATR Auditory and Visual Perception Research Laboratories, Kyoto, Japan 619-02

Yasuhiro Wada: received a B.E. and an M.E. in Engineering from Tokyo Institute of Technology in 1980 and 1982, respectively. In 1982, he joined Kawasaki Steel Co., Ltd. Since 1989, he has been on loan to ATR Auditory and Visual Perception Research Laboratories. His research interests include neural network models and motor learning control.

Search for more papers by this author
Mitsuo Kawato

Mitsuo Kawato

Members

ATR Auditory and Visual Perception Research Laboratories, Kyoto, Japan 619-02

Mitsuo Kawato: received a B.S. in Physics from Tokyo University in 1976 and an M.E. and a Ph.D. in Biophysical Engineering from Osaka University in 1978 and 1981, respectively. From 1981 to 1987, he was a member of the faculty of Osaka University. In 1987, he became University Lecturer of Biophysical Engineering, Faculty of Engineering Science, Osaka University. In 1988, he moved to ATR Auditory and Visual Perception Research Laboratories. His research interests include computational neuroscience and its application to engineering problems.

Search for more papers by this author
First published: 1992
Citations: 3

Abstract

Neural network learning processes use only a limited number of examples of a given problem. Thus, generally speaking, it is not necessarily theoretically guaranteed that the trained network can give correct answers for unknown examples. A new method of selecting the optimal neural network structure with maximum generalization capability is proposed. In statistical mathematics, several information criteria, such as AIC (Akaike's information criterion), BIC (Bayesian information criterion), and MDL (minimum description length), are used widely to select a suitable model. Applications of these criteria were quite successful, especially for linear models. These criteria assume that the model parameters are estimated correctly by using the maximum likelihood method. Unfortunately, however, this assumption does not hold for conventional iterative learning processes such as backpropagation in multilayer perceptrons or Boltzmann machine learning. Thus, we should not apply AIC directly to the selection of the optimal neural network structure.

In this paper, by expanding AIC, a new information criterion is proposed that can estimate generalization capability without the maximum likelihood estimator of synaptic weights. The cross-validation method is used to calculate the new information criterion. By computer simulation, we show that the proposed information criterion can accurately predict the generalization capability of multilayer perceptrons, and thus the optimal number of hidden units can be determined.

The full text of this article hosted at iucr.org is unavailable due to technical difficulties.