Volume 37, Issue 10 pp. 7857-7887
RESEARCH ARTICLE

Generative and discriminative infinite restricted Boltzmann machine training

Qianglong Wang

Qianglong Wang

School of Electronics and Information, Northwestern Polytechnical University, Xi'an, Shaanxi, China

Search for more papers by this author
Xiaoguang Gao

Corresponding Author

Xiaoguang Gao

School of Electronics and Information, Northwestern Polytechnical University, Xi'an, Shaanxi, China

Correspondence Xiaoguang Gao, School of Electronics and Information, Northwestern Polytechnical University, 710129 Xi'an, Shaanxi, China.

Email: [email protected]

Search for more papers by this author
Kaifang Wan

Kaifang Wan

School of Electronics and Information, Northwestern Polytechnical University, Xi'an, Shaanxi, China

Search for more papers by this author
Zijian Hu

Zijian Hu

School of Electronics and Information, Northwestern Polytechnical University, Xi'an, Shaanxi, China

Search for more papers by this author
First published: 19 May 2022

Abstract

As one of the essential deep learning models, a restricted Boltzmann machine (RBM) is a commonly used generative training model. By adaptively growing the size of the hidden units, infinite RBM (IRBM) is obtained, which possesses an excellent property of automatically choosing the hidden layer size depending on a specific task. An IRBM presents a competitive generative capability with the traditional RBM. First, a generative model called Gaussian IRBM (GIRBM) is proposed to deal with practical scenarios from the perspective of data discretization. Subsequently, a discriminative IRBM (DIRBM) and a discriminative GIRBM (DGIRBM) are established to solve classification problems by attaching extra-label units next to the input layer. They are motivated by the fact that a discriminative variant of an RBM can complete an individual framework for classification with better performance than some standard classifiers. Remarkably, the proposed models retain both generative and discriminative properties synchronously, that is, they can reconstruct data effectively and be established in considerable self-contained classifiers. The experimental results on image classification (both large and small), text identification, and facial recognition (both clean and noisy) reflect that a DIRBM and a DGIRBM are superior to some state-of-the-art RBM models in terms of the reconstruction error and the classification accuracy. Intuitively, they require models to avoid utilizing more hidden units than needed when confronted with various sizes of data, prioritizing smaller networks. In addition, the proposed models behave more robustly than other classic classifiers when dealing with noisy facial recognition.

DATA AVAILABILITY STATEMENT

The data sets generated during and/or analyzed during the current study are available in the following website.

CalTech101 Silhouettes data set: http://people.cs.umass.edu/marlin/data.shtml

MNIST: http://yann.lecun.com/exdb/mnist/

20 Newsgroups data set: https://cs.nyu.edu/roweis/data.html

20 Newsgroups data set: http://www.qwone.com/jason/20Newsgroups/

Olivetti data set: https://cs.nyu.edu/roweis/data.html

Yale Face: http://cvc.cs.yale.edu/cvc/projects/yalefaces/

The full text of this article hosted at iucr.org is unavailable due to technical difficulties.