为提高受限玻尔兹曼机(Restricted Boltzmann Machine,RBM)数据学习能力和抑制训练的特征同质化问题,提出一种随机受限玻尔兹曼机组(RandomRBM Group,RRBMG)设计.对观测数据进行随机维度组合,在随机维度组合的基础上构建子RBM群组并实施训练,随后依据神经网络的层数选择模型特征组合方式,针对浅层结构设置为均值组合方式,针对深层模型设置为隐单元叠加方式.理论分析表明,随着组内模型数目的增加,RRBMG所要学习的训练目标将逐渐接近于标准RBM的训练目标,并且能够有效减少特征同质化带来的影响;实验结果表明,与衰落机制相比,RRBMG能够有效提高RBM的特征学习能力,应用所组建的浅层结构和深层结构特征,将MNIST(Mixed National Institute of Standards and Technology)数据库实验的分类准确率分别提高了2%和0.4%.
To improve the restricted Boltzmann machine (RBM)’s data generalization ability and resolve the features homogenization problem, a random RBM group (RRBMG) design is proposed. The dimensions of observation data were randomly divided into groups, and the childRBMs were built based on the combined data group. Two methods based on the structural stories were used to compose hidden units’ layer finally, shallow structure by mean output, and deep structure through the formation of highlevel hidden units’ layer. The theoretical analysis shows that, with the increase of models’ number in the group, the training objectives of RRBMG will gradually approach the training objectives of standard RBM, and can effectively reduce the impact of feature homogeneity. The experimental results show that, compared with dropout algorithm, the proposed RRBMG can effectively improve the feature learning ability of RBM, and use the shallow structure and deep structure features to increase the classification accuracy of mixed national institute of standards and technology (MNIST) database experiment by 2% and 0.4%.
[1]LEE H, EKANADHAM C, NG A Y. Sparse deep belief net model for visual area V2[C]∥Proceedings of Advances in Neural Information Processing Systems. NY, United States: Curran Associates Inc, 2008: 873880.
[2]LUO H, SHEN R, NIU C. Sparse group restricted Boltzmann machines[C]∥Proceedings of 25th AAAI Conference on Artificial Intelligence and the 23rd Innovative Applications of Artificial Intelligence Conference. CA, United States: AI Access Foundation, 2011: 429434.
[3]JIN N, Zhang J S, ZHANG C X. A sparseresponse deep belief network based on rate distortion theory[J]. Pattern Recognition, 2014, 47(9): 31793191.
[4]BREULEUX O, BENGIO Y, VINCENT P. Quickly generating representative samples from an RBMderived process[J]. Neural Computation, 2011, 23(8): 20582073.
[5]BARTHELME S, CHOPIN N. The Poisson transform for unnormalised statistical models[J]. Statistics & Computing, 2014, 25(4): 114.
[6]胡洋. 基于马尔可夫链蒙特卡罗方法的RBM学习算法改进[D]. 上海: 上海交通大学计算机科学与工程系, 2012.
[7]TOSUN H, SHEPPARD J W. Training restricted Boltzmann machines with overlapping partitions[C]∥Joint European Conference on Machine Learning and Knowledge Discovery in Databases. Berlin: Springer, 2014:195208.
[8]罗恒. 基于协同过滤视角的受限玻尔兹曼机研究[D]. 上海: 上海交通大学 计算机科学与工程系, 2011.
[9]MIN M R, NING X, CHENG C, et al. Interpretable sparse highorder Boltzmann machines[J]. Communications IET, 2014, 10(1): 614622.
[10]MNIH V, LAROCHELLE H, HINTON G E. Conditional restricted Boltzmann machines for structured output prediction[C]∥Proceedings of the 27th Conference on Uncertainty in Artificial Intelligence. Arlington, United States: AUAI Press, 2011: 514522.
[11]SRIVASTAVA N, HINTON G, KRIZHEVSKY A, et al. Dropout: A simple way to prevent neural networks from overfitting[J]. The Journal of Machine Learning Research, 2014, 15(1): 19291958.
[12]刘凯, 张立民, 张超. 受限玻尔兹曼机的新混合稀疏惩罚机制[J]. 浙江大学学报(工学版), 2015, 49(6): 10701078.
LIU Kai, ZHANG Limin, ZHANG Chao. New hybrid sparse penalty mechanism of restricted Boltzmann machine[J]. Journal of Zhejiang University (Engineering Science), 2015, 49(6): 10701078.
[13]SALAKHUTDINOV R, HINTON G E. Deep Boltzmann machines[C]∥Proceedings of 12th International Conference on Artificial Intelligence and Statistics. MA, United States: Microtome Publishing Brookline, 2009: 448455.
[14]SRIVASTAVA N. Improving neural networks with dropout[D]. Toronto: University of Toronto, 2013.
[15]GOODFELLOW I J, WARDEFARLEY D, MIRZA M, et al. Maxout networks[J]. Computer Science, 2013, 28(3): 13191327.