Affiliation of Author(s):计算机科学与技术学院/人工智能学院/软件学院
Journal:APPLIED INTELLIGENCE
Key Words:Deep learning (DL) Mlpconv-wise supervised pre-training network in network (MPNIN) Network in network (NIN) structure Mlpconv layer Batch normalization
Abstract:Deep multi-layered neural networks have nonlinear levels that allow them to represent highly varying nonlinear functions compactly. In this paper, we propose a new deep architecture with enhanced model discrimination ability that we refer to as mlpconv-wise supervised pre-training network in network (MPNIN). The process of information abstraction is facilitated within the receptive fields for MPNIN. The proposed architecture uses the framework of the recently developed NIN structure, which slides a universal approximator, such as a multilayer perceptron with rectifier units, across an image to extract features. However, the random initialization of NIN can produce poor solutions to gradient-based optimization. We use mlpconv-wise supervised pre-training to remedy this defect because this pre-training technique may contribute to overcoming the difficulties of training deep networks by better initializing the weights in all the layers. Moreover, batch normalization is applied to reduce internal covariate shift by pre-conditioning the model. Empirical investigations are conducted on the Mixed National Institute of Standards and Technology (MNIST), the Canadian Institute for Advanced Research (CIFAR-10), CIFAR-100, the Street View House Numbers (SVHN), the US Postal (USPS), Columbia University Image Library (COIL20), COIL100 and Olivetti Research Ltd (ORL) datasets, and the results verify the effectiveness of the proposed MPNIN architecture.
ISSN No.:0924-669X
Translation or Not:no
Date of Publication:2018-01-01
Co-author:韩晓猛
Correspondence Author:dq
Date of Publication:2018-01-01
戴群
+
Education Level:南京航空航天大学
Paper Publications
Batch-normalized Mlpconv-wise supervised pre-training network in network
Date of Publication:2018-01-01 Hits: