![]() |
个人信息Personal Information
教授 博士生导师
招生学科专业:
网络空间安全 -- 【招收硕士研究生】 -- 计算机科学与技术学院
计算机科学与技术 -- 【招收博士、硕士研究生】 -- 人工智能学院
软件工程 -- 【招收硕士研究生】 -- 人工智能学院
电子信息 -- 【招收博士、硕士研究生】 -- 人工智能学院
学历:南京航空航天大学
学位:工学博士学位
所在单位:计算机科学与技术学院/人工智能学院/软件学院
电子邮箱:
Batch-normalized Mlpconv-wise supervised pre-training network in network
点击次数:
所属单位:计算机科学与技术学院/人工智能学院/软件学院
发表刊物:APPLIED INTELLIGENCE
关键字:Deep learning (DL) Mlpconv-wise supervised pre-training network in network (MPNIN) Network in network (NIN) structure Mlpconv layer Batch normalization
摘要:Deep multi-layered neural networks have nonlinear levels that allow them to represent highly varying nonlinear functions compactly. In this paper, we propose a new deep architecture with enhanced model discrimination ability that we refer to as mlpconv-wise supervised pre-training network in network (MPNIN). The process of information abstraction is facilitated within the receptive fields for MPNIN. The proposed architecture uses the framework of the recently developed NIN structure, which slides a universal approximator, such as a multilayer perceptron with rectifier units, across an image to extract features. However, the random initialization of NIN can produce poor solutions to gradient-based optimization. We use mlpconv-wise supervised pre-training to remedy this defect because this pre-training technique may contribute to overcoming the difficulties of training deep networks by better initializing the weights in all the layers. Moreover, batch normalization is applied to reduce internal covariate shift by pre-conditioning the model. Empirical investigations are conducted on the Mixed National Institute of Standards and Technology (MNIST), the Canadian Institute for Advanced Research (CIFAR-10), CIFAR-100, the Street View House Numbers (SVHN), the US Postal (USPS), Columbia University Image Library (COIL20), COIL100 and Olivetti Research Ltd (ORL) datasets, and the results verify the effectiveness of the proposed MPNIN architecture.
ISSN号:0924-669X
是否译文:否
发表时间:2018-01-01
合写作者:韩晓猛
通讯作者:戴群