作者投稿和查稿 主编审稿 专家审稿 编委审稿 远程编辑

计算机工程

• 人工智能及识别技术 • 上一篇    下一篇

前馈神经网络导数特性分析

魏 海a,杨华舒b,苏志敏a,桂 跃c,董梦思a   

  1. (昆明理工大学 a. 电力工程学院;b. 国土资源工程学院;c. 建筑工程学院,昆明 650500)
  • 收稿日期:2013-06-20 出版日期:2014-07-15 发布日期:2014-07-14
  • 作者简介:魏 海(1975-),男,副教授、博士,主研方向:神经网络;杨华舒、苏志敏,教授;桂 跃,副教授、博士,董梦思,硕士研究生。
  • 基金资助:
    国家自然科学基金资助项目(51069003);云南省应用基础研究基金资助项目(2010ZC048)。

Derivatives Feature Analysis of Feedforward Neural Networks

WEI Hai  a, YANG Hua-shu  b, SU Zhi-min  a, GUI Yue  c, DONG Meng-si  a   

  1. (a. Faculty of Electric Power Engineering; b. Faculty of Land Resources Engineering; c. Faculty of Architecture Engineering, Kunming University of Science and Technology, Kunming 650500, China)
  • Received:2013-06-20 Online:2014-07-15 Published:2014-07-14

摘要: 为分析前馈神经网络输出量的一阶、二阶偏导数特性,从一层网络结构入手,推导网络输出量的一阶偏导数,应用链式求导法则,推导多层网络输出量的一阶、二阶偏导数的计算公式。在此基础上推导网络的三阶偏导数,并针对二层结构网络,在其输出层激活函数为线性函数时,推导出该网络对输入量的高阶偏导数计算公式。实例分析结果表明,前馈神经网络一阶、二阶偏导数值的精度比网络输出值的精度要低,尤其是在区间的边界上有时会出现较大的偏差。网络的一阶、二阶偏导数值的精度也会随着隐含层神经元数量的增加明显降低,在基本相同的网络训练精度下,隐含层神经元较多的网络比神经元少的网络导数特性差。

关键词: 前馈神经网络, 偏导数, 线性激活函数, 精度, 网络体系结构, 网络输出

Abstract: In order to analyze first and second order partial derivative feature of feedforward neural networks with respect to its inputs, one layered architecture network is chosen to deduce first order partial derivative of network. Chain rule is employed to derive formulas to compute partial derivatives of multilayer architecture networks. On the basis of that, third order partial derivative of networks can be gained easily. And considering linear activation function in output layer of two layered networks, higher order partial derivatives of networks with respect of its inputs can be obtained. Case analysis shows that accuracy of first and second order partial derivative of feedforward neural networks is far less than that of output of networks, especially in the boundary area of interval of input the error between stimulation value and real value is very significant. Moreover, accuracy of first and second order derivative of network decreases greatly with increase of the number of neurons in hidden layer. Consequently, under the condition of networks with equivalent training accuracy, the networks with less neurons in hidden layer has better derivative performance than that with more neurons in hidden layer.

Key words: feedforward neural networks, partial derivative, linear activation function, accuracy, network architecture, output of networks

中图分类号: