首页    期刊浏览 2025年09月16日 星期二
登录注册

文章基本信息

  • 标题:IMPROVING THE ACCURACY OF GRADIENT DESCENT BACK PROPAGATION ALGORITHM (GDAM) ON CLASSIFICATION PROBLEMS
  • 本地全文:下载
  • 作者:M. Z. Rehman ; N. M. Nawi
  • 期刊名称:International Journal of New Computer Architectures and their Applications
  • 印刷版ISSN:2220-9085
  • 出版年度:2011
  • 卷号:1
  • 期号:4
  • 页码:838-847
  • 出版社:Society of Digital Information and Wireless Communications
  • 摘要:The traditional Back-propagation Neural Network (BPNN) Algorithm is widely used in solving many real time problems in world. But BPNN possesses a problem of slow convergence and convergence to local minima. Previously, several modifications are suggested to improve the convergence rate of Gradient Descent Back-propagation algorithm such as careful selection of initial weights and biases, learning rate, momentum, network topology, activation function and 'gain' value in the activation function. This research proposed an algorithm for improving the current working performance of Back-propagation algorithm by adaptively changing the momentum value and at the same time keeping the 'gain' parameter fixed for all nodes in the neural network. The performance of the proposed method known as 'Gradient Descent Method with Adaptive Momentum (GDAM)' is compared with the performances of 'Gradient Descent Method with Adaptive Gain (GDM-AG)' and 'Gradient Descent with Simple Momentum (GDM)'. The efficiency of the proposed method is demonstrated by simulations on five classification problems. Results show that GDAM can be used as an alternative approach for BPNN because it demonstrate better accuracy ratio on the chosen classification problems.
  • 关键词:gradient descent; neural network; adaptive ; momentum; adaptive gain; back- ; propagation
国家哲学社会科学文献中心版权所有