An Adaptive Deep Learning Optimization Method Based on Radius of Curvature

An adaptive clamping method (SGD-MS) based on the radius of curvature is designed to alleviate the local optimal oscillation problem in deep neural network, which combines the radius of curvature of the objective function and the gradient descent of the optimizer. The radius of curvature is consider...

Full description

Saved in:
Bibliographic Details
Main Authors: Jiahui Zhang, Xinhao Yang, Ke Zhang, Chenrui Wen
Format: article
Language:EN
Published: Hindawi Limited 2021
Subjects:
Online Access:https://doaj.org/article/acab22f5e532433c807e6c5f9bb9d3fc
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:An adaptive clamping method (SGD-MS) based on the radius of curvature is designed to alleviate the local optimal oscillation problem in deep neural network, which combines the radius of curvature of the objective function and the gradient descent of the optimizer. The radius of curvature is considered as the threshold to separate the momentum term or the future gradient moving average term adaptively. In addition, on this basis, we propose an accelerated version (SGD-MA), which further improves the convergence speed by using the method of aggregated momentum. Experimental results on several datasets show that the proposed methods effectively alleviate the local optimal oscillation problem and greatly improve the convergence speed and accuracy. A novel parameter updating algorithm is also provided in this paper for deep neural network.