首页 | 本学科首页   官方微博 | 高级检索  
     检索      

基于历史梯度平均方差缩减的协同参数更新方法
引用本文:谢涛,张春炯,徐永健.基于历史梯度平均方差缩减的协同参数更新方法[J].电子与信息学报,2022,43(4):956-964.
作者姓名:谢涛  张春炯  徐永健
作者单位:西南大学教育学部智慧教育研究院,重庆,400715;同济大学电子与信息工程学院,上海,201804;西南大学计算机与信息科学学院,重庆,400715
摘    要:随机梯度下降算法(SGD)随机使用一个样本估计梯度,造成较大的方差,使机器学习模型收敛减慢且训练不稳定。该文提出一种基于方差缩减的分布式SGD,命名为DisSAGD。该方法采用历史梯度平均方差缩减来更新机器学习模型中的参数,不需要完全梯度计算或额外存储,而是通过使用异步通信协议来共享跨节点的参数。为了解决全局参数分发存在的“更新滞后”问题,该文采用具有加速因子的学习速率和自适应采样策略:一方面当参数偏离最优值时,增大加速因子,加快收敛速度;另一方面,当一个工作节点比其他工作节点快时,为下一次迭代采样更多样本,使工作节点有更多时间来计算局部梯度。实验表明:DisSAGD显著减少了循环迭代的等待时间,加速了算法的收敛,其收敛速度比对照方法更快,在分布式集群中可以获得近似线性的加速。

关 键 词:梯度下降  机器学习  分布式集群  自适应采样  方差缩减

Collaborative Parameter Update Based on Average Variance Reduction of Historical Gradients
XIE Tao,ZHANG Chunjiong,XU Yongjian.Collaborative Parameter Update Based on Average Variance Reduction of Historical Gradients[J].Journal of Electronics & Information Technology,2022,43(4):956-964.
Authors:XIE Tao  ZHANG Chunjiong  XU Yongjian
Abstract:The Stochastic Gradient Descent (SGD) algorithm randomly picks up a sample to estimate gradients, creating big variance which reduces the convergence speed and makes the training unstable. A Distributed SGD based on Average variance reduction, called DisSAGD is proposed. The method uses the average variance reduction based on historical gradients to update parameters in the machine learning model, requiring little gradient calculation and additional storage, but using the asynchronous communication protocol to share parameters across nodes. In order to solve the “update staleness” problem of global parameter distribution, a learning rate with an acceleration factor and an adaptive sampling strategy are included: on the one hand, when the parameter deviates from the optimal value, the acceleration factor is increased to speed up the convergence; on the other hand, when one work node is faster than the other ones, more samples are sampled for the next iteration, so that the node has more time to calculate the local gradient. Experiments show that the DisSAGD reduces significantly the waiting time of loop iterations, accelerates the convergence of the algorithm being faster than that of the controlled methods, and obtains almost linear acceleration in distributed cluster environments.
Keywords:Gradient descent  Machine learning  Distributed cluster  Adaptive sampling  Variance reduction
本文献已被 万方数据 等数据库收录!
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号