当前位置: 澳门十大正规网投平台 > 正文
澳门十大正规网投平台

学术报告

【学术报告】A Mini-Batch Proximal Stochastic Recursive Gradient Algorithm with Diagonal Barzilai–Borwein Stepsize
发布日期:2023-09-07  浏览量:

学术报告

A Mini-Batch Proximal Stochastic Recursive Gradient Algorithm with Diagonal Barzilai–Borwein Stepsize

于腾腾(北京农学院

报告时间:202399日星期六  1515-1600


报告地点:沙河主楼E405


报告摘要: Many machine learning problems can be formulated as minimizing the sum of a function and a non-smooth regularization term. Proximal stochastic gradient methods are popular for solving such composite optimization problems. We propose a mini-batch proximal stochastic recursive gradient algorithm SRG-DBB, which incorporates the diagonal Barzilai–Borwein (DBB) stepsize strategy to capture the local geometry of the problem. The linear convergence and complexity of SRG-DBB are analyzed for strongly convex functions. We further establish the linear convergence of SRG-DBB under the non-strong convexity condition. Moreover, it is proved that SRG-DBB converges sublinearly in the convex case. Numerical experiments on standard data sets indicate that the performance of SRG-DBB is better than or comparable to the proximal stochastic recursive gradient algorithm with best-tuned scalar stepsizes or BB stepsizes. Furthermore, SRG-DBB is superior to some advanced mini-batch proximal stochastic gradient methods.

 

报告人简介:于腾腾,2016年硕士毕业于西安电子科技大学,师从刘三阳教授。2021年博士毕业于河北工业大学,师从刘新为教授。2021年10月至2023年8月在中国科学院数学与系统科学研究院从事博士后研究,合作导师袁亚湘院士。主要研究兴趣为大规模机器学习中的随机梯度算法,相关成果发表在IEEE Transactions on Neural Networks and Learning Systems、Journal of Scientific Computing、Journal of the Operations Research Society of China等期刊。


邀请人:崔春风

版权所有 Copyright © 澳门十大正规网投平台-十大网投正规信誉网址【排行榜】 地址:北京市昌平区沙河高教园南三街9号    邮编:102206    电话:61716719