Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

【PaddlePaddle Hackathon】19、在 Paddle 中新增 AdaScaleSGD #35991

Closed
TCChenlong opened this issue Sep 23, 2021 · 1 comment
Closed

【PaddlePaddle Hackathon】19、在 Paddle 中新增 AdaScaleSGD #35991

TCChenlong opened this issue Sep 23, 2021 · 1 comment

Comments

@TCChenlong
Copy link
Contributor

(此 ISSUE 为 PaddlePaddle Hackathon 活动的任务 ISSUE,更多详见PaddlePaddle Hackathon

【任务说明】

  • 任务标题:在 Paddle 中新增 AdaScale SGD

  • 技术标签:深度学习框架,python,优化算法

  • 任务难度:中等

  • 详细描述: AdaScale SGD 一种大规模分布式训练中的LR schedule 策略, 被用来加速大batch 训练时的收敛速度,目前Paddle 中并没有实现。 本任务的目标是基于 Paddle.optimize.SGD 和 Paddle.distributed 等 Paddle 基础框架实现 Paddle 中的 AdaScale SGD 优化策略,收敛最大bathc szie、收敛精度两个指标上和论文对齐。

【提交内容】

  • 任务 PR 到 Paddle

  • 相关技术文档

  • 任务单测文件

  • 调用路径:paddle.optimizer.AdaScaleSGD

【技术要求】

  • 了解 Paddle 动静态图下 Optimize pass过程、Paddle 分布式框架

  • 熟练掌握 Python

@paddle-bot-old
Copy link

您好,我们已经收到了您的问题,会安排技术人员尽快解答您的问题,请耐心等待。请您再次检查是否提供了清晰的问题描述、复现代码、环境&版本、报错信息等。同时,您也可以通过查看官网API文档常见问题历史IssueAI社区来寻求解答。祝您生活愉快~

Hi! We've received your issue and please be patient to get responded. We will arrange technicians to answer your questions as soon as possible. Please make sure that you have posted enough message to demo your request. You may also check out the APIFAQGithub Issue and AI community to get the answer.Have a nice day!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

2 participants