You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi! We've received your issue and please be patient to get responded. We will arrange technicians to answer your questions as soon as possible. Please make sure that you have posted enough message to demo your request. You may also check out the API,FAQ,Github Issue and AI community to get the answer.Have a nice day!
Since you haven't replied for more than a year, we have closed this issue/pr.
If the problem is not solved or there is a follow-up one, please reopen it at any time and we will continue to follow up.
由于您超过一年未回复,我们将关闭这个issue/pr。
若问题未解决或有后续问题,请随时重新打开,我们会继续跟进。
(此 ISSUE 为 PaddlePaddle Hackathon 第二期活动的任务 ISSUE,更多详见 【PaddlePaddle Hackathon 第二期】任务总览)
【任务说明】
任务标题:在 Paddle 中实现基于 DBTree 拓扑的 AllReduce
技术标签:深度学习框架,C++,通信拓扑
任务难度:困难
详细描述: DBTree 主要思想是利用二叉树中大约一半节点是叶子节点的特性,通过将叶子节点变换为非叶子节点,得到两颗二叉树,每个节点在其中一颗二叉树上是叶子节点,在另一颗二叉树上是非叶子节点。这种方法理论上能够提供比 Ring 算法更低的延迟。此任务的目标是在 Paddle 分布式训练框架中,实现 DBTree 结构的 AllReduce,除使用 GPU 进行训练外,最好也能支持其他异构硬件。
NCCL 参考:https://developer.nvidia.com/blog/massively-scale-deep-learning-training-nccl-2-4/
【提交内容】
【参考内容】
【技术要求】
【答疑交流】
The text was updated successfully, but these errors were encountered: