Skip to content

Conversation

liuzhenhai93
Copy link
Contributor

PR types

Others

PR changes

Others

Description

  • sharding reshard supoort v1 -> v2, v2->v1, v2->v2, v1->v1

Meiyim and others added 30 commits June 16, 2023 08:23
[Distributed]Fix trainer for distributed training
…rategy

[Dygraph] Support PP new strategy - delay_scale_loss & dp_comm_overlap
support sharding stage1 in hybrid parallel.
…ng-loop

Add time statistics for nccl-connection.
allow to use `main-grad` under TF32/FP32
…umulate

[hot-fix] resume from accumulation-step wrong
[fix] pp非mp情况下只存了pp01 model
…sume, warning when model-weight has missing keys
online hot fix, rasie Error when optimizer/lr scheduler no show in re…
liuzhenhai93 and others added 28 commits July 14, 2023 19:30
* broadcast

* polish

* polish

* polish

* polish

* polish
* fix ClipGradByAdaptiveNorm load state dict

* fix code style
@CLAassistant
Copy link

CLAassistant commented Oct 14, 2023

CLA assistant check
Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you all sign our Contributor License Agreement before we can accept your contribution.
11 out of 13 committers have signed the CLA.

✅ Meiyim
✅ haohongxiang
✅ GhostScreaming
✅ ForFishes
✅ SylarTiaNII
✅ wuhuachaocoding
✅ GuoxiaWang
✅ liuzhenhai93
✅ sneaxiy
✅ xysheng-baidu
✅ ZHUI
❌ pangengzheng
❌ QimingPeng


pangengzheng seems not to be a GitHub user. You need a GitHub account to be able to sign the CLA. If you have already a GitHub account, please add the email address used for this commit to your account.
You have signed the CLA already but the status is still pending? Let us recheck it.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.