Skip to content

[Feature]Moe alltoallv communication optimization for unquantized RL training sence & alltoallv support dpo #1389

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
wants to merge 10 commits into
base: v0.9.1-dev
Choose a base branch
from

Conversation

harygo22
Copy link

[Feature]Moe alltoallv communication optimization for unquantized RL training sence

22dimensions and others added 10 commits June 10, 2025 10:07
remove old quantization model, and new models will be added to testcase
later.

Signed-off-by: 22dimensions <[email protected]>
### What this PR does / why we need it?
Update 0.9.0rc1 contributors info

### Does this PR introduce _any_ user-facing change?
No

### How was this patch tested?
CI passed

Signed-off-by: Yikun Jiang <[email protected]>
### What this PR does / why we need it?
Make accuarcy CI and report work

### Does this PR introduce _any_ user-facing change?
No

### How was this patch tested?
Manaully review

Signed-off-by: hfadzxy <[email protected]>
Copy link

This pull request has conflicts, please resolve those before we can evaluate the pull request.

Copy link

This pull request has conflicts, please resolve those before we can evaluate the pull request.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants