-
Notifications
You must be signed in to change notification settings - Fork 4.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Update sharded_moe.py to support top2 gate with Tutel #6948
base: master
Are you sure you want to change the base?
Conversation
@microsoft-github-policy-service agree company="University of Michigan" |
Not sure if this check is needed
I see that in I didn't see any specify check for non-zero mask. |
Hi @xenshinu - FYI you'll need to update from the CLA to DCO requirements in this PR if you're able to? |
Signed-off-by: Xueshen Liu <[email protected]>
@xenshinu - looks like the formatting check is failing, could you run the pre-commit formatter? |
Tutel is forced to be unused on k > 1 since #2053
Given the fact that multiple experts per token is very common, and the gather and scatter operation without Tutel is so inefficient, I added support of tutel to top2 gate and tested on pipeline engine. This can be done for any k actually, I'll push that later when I have time to test,