-
Notifications
You must be signed in to change notification settings - Fork 38
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Eagerly Awaiting the Megatron Code Release #1
Comments
Thanks for the interest! We will publish it ASAP! |
hi, @mactavish91 Please see the example proof of concept Megatron-LM PR in here: |
Are there any plab to open source the training code for MoE mentioned in the paper? |
@huyiwen MoE training code is not related with the Muon optimizer itself, so we do not plan to release it. |
are the released intermediate checkpoints for the MoE model or the dense model? If they are for the MoE model, how should we load and train them? |
They are MoE models:
|
I really appreciate the amazing work you’ve done on the Moonshot AI open source MoE model. I’m excited about the upcoming release of the Megatron code and can't wait to explore it.
Thank you for your efforts!
The text was updated successfully, but these errors were encountered: