Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Moe converters #5

Merged
merged 4 commits into from
Oct 19, 2024
Merged

Moe converters #5

merged 4 commits into from
Oct 19, 2024

Conversation

AleHD
Copy link
Collaborator

@AleHD AleHD commented Sep 3, 2024

No description provided.

AleHD and others added 4 commits August 28, 2024 13:55
@AleHD AleHD marked this pull request as ready for review September 5, 2024 12:06
@AleHD
Copy link
Collaborator Author

AleHD commented Sep 5, 2024

Ready for review. Instructions on how to use the new converters are in the examples/xglm/README. Implementation of huggingface transformers moe xglm in examples/xglm/transformers_impl (thanks @negar-foroutan!). Let me know what you think :)

@negar-foroutan negar-foroutan merged commit 3bce1f4 into moe Oct 19, 2024
2 of 4 checks passed
@negar-foroutan negar-foroutan deleted the moe_converter branch October 19, 2024 14:16
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants