Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Pretrained MoE model #179

Open
Luodian opened this issue Aug 7, 2022 · 2 comments
Open

Pretrained MoE model #179

Luodian opened this issue Aug 7, 2022 · 2 comments
Labels
question Further information is requested

Comments

@Luodian
Copy link

Luodian commented Aug 7, 2022

hi thanks for providing such a wonderful work. However, I am curious that will you consider providing pretrained MoE models (e.g. ViT on ImageNet or machine translation tasks)

@ghostplant ghostplant added the question Further information is requested label Aug 8, 2022
@ghostplant
Copy link
Contributor

I think [SWIN Transformer] (https://github.com/microsoft/Swin-Transformer) would provide such pretrained MoE model based on Tutel.

For other language models over Fairseq, this repo currently only provides scripts that train models from scratch, since we don't really get the requirement from customers who would like to use certain pretrained models at what scale based on what datasets. But thanks for your suggestion, this is something meaningful and we'll discuss more about it.

@zeliu98
Copy link
Contributor

zeliu98 commented Aug 8, 2022

Hi @Luodian, the pretrained models can be found in: https://github.com/microsoft/Swin-Transformer/blob/main/MODELHUB.md#imagenet-22k-pretrained-swin-moe-models.
We have also provided an instruction on how to run Swin-MoE, which can be found in: https://github.com/microsoft/Swin-Transformer/blob/main/get_started.md#mixture-of-experts-support

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
question Further information is requested
Projects
None yet
Development

No branches or pull requests

3 participants