-
Notifications
You must be signed in to change notification settings - Fork 126
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
How to use the finetuned mistal model for inference with Medusa #75
Comments
As an example, you can refer to the Zephyr model ( |
@ctlllll
But in my case , i am using the mistral model-https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2 (not based on medusa). so shall i use medusa library to improve my mistral models inference time? |
You will need to train the hugging face model on the medusa heads before you can use it for inference. |
@eldhosemjoy How to train the hugging face model on the Medusa heads?can you share the reference |
You can use this script - https://github.com/FasterDecoding/Medusa/blob/main/medusa/train/train_legacy.py |
Is there no way to inference without training? I didn't have the computing resources to train, so I wanted to infer without training. |
Try this model which is ported to medusa. |
How to use the finetuned mistal model for inference with Medusa
The text was updated successfully, but these errors were encountered: