Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How can i update to vLLM v0.4.1 for llama3 support ? #66

Open
Lhemamou opened this issue Apr 30, 2024 · 5 comments
Open

How can i update to vLLM v0.4.1 for llama3 support ? #66

Lhemamou opened this issue Apr 30, 2024 · 5 comments

Comments

@Lhemamou
Copy link

Hello everyone,

I would like to update the vLLM version to v0.4.1 in order to get access to LLAMA3 but i don't know how modify the fork runpod/vllm-fork-for-sls-worker. Could you please guide me ? Happy to help in some way!

@nuckcrews
Copy link

+1

@nerdylive123
Copy link

+1, looking to figure this out soon

@houmie
Copy link

houmie commented May 7, 2024

Same issue here. There is a blocking bug on Llama3 that has been fixed with v0.4.1.

@arthrod
Copy link

arthrod commented May 7, 2024

Pretty please

@alpayariyak
Copy link
Collaborator

Hi all, thank you for raising this issue! I have just merged the vLLM 0.4.2 update into main, you can use it by changing your Docker image in your endpoint from runpod/worker-vllm:stable-cudaX.X.X to runpod/worker-vllm:dev-cudaX.X.X. From my testing thus far, everything seems in order, but if you notice any issues, please let me know. After an initial test period, I'll release the update officially to replace the default stable images. Thanks all!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

6 participants