-
Notifications
You must be signed in to change notification settings - Fork 5.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
llava broke in new version v0.1.33 #4163
Comments
Sorry you hit this issue. Looking into it now |
Related issue in llama.cpp: ggerganov/llama.cpp#7060 |
The same issue on Ollama mac, llava7b & 13b failed on the second attempt |
I noticed same after update to v0.1.33, reverting to v0.1.32 fixed issue so it have to be some kind of regression. I have simple py script with query "What is on the picture?" done over collection of photos taken by phone. For each photo there is API request to generate endpoint: From response comparisons:
I do not know what are inner workings of ollama, but it seems that previous images or context is somehow preserved for next queries.
|
for me, the first response is empty 95% of the time, if you follow up the question it works. running v0.1.33 and the moondream model |
I guess Llava is now broken for everyone around the world who wants to try it as of yesterday. If there is documentation on how to install an older version then please kindly point me to it |
For me downloading previous installer (https://github.com/ollama/ollama/releases/tag/v0.1.32) and running it installed previous version. |
Same here. 0.1.32 worked, 0.1.33 doesn't. Using |
Hi there this should be fixed in this patch #4164 for now and we'll help hunt down why this broke more broadly in llama.cpp in the meantime. It will be fixed in the next release 0.1.34 which should be out very soon |
v0.1.34 fixed it - but I am using ollama directly via ollama-python lib |
Yea fixed for me so far using various python methods including ollama-python. works in open WebUI as well, BUT it is also true that the context can confuse it, so my advice is to make a new chat when using Open Webui, and beware of the growing chat context influencing the outputs. |
That means it's not really fixed because I didn't have this problem in the previous version:
This code was working in the previous version, now it only works for the first image, after that its no longer accurate. |
@jmorganca |
What is the issue?
Ollama v0.1.33
Intel Core i9 14900K 64GB ram
Nvidia RTX 4070
llava only works for the first inference attempt. All attempts afterwards make up strange descriptions not related to the image, almost like its looking at a different picture.
This also happens with llava:13b. It will work the first time after loading. After that, broken.
This also happens on other windows machines with different Intel and Nvidia combinations.
I have updated Ollama, and redownloaded the llava models.
OS
Windows
GPU
Nvidia
CPU
Intel
Ollama version
0.1.33
The text was updated successfully, but these errors were encountered: