Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

LLaVA multimodal projection stops working after New Session #821

Closed
iScriptLex opened this issue May 4, 2024 · 4 comments
Closed

LLaVA multimodal projection stops working after New Session #821

iScriptLex opened this issue May 4, 2024 · 4 comments

Comments

@iScriptLex
Copy link

Kubuntu 22.04, GeForce 4090.
KoboldCPP version: 1.64
Models: kunoichi-7b.Q8_0.gguf, BuRP_7B-Q8_0-imat.gguf or any other Mistral-based 7B model.
Mmproj: mistral-7b-mmproj-v1.5-Q4_1.gguf
Command line:
./koboldcpp-linux-x64 --model models/BuRP_7B-Q8_0-imat.gguf --usecublas --gpulayers 10000 --contextsize 4096 --preloadstory startup.json --mmproj mmproj/mistral-7b-mmproj-v1.5-Q4_1.gguf

  1. Add image for recognition (Add Img -> Upload Image File)
  2. Add text "Describe this image" to chat
  3. Click "Generate More".

Model describes image successfully: objects, composition, etc.

  1. After that, click "New Session" (this removes uploaded image and generated text)
  2. Add some other (NOT the same) image for recognition
  3. Add text "Describe this image" to chat
  4. Click "Generate More".

Model talk nonsense that has nothing to do with the uploaded image. All subsequent uploaded images are also not recognized. Image recognition starts working again only after the KoboldCPP is fully restarted.

KoboldCPP 1.61.2 works well and doesn't have this bug.

@LostRuins
Copy link
Owner

Thanks for reporting, it's a known issue, i'm trying to find the commit that caused it.

Linked: ggerganov#7060

@LostRuins
Copy link
Owner

Hi, please try the hotfix 1.64.1 and let me know if that works.

@iScriptLex
Copy link
Author

Thank you very much, 1.64.1 works well.

@LostRuins
Copy link
Owner

Great. We can use my solution until upstream fixes it properly.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants