Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Generate a bin file in Linux. #8

Open
Rui-Hu01 opened this issue Jul 26, 2023 · 2 comments
Open

Generate a bin file in Linux. #8

Rui-Hu01 opened this issue Jul 26, 2023 · 2 comments

Comments

@Rui-Hu01
Copy link

In the "Convert the model to ggml" section of the README.md, what is the "" parameter referring to? Following the instructions in MiniGPT4, I ended up with a folder containing the model weights. What specific file should I point to with this parameter?

@Rui-Hu01
Copy link
Author

In the "Convert the model to ggml" section of the README.md, what is the "" parameter referring to? Following the instructions in MiniGPT4, I ended up with a folder containing the model weights. What specific file should I point to with this parameter?

I made a mistake. convert.py should be under llama.cpp instead of minigpt4. However, there is now a new error : 'Exception: Vocab size mismatch (model has 32001, but MiniGPT-4/model/weight_7b/tokenizer.model has 32000). Most likely you are missing added_tokens.json (should be in MiniGPT-4/model/weight_7b).'. It worked fine in Minigpt-4. How should I handle this problem? What is missing?

@Maknee
Copy link
Owner

Maknee commented Jul 28, 2023

Add a added_tokens.json with id of 32001 (I think) in the folder that contains the vocab and model.

This is a problem with how pytorch loads the file and how minigpt4 loads it I believe. I didn't explore it in depth, but after adding the json file, that should fix it and convert.py should convert to the correct weights.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants