New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
GGUF breaks - llama-3 #430
Comments
Update:
I used See reproducible notebook: https://colab.research.google.com/drive/1aNS8CgXoJZHclBEW3ZjFfiLjpmqZ14KN?usp=sharing Below is the comparison of tokenization differences between llama.cpp and HF: I also used Sadly |
Thanks for having looked into this. I've been suspicious of these |
It should be fixed! |
Findings from ggerganov/llama.cpp#7062 and Discord chats:
Notebook for repro: https://colab.research.google.com/drive/1djwQGbEJtUEZo_OuqzN_JF6xSOUKhm4q?usp=sharing
Todo:
The text was updated successfully, but these errors were encountered: