Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

PAD token missing ? #150

Closed
omkar-12bits opened this issue May 6, 2024 · 2 comments
Closed

PAD token missing ? #150

omkar-12bits opened this issue May 6, 2024 · 2 comments

Comments

@omkar-12bits
Copy link

omkar-12bits commented May 6, 2024

i tried using eos_token , unk_token and bos_token with left and right padding side but whenever the padding tokens size increases the outputs are pure garbage.

@omkar-12bits
Copy link
Author

after trying this on both mistral and mixtral its sure that paddings doesn't work very well with these models.
i was just playing with prompts at inference time and saw that paddings makes generation worst.
if inference on batches doesn't work well then how does it performs while training ? shouldn't it also produce garbage ?

@geronimi73
Copy link

can you post the code you use please?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants