You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
i tried using eos_token , unk_token and bos_token with left and right padding side but whenever the padding tokens size increases the outputs are pure garbage.
The text was updated successfully, but these errors were encountered:
after trying this on both mistral and mixtral its sure that paddings doesn't work very well with these models.
i was just playing with prompts at inference time and saw that paddings makes generation worst.
if inference on batches doesn't work well then how does it performs while training ? shouldn't it also produce garbage ?
i tried using
eos_token
,unk_token
andbos_token
withleft
andright
padding side but whenever the padding tokens size increases the outputs are pure garbage.The text was updated successfully, but these errors were encountered: