Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

About OLLAMA_PARALLEL split the max context length #4079

Open
DirtyKnightForVi opened this issue May 1, 2024 · 0 comments
Open

About OLLAMA_PARALLEL split the max context length #4079

DirtyKnightForVi opened this issue May 1, 2024 · 0 comments
Labels
bug Something isn't working

Comments

@DirtyKnightForVi
Copy link

What is the issue?

I encountered this while testing SQL QA with extremely large table, and i put all DDL into system .

When OLLAMA_PARALLEL = 4, I observed that model appears to only understand the last 4000 tokens of the DDL. This is quite different from my previous experience. My webui is open webui , it can set num_ctx to 16000, but useless.

BUT changing OLLAMA_PARALLEL=1, model can understand the whole DDL !

so , max_num_ctx = 16000 / OLLAMA_PARALLEL ? Even when the machine is free ?

OS

Linux

GPU

Nvidia

CPU

Intel

Ollama version

0.1.33-RC5

@DirtyKnightForVi DirtyKnightForVi added the bug Something isn't working label May 1, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

1 participant