Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

chatglm-turbo support 3.2k token #1757

Closed
2 tasks done
xutengfei opened this issue Dec 14, 2023 · 8 comments
Closed
2 tasks done

chatglm-turbo support 3.2k token #1757

xutengfei opened this issue Dec 14, 2023 · 8 comments
Assignees
Labels
🐞 bug Something isn't working

Comments

@xutengfei
Copy link

xutengfei commented Dec 14, 2023

Self Checks

Dify version

0.3.32

Cloud or Self Hosted

Self Hosted (Docker)

Steps to reproduce

image

✔️ Expected Behavior

image

❌ Actual Behavior

image we can not change max token of chatglm-turbo
@xutengfei xutengfei added the 🐞 bug Something isn't working label Dec 14, 2023
@crazywoola
Copy link
Member

crazywoola commented Dec 14, 2023

We will take a look into this issue. :)

---- Update----
The model did not provide a setting called max_token, so it shouldn't be able to display anyway. We will remove this setting later.

@xutengfei
Copy link
Author

xutengfei commented Dec 14, 2023

We will take a look into this issue. :)

---- Update---- The model did not provide a setting called max_token, so it shouldn't be able to display anyway. We will remove this setting later.

so we have to wait the version of 0.3.4?
now the request log show that dify did not submit the content to llm model when context larger than 512token.

@crazywoola
Copy link
Member

Yes, until we fix this.

@crazywoola
Copy link
Member

I have a question about the version, we are currently using v0.3.33, And you said you are using 0.3.2? Do you mean 0.3.32?

@xutengfei
Copy link
Author

I have a question about the version, we are currently using v0.3.33, And you said you are using 0.3.2? Do you mean 0.3.32?

yes ,i mean 0.3.32. Thanks

@crazywoola
Copy link
Member

This is a frontend bug, it won't affect what it will send to the llm. Because we do not have max token for chatglm_turbo. So it's a placeholder(default) value.
First image shows the 512 tokens, but actually it costs more than this one.
image

image

@crazywoola crazywoola assigned crazywoola and unassigned takatost Dec 14, 2023
@WSDzju
Copy link

WSDzju commented Dec 20, 2023

This is a frontend bug, it won't affect what it will send to the llm. Because we do not have max token for chatglm_turbo. So it's a placeholder(default) value. First image shows the 512 tokens, but actually it costs more than this one. image

image

I also found the problem. In the case of gpt-3.5-turbo, the Max Token can be costomized. So the Max Token option will be added for Chatglm? I also try the lastest version 0.3.34, the option have not provided yet. BTW, the ouput length of chatglm-turbo is relatively short, although I have provided considerable context docs and prompted to give the answer as long as possible. So the situation is only related to LLM itself ?

@crazywoola
Copy link
Member

We have update the UI, and there is no max_token in ChatGLM. This should be resolved in v0.4.x

image

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
🐞 bug Something isn't working
Projects
None yet
Development

No branches or pull requests

4 participants