Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Thanks and questions... #1

Open
beejaya opened this issue Jun 30, 2023 · 1 comment
Open

Thanks and questions... #1

beejaya opened this issue Jun 30, 2023 · 1 comment

Comments

@beejaya
Copy link

beejaya commented Jun 30, 2023

@mayooear thanks for this nice project. Very helpful. Have a couple of items:

  1. The current setup is missing a required package: sentence_transformers
  2. Any thoughts on how to run this in GPU?
@saradhix
Copy link

saradhix commented Jun 30, 2023

The required package sentence-transformers is available. There is a hyphen instead of underscore in the package name.
Currently the model is the quantized version of the mpt-30b-chat model which is being loaded with Ctransformers. You can use the original huggingface llm with this code

import transformers
llm = transformers.AutoModelForCausalLM.from_pretrained(
  'mosaicml/mpt-30b-chat',
  trust_remote_code=True
)

You should replace lines 73 to 78 in question_answer_docs.py with the above code. It should work fine ( I haven't tried) with GPU if you have the GPU which can hold the 30b parameter model.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants