Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Fine Tuning on Custom Data ipynb #87

Closed
samarthsarin opened this issue Mar 30, 2023 · 17 comments
Closed

Fine Tuning on Custom Data ipynb #87

samarthsarin opened this issue Mar 30, 2023 · 17 comments
Labels
enhancement New feature or request training gpt4all-training issues

Comments

@samarthsarin
Copy link

Can you please provide a ipynb notebook which shows steps for fine tuning this model on custom data?

@FiveTechSoft

This comment was marked as spam.

@daleevans
Copy link

The entire process is just:

make your data, in the jsonl format like you get when you download the standard data
edit configs/train/finetune_lora.yaml to point to your new data file and set up your wandb/hf account info
possibly edit configs/deepspeed/ds_config.json depending on your local GPU/CPU/memory (batch sizes, and maybe set stage3_gather_16bit_weights_on_model_save and cpu offload)
run train.py

If you don't have wandb or hf, you may need to comment some lines out in train.py

@samarthsarin
Copy link
Author

I got your point. Still for better documentation I would be really grateful if some jupyter nptebook can be provided as majority of the audience here is looking for one such fine tuning code. We all would be really grateful if you can provide one such code for fine tuning gpt4all in a jupyter notebook.

Thank you

@magedhelmy1
Copy link

Hi @zanussbaum, any advise on how to move forward with this?

@ajayarunachalam

This comment was marked as spam.

9 similar comments
@surajtripathy07

This comment was marked as spam.

@MiniWho

This comment was marked as spam.

@NikhilShaw

This comment was marked as spam.

@naman-TI

This comment was marked as spam.

@KiranNadig62

This comment was marked as spam.

@ybdesire

This comment was marked as spam.

@klintT

This comment was marked as spam.

@joseberlines

This comment was marked as spam.

@irfanmaulanaak

This comment was marked as spam.

@Dineshk011287
Copy link

I got your point. Still for better documentation I would be really grateful if some jupyter nptebook can be provided as majority of the audience here is looking for one such fine tuning code. We all would be really grateful if you can provide one such code for fine tuning gpt4all in a jupyter notebook.

+1
I am also looking for this. If any documentation or jupyter notebook would definitely help

@niansa niansa added enhancement New feature or request training gpt4all-training issues labels Aug 10, 2023
@nomic-ai nomic-ai deleted a comment from WordDealer Oct 13, 2023
@nomic-ai nomic-ai deleted a comment from fredcobain Oct 13, 2023
@cebtenzzre cebtenzzre pinned this issue Oct 13, 2023
@windowshopr

This comment was marked as spam.

@cebtenzzre cebtenzzre unpinned this issue May 9, 2024
@cebtenzzre
Copy link
Member

Closing this issue as stale. A lot has changed since Nomic last trained a text completion model.

@cebtenzzre cebtenzzre closed this as not planned Won't fix, can't repro, duplicate, stale May 9, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request training gpt4all-training issues
Projects
None yet
Development

No branches or pull requests