Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

What's the trainset is used to obtain “Model with contextg extension via improved LoRA fine-tuning” (LoRA+)? #184

Open
ZackZikaiXiao opened this issue Apr 22, 2024 · 0 comments

Comments

@ZackZikaiXiao
Copy link

ZackZikaiXiao commented Apr 22, 2024

Hi, thanks for the great work. I have a question regarding the used trainset for different types of models (Fully fine-tuned, Lora+, models for extra experiments in paper).

In the ReadMe, it states, "There is no need to make supervised fine-tuning upon the fine-tuned context extended models. It is all right to directly use the base model as Llama2-chat models, as the amount of long instruction following data is enough for SFT." While in the paper, Figure 5's caption suggests that Lora+ is trained with RedPajama.

I'm seeking clarification on the following points:

  1. Do the released models refer to those that have undergone unsupervised fine-tuning on RedPajama and then tested on PG19?
  2. Is Table 9, which evaluates the LongBench benchmark, the only one involving supervised fine-tuning with LongAlpaca-12k based on models fine-tuned with RedPajama?
  3. Where can I find the performance of using only LongAlpaca-12k to derive the Lora adapter, embeds, and norm layer?
RedPajama (unsupervised) LongAlpaca-12k (supervised)
Fully fine-tuned (readme)
Lora+ (readme)
Models for LongBench benchmark (paper)

I've drafted a table to summarize my understanding of the training configurations mentioned in both the ReadMe and the paper. Could you please confirm if this representation is correct?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant