Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Feature] Should we remove other dependency of flashattention? #164

Open
1 task
sunpengsdu opened this issue Apr 1, 2024 · 0 comments
Open
1 task

[Feature] Should we remove other dependency of flashattention? #164

sunpengsdu opened this issue Apr 1, 2024 · 0 comments
Assignees
Labels
enhancement New feature or request

Comments

@sunpengsdu
Copy link
Contributor

Describe the feature

Should we remove other dependency of flash-attention, and only keep the core attention related ops?

If possible, we can only use pip to install flash-attention, avoiding a lot of compiling operations.

To seek whether it is possible, we need to check whether it would reduce the training performance a lot.

Will you implement it?

  • I would like to implement this feature and create a PR!
@sunpengsdu sunpengsdu added the enhancement New feature or request label Apr 1, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

6 participants