-
Notifications
You must be signed in to change notification settings - Fork 293
Open
Labels
scoping requiredFeatures that need significant design and planning before being actionableFeatures that need significant design and planning before being actionabletype:featureNew feature or requestNew feature or request
Description
Is your feature request related to a problem? Please describe.
Flash Attention 2 is a library that provides attention operation kernels for faster and more memory efficient inference and training:
Describe the solution you'd like
(https://github.com/Dao-AILab/flash-attention)
innat, shanky-kapoor and BogdanTurbal
Metadata
Metadata
Assignees
Labels
scoping requiredFeatures that need significant design and planning before being actionableFeatures that need significant design and planning before being actionabletype:featureNew feature or requestNew feature or request