FlashAttention-2 support
#7
by
afaulconbridge
- opened
Please add FlashAttention-2 support https://huggingface.co./docs/transformers/perf_infer_gpu_one#flashattention-2
Feel free to open an issue on Github
Please add FlashAttention-2 support https://huggingface.co./docs/transformers/perf_infer_gpu_one#flashattention-2
Feel free to open an issue on Github