feat: change minimal supported CUDA version to 11.7 (#1206)
This commit is contained in:
parent
3cea2fb6ee
commit
e371bea04f
@ -98,7 +98,7 @@ MAX_JOBS=4 pip install flash-attn --no-build-isolation
|
|||||||
|
|
||||||
### NVIDIA CUDA Support
|
### NVIDIA CUDA Support
|
||||||
**Requirements:**
|
**Requirements:**
|
||||||
- CUDA 11.6 and above.
|
- CUDA 11.7 and above.
|
||||||
|
|
||||||
We recommend the
|
We recommend the
|
||||||
[Pytorch](https://catalog.ngc.nvidia.com/orgs/nvidia/containers/pytorch)
|
[Pytorch](https://catalog.ngc.nvidia.com/orgs/nvidia/containers/pytorch)
|
||||||
|
|||||||
4
setup.py
4
setup.py
@ -156,9 +156,9 @@ if not SKIP_CUDA_BUILD and not IS_ROCM:
|
|||||||
cc_flag = []
|
cc_flag = []
|
||||||
if CUDA_HOME is not None:
|
if CUDA_HOME is not None:
|
||||||
_, bare_metal_version = get_cuda_bare_metal_version(CUDA_HOME)
|
_, bare_metal_version = get_cuda_bare_metal_version(CUDA_HOME)
|
||||||
if bare_metal_version < Version("11.6"):
|
if bare_metal_version < Version("11.7"):
|
||||||
raise RuntimeError(
|
raise RuntimeError(
|
||||||
"FlashAttention is only supported on CUDA 11.6 and above. "
|
"FlashAttention is only supported on CUDA 11.7 and above. "
|
||||||
"Note: make sure nvcc has a supported version by running nvcc -V."
|
"Note: make sure nvcc has a supported version by running nvcc -V."
|
||||||
)
|
)
|
||||||
# cc_flag.append("-gencode")
|
# cc_flag.append("-gencode")
|
||||||
|
|||||||
Loading…
Reference in New Issue
Block a user