mirror of
https://github.com/hpcaitech/ColossalAI.git
synced 2025-09-10 05:20:33 +00:00
[shardformer] update colo attention to support custom mask (#5510)
* [feature] refactor colo attention (#5462) * [extension] update api * [feature] add colo attention * [feature] update sdpa * [feature] update npu attention * [feature] update flash-attn * [test] add flash attn test * [test] update flash attn test * [shardformer] update modeling to fit colo attention (#5465) * [misc] refactor folder structure * [shardformer] update llama flash-attn * [shardformer] fix llama policy * [devops] update tensornvme install * [test] update llama test * [shardformer] update colo attn kernel dispatch * [shardformer] update blip2 * [shardformer] update chatglm * [shardformer] update gpt2 * [shardformer] update gptj * [shardformer] update opt * [shardformer] update vit * [shardformer] update colo attention mask prep * [shardformer] update whisper * [test] fix shardformer tests (#5514) * [test] fix shardformer tests * [test] fix shardformer tests
This commit is contained in:
@@ -22,7 +22,7 @@ class _CudaExtension(_CppExtension):
|
||||
This function should return a list of nvcc compilation flags for extensions.
|
||||
"""
|
||||
|
||||
def is_hardware_available(self) -> bool:
|
||||
def is_available(self) -> bool:
|
||||
# cuda extension can only be built if cuda is available
|
||||
try:
|
||||
import torch
|
||||
@@ -32,7 +32,7 @@ class _CudaExtension(_CppExtension):
|
||||
cuda_available = False
|
||||
return cuda_available
|
||||
|
||||
def assert_hardware_compatible(self) -> None:
|
||||
def assert_compatible(self) -> None:
|
||||
from torch.utils.cpp_extension import CUDA_HOME
|
||||
|
||||
if not CUDA_HOME:
|
||||
|
Reference in New Issue
Block a user