mirror of
https://github.com/hpcaitech/ColossalAI.git
synced 2025-06-20 04:32:47 +00:00
Merge pull request #6064 from wangbluo/fix_attn
[sp] : fix the attention kernel for sp
This commit is contained in:
commit
63314ce4e4
@ -121,6 +121,7 @@ class ColoAttention:
|
|||||||
)
|
)
|
||||||
|
|
||||||
if size >= MEMORY_BOUND:
|
if size >= MEMORY_BOUND:
|
||||||
|
if isinstance(ColoAttention._flash_kernel_dispatch, KernelLoader):
|
||||||
ColoAttention._flash_kernel_dispatch = ColoAttention._flash_kernel_dispatch.load()
|
ColoAttention._flash_kernel_dispatch = ColoAttention._flash_kernel_dispatch.load()
|
||||||
# lazy load
|
# lazy load
|
||||||
if isinstance(ColoAttention._kernel_dispatch_map[dtype][mask_type], KernelLoader):
|
if isinstance(ColoAttention._kernel_dispatch_map[dtype][mask_type], KernelLoader):
|
||||||
|
Loading…
Reference in New Issue
Block a user