mirror of https://github.com/hpcaitech/ColossalAI
Merge pull request #6064 from wangbluo/fix_attn
[sp] : fix the attention kernel for sppull/6062/head
commit
63314ce4e4
|
@ -121,7 +121,8 @@ class ColoAttention:
|
|||
)
|
||||
|
||||
if size >= MEMORY_BOUND:
|
||||
ColoAttention._flash_kernel_dispatch = ColoAttention._flash_kernel_dispatch.load()
|
||||
if isinstance(ColoAttention._flash_kernel_dispatch, KernelLoader):
|
||||
ColoAttention._flash_kernel_dispatch = ColoAttention._flash_kernel_dispatch.load()
|
||||
# lazy load
|
||||
if isinstance(ColoAttention._kernel_dispatch_map[dtype][mask_type], KernelLoader):
|
||||
ColoAttention._kernel_dispatch_map[dtype][mask_type] = ColoAttention._kernel_dispatch_map[dtype][
|
||||
|
|
Loading…
Reference in New Issue