mirror of https://github.com/hpcaitech/ColossalAI
21 lines
527 B
Python
21 lines
527 B
Python
|
from .flash_attention_dao_cuda import FlashAttentionDaoCudaExtension
|
||
|
from .flash_attention_npu import FlashAttentionNpuExtension
|
||
|
from .flash_attention_xformers_cuda import FlashAttentionXformersCudaExtension
|
||
|
|
||
|
try:
|
||
|
import flash_attention # noqa
|
||
|
|
||
|
HAS_FLASH_ATTN = True
|
||
|
except:
|
||
|
HAS_FLASH_ATTN = False
|
||
|
|
||
|
try:
|
||
|
import xformers # noqa
|
||
|
|
||
|
HAS_MEM_EFF_ATTN = True
|
||
|
except:
|
||
|
HAS_MEM_EFF_ATTN = False
|
||
|
|
||
|
|
||
|
__all__ = ["FlashAttentionDaoCudaExtension", "FlashAttentionXformersCudaExtension", "FlashAttentionNpuExtension"]
|