from .flash_attention_dao_cuda import FlashAttentionDaoCudaExtension from .flash_attention_npu import FlashAttentionNpuExtension from .flash_attention_xformers_cuda import FlashAttentionXformersCudaExtension try: import flash_attention # noqa HAS_FLASH_ATTN = True except: HAS_FLASH_ATTN = False try: import xformers # noqa HAS_MEM_EFF_ATTN = True except: HAS_MEM_EFF_ATTN = False __all__ = ["FlashAttentionDaoCudaExtension", "FlashAttentionXformersCudaExtension", "FlashAttentionNpuExtension"]