mirror of https://github.com/hpcaitech/ColossalAI
Browse Source
* [hotfix] remove triton kernels from kernel init * revise bloom/llama kernel imports for inferpull/4727/head
Yuanheng Zhao
1 year ago
committed by
GitHub
6 changed files with 28 additions and 30 deletions
@ -1,14 +1,7 @@
|
||||
from .cuda_native import FusedScaleMaskSoftmax, LayerNorm, MultiHeadAttention |
||||
from .triton import llama_context_attn_fwd, bloom_context_attn_fwd |
||||
from .triton import softmax |
||||
from .triton import copy_kv_cache_to_dest |
||||
|
||||
__all__ = [ |
||||
"LayerNorm", |
||||
"FusedScaleMaskSoftmax", |
||||
"MultiHeadAttention", |
||||
"llama_context_attn_fwd", |
||||
"bloom_context_attn_fwd", |
||||
"softmax", |
||||
"copy_kv_cache_to_dest", |
||||
] |
||||
|
Loading…
Reference in new issue