You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
ColossalAI/colossalai/kernel/triton
Jianghai 7ddd8b37f0
fix (#5311)
10 months ago
..
__init__.py [inference]Optimize the usage of the mid tensors space in flash attn (#5304) 10 months ago
context_attn_unpad.py [inference]Optimize the usage of the mid tensors space in flash attn (#5304) 10 months ago
custom_autotune.py
flash_decoding.py [inference]Optimize the usage of the mid tensors space in flash attn (#5304) 10 months ago
fused_rotary_embedding.py fix (#5311) 10 months ago
gptq_triton.py
kvcache_copy.py
llama_act_combine_kernel.py
no_pad_rotary_embedding.py fix (#5311) 10 months ago
qkv_matmul_kernel.py
rms_layernorm.py
rotary_cache_copy.py fix (#5311) 10 months ago
softmax.py