ColossalAI/applications/Colossal-LLaMA-2/colossal_llama2
Zian(Andy) Zheng 7768afbad0 Update flash_attention_patch.py
To be compatible with the new change in the Transformers library, where a new argument 'padding_mask' was added to forward function of attention layer.
https://github.com/huggingface/transformers/pull/25598
2023-10-16 14:00:45 +08:00
..
dataset initial commit: add colossal llama 2 (#4784) 2023-09-24 23:12:26 +08:00
model initial commit: add colossal llama 2 (#4784) 2023-09-24 23:12:26 +08:00
tokenizer initial commit: add colossal llama 2 (#4784) 2023-09-24 23:12:26 +08:00
utils Update flash_attention_patch.py 2023-10-16 14:00:45 +08:00
__init__.py initial commit: add colossal llama 2 (#4784) 2023-09-24 23:12:26 +08:00