From fdec650bb471062308d9948a5b2740a65914b4ea Mon Sep 17 00:00:00 2001 From: Xu Kai Date: Wed, 11 Oct 2023 17:43:01 +0800 Subject: [PATCH] fix test llama (#4884) --- tests/test_infer/test_llama_infer.py | 1 - 1 file changed, 1 deletion(-) diff --git a/tests/test_infer/test_llama_infer.py b/tests/test_infer/test_llama_infer.py index 13bdf0399..b424525a3 100644 --- a/tests/test_infer/test_llama_infer.py +++ b/tests/test_infer/test_llama_infer.py @@ -38,7 +38,6 @@ def run_llama_test(test_config): enable_tensor_parallelism=True if test_config["tp_size"] > 1 else False, inference_only=True ) infer_engine = TPInferEngine(model, shard_config, BATCH_SIZE, MAX_INPUT_LEN, MAX_OUTPUT_LEN) - init_to_get_rotary(model.model, base=10000) generate_kwargs = dict(max_new_tokens=MAX_OUTPUT_LEN, do_sample=False) input_tokens = {