From 1f60061c971fbe7f1b145f99130ae17dd3bba05e Mon Sep 17 00:00:00 2001 From: ydshieh Date: Wed, 7 Feb 2024 11:11:52 +0100 Subject: [PATCH] docker --- tests/models/llama/test_inference_llama.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/tests/models/llama/test_inference_llama.py b/tests/models/llama/test_inference_llama.py index 88346844..cbb33f5f 100644 --- a/tests/models/llama/test_inference_llama.py +++ b/tests/models/llama/test_inference_llama.py @@ -25,6 +25,6 @@ def test_foo(self): generate_ids = model.generate(inputs.input_ids, max_length=30) output = tokenizer.batch_decode(generate_ids, skip_special_tokens=True, clean_up_tokenization_spaces=False)[0] - expected_output = "Hey, are you conscious? Can you talk to me?\nI'm not conscious, but I can talk to you." + expected_output = "I'm here, but I'm not sure I'm conscious." assert output == expected_output