From 8c2ab31579c124af0ae7cab92e3782f74578fc2e Mon Sep 17 00:00:00 2001 From: Andrew Zhu Date: Wed, 6 Sep 2023 12:08:11 -0400 Subject: [PATCH] build: llama output format --- tests/test_llama.py | 10 +++++----- 1 file changed, 5 insertions(+), 5 deletions(-) diff --git a/tests/test_llama.py b/tests/test_llama.py index be3451c..f44c612 100644 --- a/tests/test_llama.py +++ b/tests/test_llama.py @@ -46,9 +46,9 @@ async def test_llama(create_kani, gh_log): resp = await ai.chat_round_str("What are some cool things to do in Tokyo?") gh_log.write( "# LLaMA Basic\n" - "*This is a real output from kani running LLaMA v2 on GitHub Actions.*\n" - "---\n" - ">>> What are some cool things to do in Tokyo?\n" + "*This is a real output from kani running LLaMA v2 on GitHub Actions.*\n\n" + "---\n\n" + "> What are some cool things to do in Tokyo?\n\n" ) gh_log.write(resp) gh_log.write("\n\n---\n\n") @@ -65,8 +65,8 @@ async def test_chatting_llamas(create_kani, gh_log): tourist_response = tourist.chat_history[-1].content gh_log.write( "# LLaMAs Visit Tokyo\n" - "*These are real outputs from two kani running LLaMA v2 on GitHub Actions.*\n" - "---\n" + "*These are real outputs from two kani running LLaMA v2 on GitHub Actions.*\n\n" + "---\n\n" f"### Tourist\n{tourist_response}\n" ) for _ in range(5):