From f51bf942031de97f3da1364c7c3d90d2100f8714 Mon Sep 17 00:00:00 2001 From: Roy Date: Wed, 30 Oct 2024 02:16:33 +0530 Subject: [PATCH 01/29] THUDM/chatglm3-6b_added_in_nightly_models --- .github/workflows/causal_lm_cpp.yml | 107 ++++++++++++++++++++++ tests/python_tests/ov_genai_test_utils.py | 1 + 2 files changed, 108 insertions(+) diff --git a/.github/workflows/causal_lm_cpp.yml b/.github/workflows/causal_lm_cpp.yml index 9856294340..a2086779f2 100644 --- a/.github/workflows/causal_lm_cpp.yml +++ b/.github/workflows/causal_lm_cpp.yml @@ -274,6 +274,41 @@ jobs: && call .\ov\setupvars.bat && python samples\python\greedy_causal_lm\lora.py .\TinyLlama\TinyLlama-1.1B-intermediate-step-1431k-3T\ adapter_model.safetensors "How to create a table with two columns, one of them has type float, another one has type int?" + cpp-greedy_causal_lm-Chatglm3-6b: + runs-on: ubuntu-24.04 + defaults: + run: + shell: bash + steps: + - uses: actions/checkout@v4 + with: + submodules: recursive + - uses: actions/setup-python@v4 + with: + python-version: 3.11 + - name: Install OpenVINO + run: | + mkdir ./ov/ + curl ${{ env.l_ov_link }} | tar --directory ./ov/ --strip-components 1 -xz + sudo ./ov/install_dependencies/install_openvino_dependencies.sh + - name: Build app + run: | + source ./ov/setupvars.sh + cmake -DCMAKE_BUILD_TYPE=Release -S ./ -B ./build/ + cmake --build ./build/ --config Release -j + - name: Download and convert and model + run: | + source ./ov/setupvars.sh + python -m pip install --upgrade-strategy eager -r ./samples/requirements.txt + python -m pip install ./thirdparty/openvino_tokenizers/[transformers] --pre --extra-index-url https://storage.openvinotoolkit.org/simple/wheels/nightly + optimum-cli export openvino --trust-remote-code --task text-generation-with-past --model THUDM/chatglm3-6b chatglm3-6b + - run: > + . ./ov/setupvars.sh + && timeout 2m ./build/samples/cpp/greedy_causal_lm/greedy_causal_lm ./chatglm3-6b/ 69 | diff <(timeout 2m samples/python/greedy_causal_lm/greedy_causal_lm.py ./chatglm3-6b/ 69) - + env: + PYTHONPATH: "./build" + + cpp-greedy_causal_lm-Qwen-7B-Chat: runs-on: ubuntu-20.04-16-cores defaults: @@ -535,6 +570,78 @@ jobs: " echo "Prompt lookup" passed + cpp-prompt_lookup_decoding_lm-ubuntu-Chatglm3-6b: + runs-on: ubuntu-20.04-16-cores + defaults: + run: + shell: bash + steps: + - uses: actions/checkout@v4 + with: + submodules: recursive + - uses: actions/setup-python@v4 + with: + python-version: 3.12 + - name: Install OpenVINO + run: | + mkdir ./ov/ + curl ${{ env.l_ov_link }} | tar --directory ./ov/ --strip-components 1 -xz + sudo ./ov/install_dependencies/install_openvino_dependencies.sh + - name: Build app + run: | + source ./ov/setupvars.sh + cmake -DCMAKE_BUILD_TYPE=Release -S ./ -B ./build/ + cmake --build ./build/ --config Release -j + - name: Download and convert and model + run: | + source ./ov/setupvars.sh + python -m pip install --upgrade-strategy eager -r ./samples/requirements.txt + python -m pip install ./thirdparty/openvino_tokenizers/[transformers] --pre --extra-index-url https://storage.openvinotoolkit.org/simple/wheels/nightly + optimum-cli export openvino --trust-remote-code --weight-format fp16 --model TinyLlama/TinyLlama-1.1B-Chat-v1.0 TinyLlama-1.1B-Chat-v1.0 + optimum-cli export openvino --trust-remote-code --task text-generation-with-past --model THUDM/chatglm3-6b chatglm3-6b + - name: run and compare + run: | + source ./ov/setupvars.sh + + echo 'Code:```python + def add(a, b): + return a + b + ``` + Question: Can you please add 2 and 3 + A:' > ./prompt.txt + + ./build/samples/cpp/prompt_lookup_decoding_lm/prompt_lookup_decoding_lm ./TinyLlama-1.1B-Chat-v1.0/ "$( predictions_prompt_lookup.txt + ./build/samples/cpp/greedy_causal_lm/greedy_causal_lm ./TinyLlama-1.1B-Chat-v1.0/ "$( predictions_greedy.txt + python -c " + with open('predictions_greedy.txt', 'r') as f: + predicted_greedy = f.readline() + with open('predictions_prompt_lookup.txt', 'r') as f: + predicted_prompt_lookup = f.readline() + assert predicted_greedy == predicted_prompt_lookup + " + echo "Prompt lookup" passed + - name: run and compare (model with seq_length_axis = 1) + run: | + source ./ov/setupvars.sh + + echo 'Code:```python + def add(a, b): + return a + b + ``` + Question: Can you please add 2 and 3 + A:' > ./prompt.txt + + ./build/samples/cpp/prompt_lookup_decoding_lm/prompt_lookup_decoding_lm ./Qwen-7B-Chat/ "$( predictions_prompt_lookup.txt + ./build/samples/cpp/greedy_causal_lm/greedy_causal_lm ./Qwen-7B-Chat/ "$( predictions_greedy.txt + python -c " + with open('predictions_greedy.txt', 'r') as f: + predicted_greedy = f.readline() + with open('predictions_prompt_lookup.txt', 'r') as f: + predicted_prompt_lookup = f.readline() + assert predicted_greedy == predicted_prompt_lookup + " + echo "Prompt lookup" passed + cpp-Phi-1_5: runs-on: ubuntu-20.04-16-cores defaults: diff --git a/tests/python_tests/ov_genai_test_utils.py b/tests/python_tests/ov_genai_test_utils.py index 5730def0c6..d9e28af400 100644 --- a/tests/python_tests/ov_genai_test_utils.py +++ b/tests/python_tests/ov_genai_test_utils.py @@ -25,6 +25,7 @@ def get_models_list(): "microsoft/phi-1_5", "microsoft/phi-2", "THUDM/chatglm2-6b", + "THUDM/chatglm3-6b", "Qwen/Qwen2-0.5B-Instruct", "Qwen/Qwen-7B-Chat", "Qwen/Qwen1.5-7B-Chat", From 064d8ad45730fdd0638ad4d3a08dcc5d29a8e821 Mon Sep 17 00:00:00 2001 From: Roy Date: Wed, 30 Oct 2024 18:06:27 +0530 Subject: [PATCH 02/29] checked --- .github/workflows/causal_lm_cpp.yml | 4 ++-- tests/python_tests/ov_genai_test_utils.py | 2 +- 2 files changed, 3 insertions(+), 3 deletions(-) diff --git a/.github/workflows/causal_lm_cpp.yml b/.github/workflows/causal_lm_cpp.yml index a2086779f2..ed8b5b5c02 100644 --- a/.github/workflows/causal_lm_cpp.yml +++ b/.github/workflows/causal_lm_cpp.yml @@ -1077,9 +1077,9 @@ jobs: Overall_Status: name: ci/gha_overall_status_causal_lm - needs: [cpp-multinomial-greedy_causal_lm-ubuntu, cpp-beam_search_causal_lm-ubuntu, cpp-greedy_causal_lm-windows, + needs: [cpp-multinomial-greedy_causal_lm-ubuntu, cpp-beam_search_causal_lm-ubuntu, cpp-greedy_causal_lm-windows,cpp-greedy_causal_lm-Chatglm3-6b, cpp-greedy_causal_lm-Qwen-7B-Chat, cpp-beam_search_causal_lm-Qwen1_5-7B-Chat, cpp-beam_search_causal_lm-Phi-2, - cpp-beam_search_causal_lm-notus-7b-v1, cpp-speculative_decoding_lm-ubuntu, cpp-prompt_lookup_decoding_lm-ubuntu, + cpp-beam_search_causal_lm-notus-7b-v1, cpp-speculative_decoding_lm-ubuntu, cpp-prompt_lookup_decoding_lm-ubuntu,cpp-prompt_lookup_decoding_lm-ubuntu-Chatglm3-6b, cpp-Phi-1_5, cpp-greedy_causal_lm-redpajama-3b-chat, cpp-chat_sample-ubuntu, cpp-continuous-batching-ubuntu, visual_language_chat_sample-ubuntu, cpp-continuous-batching-windows, cpp-continuous-batching-macos] diff --git a/tests/python_tests/ov_genai_test_utils.py b/tests/python_tests/ov_genai_test_utils.py index d9e28af400..59c11cb528 100644 --- a/tests/python_tests/ov_genai_test_utils.py +++ b/tests/python_tests/ov_genai_test_utils.py @@ -25,7 +25,7 @@ def get_models_list(): "microsoft/phi-1_5", "microsoft/phi-2", "THUDM/chatglm2-6b", - "THUDM/chatglm3-6b", + "THUDM/chatglm3-6b", # no beam_search "Qwen/Qwen2-0.5B-Instruct", "Qwen/Qwen-7B-Chat", "Qwen/Qwen1.5-7B-Chat", From a3a01ee3a5dfd284e6dfd239407fe7a2da8dd010 Mon Sep 17 00:00:00 2001 From: Roy Date: Fri, 1 Nov 2024 01:31:48 +0530 Subject: [PATCH 03/29] extended SUPPORTED_MODELS --- src/docs/SUPPORTED_MODELS.md | 3 +++ 1 file changed, 3 insertions(+) diff --git a/src/docs/SUPPORTED_MODELS.md b/src/docs/SUPPORTED_MODELS.md index e1f2483d8e..3edc41337c 100644 --- a/src/docs/SUPPORTED_MODELS.md +++ b/src/docs/SUPPORTED_MODELS.md @@ -157,6 +157,9 @@ The pipeline can work with other similar topologies produced by `optimum-intel` > [!NOTE] > Models should belong to the same family and have the same tokenizers. +>[!Note] +> The `beam_search_causal_lm` is not supported in the `ChatGLM3-6B model`. + ## Text 2 image models From 53711c5f996d6a13183b94bba838d0eb26b800e2 Mon Sep 17 00:00:00 2001 From: Roy Date: Fri, 1 Nov 2024 14:54:08 +0530 Subject: [PATCH 04/29] updated -causal_lm_cpp.yml-removed-cpp-prompt_lookup_decoding_lm-ubuntu-Chatglm3-6b-and-merged-in-cpp-prompt_lookup_decoding_lm-ubuntu --- .github/workflows/causal_lm_cpp.yml | 59 +++-------------------------- 1 file changed, 5 insertions(+), 54 deletions(-) diff --git a/.github/workflows/causal_lm_cpp.yml b/.github/workflows/causal_lm_cpp.yml index ed8b5b5c02..9999002403 100644 --- a/.github/workflows/causal_lm_cpp.yml +++ b/.github/workflows/causal_lm_cpp.yml @@ -498,6 +498,7 @@ jobs: env: PYTHONPATH: "./build/:$PYTHONPATH" LD_LIBRARY_PATH: "./build/openvino_genai/:$LD_LIBRARY_PATH" + cpp-prompt_lookup_decoding_lm-ubuntu: runs-on: ubuntu-20.04-16-cores defaults: @@ -527,6 +528,7 @@ jobs: python -m pip install ./thirdparty/openvino_tokenizers/[transformers] --pre --extra-index-url https://storage.openvinotoolkit.org/simple/wheels/nightly optimum-cli export openvino --trust-remote-code --weight-format fp16 --model TinyLlama/TinyLlama-1.1B-Chat-v1.0 TinyLlama-1.1B-Chat-v1.0 optimum-cli export openvino --trust-remote-code --weight-format fp16 --model Qwen/Qwen-7B-Chat Qwen-7B-Chat --task text-generation-with-past + optimum-cli export openvino --trust-remote-code --task text-generation-with-past --model THUDM/chatglm3-6b chatglm3-6b - name: run and compare run: | source ./ov/setupvars.sh @@ -569,36 +571,6 @@ jobs: assert predicted_greedy == predicted_prompt_lookup " echo "Prompt lookup" passed - - cpp-prompt_lookup_decoding_lm-ubuntu-Chatglm3-6b: - runs-on: ubuntu-20.04-16-cores - defaults: - run: - shell: bash - steps: - - uses: actions/checkout@v4 - with: - submodules: recursive - - uses: actions/setup-python@v4 - with: - python-version: 3.12 - - name: Install OpenVINO - run: | - mkdir ./ov/ - curl ${{ env.l_ov_link }} | tar --directory ./ov/ --strip-components 1 -xz - sudo ./ov/install_dependencies/install_openvino_dependencies.sh - - name: Build app - run: | - source ./ov/setupvars.sh - cmake -DCMAKE_BUILD_TYPE=Release -S ./ -B ./build/ - cmake --build ./build/ --config Release -j - - name: Download and convert and model - run: | - source ./ov/setupvars.sh - python -m pip install --upgrade-strategy eager -r ./samples/requirements.txt - python -m pip install ./thirdparty/openvino_tokenizers/[transformers] --pre --extra-index-url https://storage.openvinotoolkit.org/simple/wheels/nightly - optimum-cli export openvino --trust-remote-code --weight-format fp16 --model TinyLlama/TinyLlama-1.1B-Chat-v1.0 TinyLlama-1.1B-Chat-v1.0 - optimum-cli export openvino --trust-remote-code --task text-generation-with-past --model THUDM/chatglm3-6b chatglm3-6b - name: run and compare run: | source ./ov/setupvars.sh @@ -610,29 +582,8 @@ jobs: Question: Can you please add 2 and 3 A:' > ./prompt.txt - ./build/samples/cpp/prompt_lookup_decoding_lm/prompt_lookup_decoding_lm ./TinyLlama-1.1B-Chat-v1.0/ "$( predictions_prompt_lookup.txt - ./build/samples/cpp/greedy_causal_lm/greedy_causal_lm ./TinyLlama-1.1B-Chat-v1.0/ "$( predictions_greedy.txt - python -c " - with open('predictions_greedy.txt', 'r') as f: - predicted_greedy = f.readline() - with open('predictions_prompt_lookup.txt', 'r') as f: - predicted_prompt_lookup = f.readline() - assert predicted_greedy == predicted_prompt_lookup - " - echo "Prompt lookup" passed - - name: run and compare (model with seq_length_axis = 1) - run: | - source ./ov/setupvars.sh - - echo 'Code:```python - def add(a, b): - return a + b - ``` - Question: Can you please add 2 and 3 - A:' > ./prompt.txt - - ./build/samples/cpp/prompt_lookup_decoding_lm/prompt_lookup_decoding_lm ./Qwen-7B-Chat/ "$( predictions_prompt_lookup.txt - ./build/samples/cpp/greedy_causal_lm/greedy_causal_lm ./Qwen-7B-Chat/ "$( predictions_greedy.txt + ./build/samples/cpp/prompt_lookup_decoding_lm/prompt_lookup_decoding_lm ./chatglm3-6b/ "$( predictions_prompt_lookup.txt + ./build/samples/cpp/greedy_causal_lm/greedy_causal_lm ./chatglm3-6b/ "$( predictions_greedy.txt python -c " with open('predictions_greedy.txt', 'r') as f: predicted_greedy = f.readline() @@ -1079,7 +1030,7 @@ jobs: name: ci/gha_overall_status_causal_lm needs: [cpp-multinomial-greedy_causal_lm-ubuntu, cpp-beam_search_causal_lm-ubuntu, cpp-greedy_causal_lm-windows,cpp-greedy_causal_lm-Chatglm3-6b, cpp-greedy_causal_lm-Qwen-7B-Chat, cpp-beam_search_causal_lm-Qwen1_5-7B-Chat, cpp-beam_search_causal_lm-Phi-2, - cpp-beam_search_causal_lm-notus-7b-v1, cpp-speculative_decoding_lm-ubuntu, cpp-prompt_lookup_decoding_lm-ubuntu,cpp-prompt_lookup_decoding_lm-ubuntu-Chatglm3-6b, + cpp-beam_search_causal_lm-notus-7b-v1, cpp-speculative_decoding_lm-ubuntu, cpp-prompt_lookup_decoding_lm-ubuntu, cpp-Phi-1_5, cpp-greedy_causal_lm-redpajama-3b-chat, cpp-chat_sample-ubuntu, cpp-continuous-batching-ubuntu, visual_language_chat_sample-ubuntu, cpp-continuous-batching-windows, cpp-continuous-batching-macos] From 4532c96536747a3d3e4a958da56a2fbd0e7a7121 Mon Sep 17 00:00:00 2001 From: Roy Date: Fri, 1 Nov 2024 15:03:36 +0530 Subject: [PATCH 05/29] updated_supported_model.md_as_asked --- src/docs/SUPPORTED_MODELS.md | 7 +++++-- 1 file changed, 5 insertions(+), 2 deletions(-) diff --git a/src/docs/SUPPORTED_MODELS.md b/src/docs/SUPPORTED_MODELS.md index 3edc41337c..abc6acbc8e 100644 --- a/src/docs/SUPPORTED_MODELS.md +++ b/src/docs/SUPPORTED_MODELS.md @@ -157,8 +157,11 @@ The pipeline can work with other similar topologies produced by `optimum-intel` > [!NOTE] > Models should belong to the same family and have the same tokenizers. ->[!Note] -> The `beam_search_causal_lm` is not supported in the `ChatGLM3-6B model`. +>[!NOTE] +>`optimum-cli` requires the `--task text-generation-with-past` argument for the `THUDM/chatglm3-6b` model + +>[!NOTE] +> The **beam search** is not supported for `ChatGLM3-6B model`. ## Text 2 image models From e2c73b8a31fe38d64b61aa4667a415973a98aeac Mon Sep 17 00:00:00 2001 From: Aniruddha521 <126981371+Aniruddha521@users.noreply.github.com> Date: Fri, 1 Nov 2024 15:11:33 +0530 Subject: [PATCH 06/29] Update SUPPORTED_MODELS.md --- src/docs/SUPPORTED_MODELS.md | 10 +++------- 1 file changed, 3 insertions(+), 7 deletions(-) diff --git a/src/docs/SUPPORTED_MODELS.md b/src/docs/SUPPORTED_MODELS.md index 2444b5a041..e53437be80 100644 --- a/src/docs/SUPPORTED_MODELS.md +++ b/src/docs/SUPPORTED_MODELS.md @@ -155,13 +155,9 @@ The pipeline can work with other similar topologies produced by `optimum-intel` 4. `position_ids` (optional) encodes a position of currently generating token in the sequence and a single `logits` output. > [!NOTE] -> Models should belong to the same family and have the same tokenizers. - ->[!NOTE] ->`optimum-cli` requires the `--task text-generation-with-past` argument for the `THUDM/chatglm3-6b` model - ->[!NOTE] -> The **beam search** is not supported for `ChatGLM3-6B model`. +>* Models should belong to the same family and have the same tokenizers. +>* `optimum-cli` requires the `--task text-generation-with-past` argument for the `THUDM/chatglm3-6b` model +>* The **beam search** is not supported for `ChatGLM3-6B model`. ## Text 2 image models From c190798b2e539b34a843fd4ad71ab6ec715f305e Mon Sep 17 00:00:00 2001 From: Roy Date: Fri, 1 Nov 2024 18:49:55 +0530 Subject: [PATCH 07/29] updated-causal-lm --- .github/workflows/causal_lm_cpp.yml | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/.github/workflows/causal_lm_cpp.yml b/.github/workflows/causal_lm_cpp.yml index 9999002403..361c741a39 100644 --- a/.github/workflows/causal_lm_cpp.yml +++ b/.github/workflows/causal_lm_cpp.yml @@ -275,7 +275,7 @@ jobs: && python samples\python\greedy_causal_lm\lora.py .\TinyLlama\TinyLlama-1.1B-intermediate-step-1431k-3T\ adapter_model.safetensors "How to create a table with two columns, one of them has type float, another one has type int?" cpp-greedy_causal_lm-Chatglm3-6b: - runs-on: ubuntu-24.04 + runs-on: ubuntu-20.04-16-cores defaults: run: shell: bash @@ -301,7 +301,7 @@ jobs: source ./ov/setupvars.sh python -m pip install --upgrade-strategy eager -r ./samples/requirements.txt python -m pip install ./thirdparty/openvino_tokenizers/[transformers] --pre --extra-index-url https://storage.openvinotoolkit.org/simple/wheels/nightly - optimum-cli export openvino --trust-remote-code --task text-generation-with-past --model THUDM/chatglm3-6b chatglm3-6b + optimum-cli export openvino --trust-remote-code --weight-format fp16 --model THUDM/chatglm3-6b chatglm3-6b --task text-generation-with-past - run: > . ./ov/setupvars.sh && timeout 2m ./build/samples/cpp/greedy_causal_lm/greedy_causal_lm ./chatglm3-6b/ 69 | diff <(timeout 2m samples/python/greedy_causal_lm/greedy_causal_lm.py ./chatglm3-6b/ 69) - @@ -528,7 +528,7 @@ jobs: python -m pip install ./thirdparty/openvino_tokenizers/[transformers] --pre --extra-index-url https://storage.openvinotoolkit.org/simple/wheels/nightly optimum-cli export openvino --trust-remote-code --weight-format fp16 --model TinyLlama/TinyLlama-1.1B-Chat-v1.0 TinyLlama-1.1B-Chat-v1.0 optimum-cli export openvino --trust-remote-code --weight-format fp16 --model Qwen/Qwen-7B-Chat Qwen-7B-Chat --task text-generation-with-past - optimum-cli export openvino --trust-remote-code --task text-generation-with-past --model THUDM/chatglm3-6b chatglm3-6b + optimum-cli export openvino --trust-remote-code --weight-format fp16 --model THUDM/chatglm3-6b chatglm3-6b --task text-generation-with-past - name: run and compare run: | source ./ov/setupvars.sh From 2e2b2937ba2c7ccde0df3977bdbcef21a366017e Mon Sep 17 00:00:00 2001 From: Aniruddha521 <126981371+Aniruddha521@users.noreply.github.com> Date: Mon, 4 Nov 2024 21:21:10 +0530 Subject: [PATCH 08/29] Update causal_lm_cpp.yml --- .github/workflows/causal_lm_cpp.yml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/.github/workflows/causal_lm_cpp.yml b/.github/workflows/causal_lm_cpp.yml index 361c741a39..fe6191ad4c 100644 --- a/.github/workflows/causal_lm_cpp.yml +++ b/.github/workflows/causal_lm_cpp.yml @@ -304,7 +304,7 @@ jobs: optimum-cli export openvino --trust-remote-code --weight-format fp16 --model THUDM/chatglm3-6b chatglm3-6b --task text-generation-with-past - run: > . ./ov/setupvars.sh - && timeout 2m ./build/samples/cpp/greedy_causal_lm/greedy_causal_lm ./chatglm3-6b/ 69 | diff <(timeout 2m samples/python/greedy_causal_lm/greedy_causal_lm.py ./chatglm3-6b/ 69) - + && timeout 2m ./build/samples/cpp/greedy_causal_lm/greedy_causal_lm ./chatglm3-6b/ 69 "why is sun yellow?" env: PYTHONPATH: "./build" From 929275725c1626dba7ff393380438dbf3305658e Mon Sep 17 00:00:00 2001 From: Roy Date: Tue, 5 Nov 2024 11:11:27 +0530 Subject: [PATCH 09/29] updating branch --- thirdparty/openvino_tokenizers | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/thirdparty/openvino_tokenizers b/thirdparty/openvino_tokenizers index 16da7f3901..1ddb12edca 160000 --- a/thirdparty/openvino_tokenizers +++ b/thirdparty/openvino_tokenizers @@ -1 +1 @@ -Subproject commit 16da7f39010daa04809f9552fa00f53ac521439b +Subproject commit 1ddb12edca0e6a9fb02b37a3e1565111dfd22d98 From 4028323f993cf307c6fc13d079cf92b099f49b40 Mon Sep 17 00:00:00 2001 From: Aniruddha521 <126981371+Aniruddha521@users.noreply.github.com> Date: Tue, 5 Nov 2024 15:51:57 +0530 Subject: [PATCH 10/29] Updated cpp-prompt_lookup_decoding_lm-ubuntu and cpp-greedy_causal_lm-Chatglm3-6b incausal_lm_cpp.yml --- .github/workflows/causal_lm_cpp.yml | 31 ++++++++++++++++++++++++----- 1 file changed, 26 insertions(+), 5 deletions(-) diff --git a/.github/workflows/causal_lm_cpp.yml b/.github/workflows/causal_lm_cpp.yml index cc94247074..6e7f90b2e1 100644 --- a/.github/workflows/causal_lm_cpp.yml +++ b/.github/workflows/causal_lm_cpp.yml @@ -302,13 +302,32 @@ jobs: python -m pip install --upgrade-strategy eager -r ./samples/requirements.txt python -m pip install ./thirdparty/openvino_tokenizers/[transformers] --pre --extra-index-url https://storage.openvinotoolkit.org/simple/wheels/nightly optimum-cli export openvino --trust-remote-code --weight-format fp16 --model THUDM/chatglm3-6b chatglm3-6b --task text-generation-with-past - - run: > + - name: Run greedy_causal_lam C++ sample - Chatglm3-6b + run: > . ./ov/setupvars.sh - && timeout 2m ./build/samples/cpp/greedy_causal_lm/greedy_causal_lm ./chatglm3-6b/ 69 "why is sun yellow?" + && ./build/samples/cpp/greedy_causal_lm/greedy_causal_lm ./chatglm3-6b/ "Why sun is yellow?" | tee cpp.txt + timeout-minutes: 2 + - name: Run greedy_causal_lam Python sample - Chatglm3-6b + run: > + . ./ov/setupvars.sh + && /samples/cpp/greedy_causal_lm/greedy_causal_lm ./chatglm3-6b/ "Why sun is yellow?" | tee py.txt env: PYTHONPATH: "./build" - - + - name: Visualizing and Comparing generated output + run: > + python -c" + with open("cpp.txt", 'r') as f: + greedy_cp = f.read() + with open("py.txt", 'r') as f: + greedy_py = f.read() + print(f"Generated with C++ sample: {greedy_cp}") + print(f"Generated with Python sample: {greedy_py}") + " + diff cp.txt py.txt + echo "Why sun is yellow?" passed + env: + PYTHONPATH: "./build" + cpp-greedy_causal_lm-Qwen-7B-Chat: runs-on: ubuntu-20.04-16-cores defaults: @@ -584,12 +603,14 @@ jobs: ./build/samples/cpp/prompt_lookup_decoding_lm/prompt_lookup_decoding_lm ./chatglm3-6b/ "$( predictions_prompt_lookup.txt ./build/samples/cpp/greedy_causal_lm/greedy_causal_lm ./chatglm3-6b/ "$( predictions_greedy.txt + diff predicted_prompt_lookup predictions_greedy.txt python -c " with open('predictions_greedy.txt', 'r') as f: predicted_greedy = f.readline() with open('predictions_prompt_lookup.txt', 'r') as f: predicted_prompt_lookup = f.readline() - assert predicted_greedy == predicted_prompt_lookup + print(predicted_greedy) + print(predicted_prompt_lookup) " echo "Prompt lookup" passed From b70302cc4a7fefbb6b6437904eaa10e9614b88cd Mon Sep 17 00:00:00 2001 From: Aniruddha521 <126981371+Aniruddha521@users.noreply.github.com> Date: Tue, 5 Nov 2024 16:32:43 +0530 Subject: [PATCH 11/29] Update causal_lm_cpp.yml --- .github/workflows/causal_lm_cpp.yml | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/.github/workflows/causal_lm_cpp.yml b/.github/workflows/causal_lm_cpp.yml index 6e7f90b2e1..ed9e287f12 100644 --- a/.github/workflows/causal_lm_cpp.yml +++ b/.github/workflows/causal_lm_cpp.yml @@ -310,7 +310,7 @@ jobs: - name: Run greedy_causal_lam Python sample - Chatglm3-6b run: > . ./ov/setupvars.sh - && /samples/cpp/greedy_causal_lm/greedy_causal_lm ./chatglm3-6b/ "Why sun is yellow?" | tee py.txt + && samples/python/greedy_causal_lm/greedy_causal_lm.py ./chatglm3-6b/ "Why sun is yellow?" | tee py.txt env: PYTHONPATH: "./build" - name: Visualizing and Comparing generated output @@ -603,7 +603,7 @@ jobs: ./build/samples/cpp/prompt_lookup_decoding_lm/prompt_lookup_decoding_lm ./chatglm3-6b/ "$( predictions_prompt_lookup.txt ./build/samples/cpp/greedy_causal_lm/greedy_causal_lm ./chatglm3-6b/ "$( predictions_greedy.txt - diff predicted_prompt_lookup predictions_greedy.txt + diff predicted_prompt_lookup.txt predictions_greedy.txt python -c " with open('predictions_greedy.txt', 'r') as f: predicted_greedy = f.readline() From f2519857331dd8ab24d106b88dee7dc1a22393ac Mon Sep 17 00:00:00 2001 From: Aniruddha521 <126981371+Aniruddha521@users.noreply.github.com> Date: Tue, 5 Nov 2024 17:09:50 +0530 Subject: [PATCH 12/29] Update causal_lm_cpp.yml --- .github/workflows/causal_lm_cpp.yml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/.github/workflows/causal_lm_cpp.yml b/.github/workflows/causal_lm_cpp.yml index ed9e287f12..0c3dbb1d08 100644 --- a/.github/workflows/causal_lm_cpp.yml +++ b/.github/workflows/causal_lm_cpp.yml @@ -315,7 +315,7 @@ jobs: PYTHONPATH: "./build" - name: Visualizing and Comparing generated output run: > - python -c" + python -c " with open("cpp.txt", 'r') as f: greedy_cp = f.read() with open("py.txt", 'r') as f: From 87db4568e3f32380e78e10afe2eb2e453502b4b6 Mon Sep 17 00:00:00 2001 From: Roy Date: Tue, 5 Nov 2024 23:58:22 +0530 Subject: [PATCH 13/29] updated causal_lm_cpp.yml --- .github/workflows/causal_lm_cpp.yml | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/.github/workflows/causal_lm_cpp.yml b/.github/workflows/causal_lm_cpp.yml index 0c3dbb1d08..aa0629f38e 100644 --- a/.github/workflows/causal_lm_cpp.yml +++ b/.github/workflows/causal_lm_cpp.yml @@ -314,7 +314,7 @@ jobs: env: PYTHONPATH: "./build" - name: Visualizing and Comparing generated output - run: > + run: | python -c " with open("cpp.txt", 'r') as f: greedy_cp = f.read() @@ -603,7 +603,7 @@ jobs: ./build/samples/cpp/prompt_lookup_decoding_lm/prompt_lookup_decoding_lm ./chatglm3-6b/ "$( predictions_prompt_lookup.txt ./build/samples/cpp/greedy_causal_lm/greedy_causal_lm ./chatglm3-6b/ "$( predictions_greedy.txt - diff predicted_prompt_lookup.txt predictions_greedy.txt + diff predictions_prompt_lookup.txt predictions_greedy.txt python -c " with open('predictions_greedy.txt', 'r') as f: predicted_greedy = f.readline() From 65b78912209d2ca340b8fc5639e289d3e8cb8f79 Mon Sep 17 00:00:00 2001 From: Aniruddha521 <126981371+Aniruddha521@users.noreply.github.com> Date: Thu, 7 Nov 2024 12:50:51 +0530 Subject: [PATCH 14/29] Update causal_lm_cpp.yml --- .github/workflows/causal_lm_cpp.yml | 10 +++++----- 1 file changed, 5 insertions(+), 5 deletions(-) diff --git a/.github/workflows/causal_lm_cpp.yml b/.github/workflows/causal_lm_cpp.yml index d21bd70255..68f6fc67de 100644 --- a/.github/workflows/causal_lm_cpp.yml +++ b/.github/workflows/causal_lm_cpp.yml @@ -1020,11 +1020,11 @@ jobs: Overall_Status: name: ci/gha_overall_status_causal_lm - needs: [cpp-multinomial-greedy_causal_lm-ubuntu, cpp-beam_search_causal_lm-ubuntu, cpp-greedy_causal_lm-windows,cpp-greedy_causal_lm-Chatglm3-6b, - cpp-greedy_causal_lm-Qwen-7B-Chat, cpp-beam_search_causal_lm-Qwen1_5-7B-Chat, cpp-beam_search_causal_lm-Phi-2, - cpp-beam_search_causal_lm-notus-7b-v1, cpp-speculative_decoding_lm-ubuntu, cpp-prompt_lookup_decoding_lm-ubuntu, - cpp-Phi-1_5, cpp-greedy_causal_lm-redpajama-3b-chat, cpp-chat_sample-ubuntu, cpp-continuous-batching-ubuntu, - visual_language_chat_sample-ubuntu, + needs: [cpp-multinomial-greedy_causal_lm-ubuntu, cpp-beam_search_causal_lm-ubuntu, cpp-greedy_causal_lm-windows, + cpp-greedy_causal_lm-Chatglm3-6b, cpp-greedy_causal_lm-Qwen-7B-Chat, cpp-beam_search_causal_lm-Qwen1_5-7B-Chat, + cpp-beam_search_causal_lm-Phi-2, cpp-beam_search_causal_lm-notus-7b-v1, cpp-speculative_decoding_lm-ubuntu, + cpp-prompt_lookup_decoding_lm-ubuntu, cpp-Phi-1_5, cpp-greedy_causal_lm-redpajama-3b-chat, cpp-chat_sample-ubuntu, + cpp-continuous-batching-ubuntu, visual_language_chat_sample-ubuntu, cpp-continuous-batching-windows, cpp-continuous-batching-macos] if: ${{ always() }} runs-on: ubuntu-latest From ccd9ba9001665ae75d4aaa85d6a3e8fe7f03cbc9 Mon Sep 17 00:00:00 2001 From: Roy Date: Tue, 12 Nov 2024 14:24:34 +0530 Subject: [PATCH 15/29] tokenizers --- thirdparty/openvino_tokenizers | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/thirdparty/openvino_tokenizers b/thirdparty/openvino_tokenizers index 1ddb12edca..3fe6677f27 160000 --- a/thirdparty/openvino_tokenizers +++ b/thirdparty/openvino_tokenizers @@ -1 +1 @@ -Subproject commit 1ddb12edca0e6a9fb02b37a3e1565111dfd22d98 +Subproject commit 3fe6677f27c7f5eb149b959d5cd80dff76e9924b From cca410a103ff3080e9bffa9029cdea841b23a7d4 Mon Sep 17 00:00:00 2001 From: Roy Date: Tue, 12 Nov 2024 19:25:36 +0530 Subject: [PATCH 16/29] slight modification --- .github/workflows/causal_lm_cpp.yml | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/.github/workflows/causal_lm_cpp.yml b/.github/workflows/causal_lm_cpp.yml index 5752b443bc..ccf81abef1 100644 --- a/.github/workflows/causal_lm_cpp.yml +++ b/.github/workflows/causal_lm_cpp.yml @@ -320,8 +320,8 @@ jobs: greedy_cp = f.read() with open("py.txt", 'r') as f: greedy_py = f.read() - print(f"Generated with C++ sample: {greedy_cp}") - print(f"Generated with Python sample: {greedy_py}") + print(greedy_cp) + print(greedy_py) " diff cp.txt py.txt echo "Why sun is yellow?" passed From bdd716ca9a253ae32287ae6f39920148d503b986 Mon Sep 17 00:00:00 2001 From: Aniruddha521 <126981371+Aniruddha521@users.noreply.github.com> Date: Tue, 5 Nov 2024 16:32:43 +0530 Subject: [PATCH 17/29] Update causal_lm_cpp.yml solved conflicts --- thirdparty/openvino_tokenizers | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/thirdparty/openvino_tokenizers b/thirdparty/openvino_tokenizers index 306dcd8dae..3fe6677f27 160000 --- a/thirdparty/openvino_tokenizers +++ b/thirdparty/openvino_tokenizers @@ -1 +1 @@ -Subproject commit 306dcd8daec36bbc680c50c68de1e954f42b0ab8 +Subproject commit 3fe6677f27c7f5eb149b959d5cd80dff76e9924b From 35fd401b8e986d2863a43d6cd75d5579ae2eaf2b Mon Sep 17 00:00:00 2001 From: Roy Date: Tue, 12 Nov 2024 19:53:26 +0530 Subject: [PATCH 18/29] checking --- .github/workflows/causal_lm_cpp.yml | 6 +++++- 1 file changed, 5 insertions(+), 1 deletion(-) diff --git a/.github/workflows/causal_lm_cpp.yml b/.github/workflows/causal_lm_cpp.yml index ccf81abef1..5c14e22a48 100644 --- a/.github/workflows/causal_lm_cpp.yml +++ b/.github/workflows/causal_lm_cpp.yml @@ -320,7 +320,9 @@ jobs: greedy_cp = f.read() with open("py.txt", 'r') as f: greedy_py = f.read() + print('In c++ generated') print(greedy_cp) + print('In python generated') print(greedy_py) " diff cp.txt py.txt @@ -603,15 +605,17 @@ jobs: ./build/samples/cpp/prompt_lookup_decoding_lm/prompt_lookup_decoding_lm ./chatglm3-6b/ "$( predictions_prompt_lookup.txt ./build/samples/cpp/greedy_causal_lm/greedy_causal_lm ./chatglm3-6b/ "$( predictions_greedy.txt - diff predictions_prompt_lookup.txt predictions_greedy.txt python -c " with open('predictions_greedy.txt', 'r') as f: predicted_greedy = f.readline() with open('predictions_prompt_lookup.txt', 'r') as f: predicted_prompt_lookup = f.readline() + print('greedy generated') print(predicted_greedy) + print('Prompt lookup generated') print(predicted_prompt_lookup) " + diff predictions_prompt_lookup.txt predictions_greedy.txt echo "Prompt lookup" passed cpp-Phi-1_5: From 55f493b31a3344ce93f1860b906d8be82133a8e2 Mon Sep 17 00:00:00 2001 From: Roy Date: Tue, 12 Nov 2024 20:56:39 +0530 Subject: [PATCH 19/29] usual check --- .github/workflows/causal_lm_cpp.yml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/.github/workflows/causal_lm_cpp.yml b/.github/workflows/causal_lm_cpp.yml index 5c14e22a48..d723c28244 100644 --- a/.github/workflows/causal_lm_cpp.yml +++ b/.github/workflows/causal_lm_cpp.yml @@ -325,7 +325,7 @@ jobs: print('In python generated') print(greedy_py) " - diff cp.txt py.txt + diff cpp.txt py.txt echo "Why sun is yellow?" passed env: PYTHONPATH: "./build" From cafa42a0dadf84618471e27b98077c6f63d5bb91 Mon Sep 17 00:00:00 2001 From: Aniruddha521 <126981371+Aniruddha521@users.noreply.github.com> Date: Wed, 13 Nov 2024 10:04:14 +0530 Subject: [PATCH 20/29] Update causal_lm_cpp.yml --- .github/workflows/causal_lm_cpp.yml | 18 ++++++------------ 1 file changed, 6 insertions(+), 12 deletions(-) diff --git a/.github/workflows/causal_lm_cpp.yml b/.github/workflows/causal_lm_cpp.yml index d723c28244..138c16e18f 100644 --- a/.github/workflows/causal_lm_cpp.yml +++ b/.github/workflows/causal_lm_cpp.yml @@ -302,19 +302,13 @@ jobs: python -m pip install --upgrade-strategy eager -r ./samples/requirements.txt python -m pip install ./thirdparty/openvino_tokenizers/[transformers] --pre --extra-index-url https://storage.openvinotoolkit.org/simple/wheels/nightly optimum-cli export openvino --trust-remote-code --weight-format fp16 --model THUDM/chatglm3-6b chatglm3-6b --task text-generation-with-past - - name: Run greedy_causal_lam C++ sample - Chatglm3-6b - run: > - . ./ov/setupvars.sh - && ./build/samples/cpp/greedy_causal_lm/greedy_causal_lm ./chatglm3-6b/ "Why sun is yellow?" | tee cpp.txt - timeout-minutes: 2 - - name: Run greedy_causal_lam Python sample - Chatglm3-6b - run: > - . ./ov/setupvars.sh - && samples/python/greedy_causal_lm/greedy_causal_lm.py ./chatglm3-6b/ "Why sun is yellow?" | tee py.txt - env: - PYTHONPATH: "./build" - - name: Visualizing and Comparing generated output + - name: Run C++ and Python Samples and Comparing generated output - Chatglm3-6b run: | + . ./ov/setupvars.sh + echo "why sun is yellow?" > ./prompt.txt + + ./build/samples/cpp/greedy_causal_lm/greedy_causal_lm ./chatglm3-6b/ "$( Date: Tue, 12 Nov 2024 20:56:39 +0530 Subject: [PATCH 21/29] usual check updated submodule with git checkout a0268cd5c5fe71ccbc4dc773b502075867c859fe --- .github/workflows/causal_lm_cpp.yml | 2 +- thirdparty/openvino_tokenizers | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/.github/workflows/causal_lm_cpp.yml b/.github/workflows/causal_lm_cpp.yml index 5c14e22a48..d723c28244 100644 --- a/.github/workflows/causal_lm_cpp.yml +++ b/.github/workflows/causal_lm_cpp.yml @@ -325,7 +325,7 @@ jobs: print('In python generated') print(greedy_py) " - diff cp.txt py.txt + diff cpp.txt py.txt echo "Why sun is yellow?" passed env: PYTHONPATH: "./build" diff --git a/thirdparty/openvino_tokenizers b/thirdparty/openvino_tokenizers index 3fe6677f27..306dcd8dae 160000 --- a/thirdparty/openvino_tokenizers +++ b/thirdparty/openvino_tokenizers @@ -1 +1 @@ -Subproject commit 3fe6677f27c7f5eb149b959d5cd80dff76e9924b +Subproject commit 306dcd8daec36bbc680c50c68de1e954f42b0ab8 From 0d6e58e44d6c635ef025da9e0451bfdaf234faf3 Mon Sep 17 00:00:00 2001 From: Roy Date: Wed, 13 Nov 2024 12:45:47 +0530 Subject: [PATCH 22/29] minor changes --- .github/workflows/causal_lm_cpp.yml | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-) diff --git a/.github/workflows/causal_lm_cpp.yml b/.github/workflows/causal_lm_cpp.yml index 138c16e18f..ef13c60146 100644 --- a/.github/workflows/causal_lm_cpp.yml +++ b/.github/workflows/causal_lm_cpp.yml @@ -307,12 +307,12 @@ jobs: . ./ov/setupvars.sh echo "why sun is yellow?" > ./prompt.txt - ./build/samples/cpp/greedy_causal_lm/greedy_causal_lm ./chatglm3-6b/ "$( cpp.txt + samples/python/greedy_causal_lm/greedy_causal_lm.py ./chatglm3-6b/ "$( py.txt python -c " - with open("cpp.txt", 'r') as f: + with open('cpp.txt', 'r') as f: greedy_cp = f.read() - with open("py.txt", 'r') as f: + with open('py.txt', 'r') as f: greedy_py = f.read() print('In c++ generated') print(greedy_cp) From 0f39e8c3ee23044ff6ce0c1c3880f20ce0cbaef3 Mon Sep 17 00:00:00 2001 From: Roy Date: Sat, 23 Nov 2024 03:55:53 +0530 Subject: [PATCH 23/29] few modification --- .github/workflows/causal_lm_cpp.yml | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-) diff --git a/.github/workflows/causal_lm_cpp.yml b/.github/workflows/causal_lm_cpp.yml index ef13c60146..eca6185dda 100644 --- a/.github/workflows/causal_lm_cpp.yml +++ b/.github/workflows/causal_lm_cpp.yml @@ -307,8 +307,8 @@ jobs: . ./ov/setupvars.sh echo "why sun is yellow?" > ./prompt.txt - ./build/samples/cpp/greedy_causal_lm/greedy_causal_lm ./chatglm3-6b/ "$( cpp.txt - samples/python/greedy_causal_lm/greedy_causal_lm.py ./chatglm3-6b/ "$( py.txt + ./build/samples/cpp/greedy_causal_lm/greedy_causal_lm ./chatglm3-6b/ "$( ./prompt.txt - ./build/samples/cpp/prompt_lookup_decoding_lm/prompt_lookup_decoding_lm ./chatglm3-6b/ "$( predictions_prompt_lookup.txt - ./build/samples/cpp/greedy_causal_lm/greedy_causal_lm ./chatglm3-6b/ "$( predictions_greedy.txt + ./build/samples/cpp/prompt_lookup_decoding_lm/prompt_lookup_decoding_lm ./chatglm3-6b/ "$( Date: Sat, 23 Nov 2024 22:23:42 +0530 Subject: [PATCH 24/29] Update causal_lm_cpp.yml --- .github/workflows/causal_lm_cpp.yml | 4 ++++ 1 file changed, 4 insertions(+) diff --git a/.github/workflows/causal_lm_cpp.yml b/.github/workflows/causal_lm_cpp.yml index 71bbaaaaee..be1de0b933 100644 --- a/.github/workflows/causal_lm_cpp.yml +++ b/.github/workflows/causal_lm_cpp.yml @@ -597,8 +597,11 @@ jobs: Question: Can you please add 2 and 3 A:' > ./prompt.txt + echo '-------------------------------Prompt lookup Generated-----------------------------------------' ./build/samples/cpp/prompt_lookup_decoding_lm/prompt_lookup_decoding_lm ./chatglm3-6b/ "$( Date: Tue, 26 Nov 2024 16:39:56 +0530 Subject: [PATCH 25/29] Update causal_lm_cpp.yml --- .github/workflows/causal_lm_cpp.yml | 5 +++-- 1 file changed, 3 insertions(+), 2 deletions(-) diff --git a/.github/workflows/causal_lm_cpp.yml b/.github/workflows/causal_lm_cpp.yml index be1de0b933..d2e765cdbb 100644 --- a/.github/workflows/causal_lm_cpp.yml +++ b/.github/workflows/causal_lm_cpp.yml @@ -611,9 +611,10 @@ jobs: print(predicted_greedy) print('Prompt lookup generated') print(predicted_prompt_lookup) + print('--------------------------------------------------------------------') + assert predicted_greedy == predicted_prompt_lookup + print('Passes') " - echo '-------------------------------Diff--------------------------------------' - diff predictions_prompt_lookup.txt predictions_greedy.txt echo "Prompt lookup" passed cpp-Phi-1_5: From 9fa72e62ff12dda0858c380e4bfe2f3fb50b08f6 Mon Sep 17 00:00:00 2001 From: Roy Date: Sat, 30 Nov 2024 03:34:25 +0530 Subject: [PATCH 26/29] modification in github workflow(casual_lm_cpp) --- .github/workflows/causal_lm_cpp.yml | 19 +++---------------- 1 file changed, 3 insertions(+), 16 deletions(-) diff --git a/.github/workflows/causal_lm_cpp.yml b/.github/workflows/causal_lm_cpp.yml index 72aed42b4c..a0216d777b 100644 --- a/.github/workflows/causal_lm_cpp.yml +++ b/.github/workflows/causal_lm_cpp.yml @@ -306,19 +306,11 @@ jobs: run: | . ./ov/setupvars.sh echo "why sun is yellow?" > ./prompt.txt - + echo '--------------------------------C++ Generated--------------------------------------------------' ./build/samples/cpp/greedy_causal_lm/greedy_causal_lm ./chatglm3-6b/ "$( Date: Wed, 4 Dec 2024 12:45:38 +0530 Subject: [PATCH 27/29] Suggested changes --- .github/workflows/causal_lm_cpp.yml | 3 +-- src/docs/SUPPORTED_MODELS.md | 3 +-- 2 files changed, 2 insertions(+), 4 deletions(-) diff --git a/.github/workflows/causal_lm_cpp.yml b/.github/workflows/causal_lm_cpp.yml index a0216d777b..8b27b0484a 100644 --- a/.github/workflows/causal_lm_cpp.yml +++ b/.github/workflows/causal_lm_cpp.yml @@ -275,7 +275,7 @@ jobs: && python samples\python\greedy_causal_lm\lora.py .\TinyLlama\TinyLlama-1.1B-intermediate-step-1431k-3T\ adapter_model.safetensors "How to create a table with two columns, one of them has type float, another one has type int?" cpp-greedy_causal_lm-Chatglm3-6b: - runs-on: ubuntu-20.04-16-cores + runs-on: ubuntu-20.04-4-cores defaults: run: shell: bash @@ -312,7 +312,6 @@ jobs: samples/python/greedy_causal_lm/greedy_causal_lm.py ./chatglm3-6b/ "$( [!NOTE] >* Models should belong to the same family and have the same tokenizers. >* `optimum-cli` requires the `--task text-generation-with-past` argument for the `THUDM/chatglm3-6b` model ->* The **beam search** is not supported for `ChatGLM3-6B model`. - +>* The **beam search** is not supported for `THUDM/chatglm3-6b` model. ## Text 2 image models
From bad9759008d4abc74a73c7d8a7590634ddce62d8 Mon Sep 17 00:00:00 2001 From: Roy Date: Fri, 6 Dec 2024 12:12:23 +0530 Subject: [PATCH 28/29] tokenizers --- thirdparty/openvino_tokenizers | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/thirdparty/openvino_tokenizers b/thirdparty/openvino_tokenizers index 3fe6677f27..306dcd8dae 160000 --- a/thirdparty/openvino_tokenizers +++ b/thirdparty/openvino_tokenizers @@ -1 +1 @@ -Subproject commit 3fe6677f27c7f5eb149b959d5cd80dff76e9924b +Subproject commit 306dcd8daec36bbc680c50c68de1e954f42b0ab8 From 0ba2718b924bd1527923a681a010b903a3221f2e Mon Sep 17 00:00:00 2001 From: Roy Date: Fri, 6 Dec 2024 12:16:31 +0530 Subject: [PATCH 29/29] cpp-prompt_lookup_decoding_lm-ubuntu --- .github/workflows/causal_lm_cpp.yml | 9 +-------- 1 file changed, 1 insertion(+), 8 deletions(-) diff --git a/.github/workflows/causal_lm_cpp.yml b/.github/workflows/causal_lm_cpp.yml index 35a7eba698..8b490b0002 100644 --- a/.github/workflows/causal_lm_cpp.yml +++ b/.github/workflows/causal_lm_cpp.yml @@ -593,14 +593,7 @@ jobs: echo '-------------------------------Greedy Generated------------------------------------------------' ./build/samples/cpp/greedy_causal_lm/greedy_causal_lm ./chatglm3-6b/ "$(