https://drive.google.com/file/d/1HHDR1nMYoHCz5FNrppv3oee9E4CWZZUL/view?usp=drive_link
pip install -r requirements.txt
-
Put the dataset in
LAMBADA/data
; -
Set your openai api_key in
lambada_chatgpt.py
lambada_002.py
lambada_003.py
L6, your Claude userOAuthToken and channel_id intest_lambada.sh
L6, and your model-path intest_lambada.sh
L7-13; -
To get the evaluation scores of all models:
bash test_lambada.sh
-
All generated outputs are saved in
LAMBADA/generation
.
https://drive.google.com/file/d/1p2jG-h8NTUOCh-X_rIZHjV_wjzfGWDE5/view?usp=drive_link
pip install -r requirements.txt
-
Put the dataset in
WMT22/data
; -
Set your openai api_key in
wmt_chatgpt.py
wmt_002.py
wmt_003.py
L6, your Claude userOAuthToken and channel_id intest_wmt.sh
L9, and your model-path intest_wmt.sh
L11-23; -
To get the evaluation scores of all models:
bash test_wmt.sh
-
All generated outputs are saved in
WMT22/generation
.
https://drive.google.com/file/d/1SI7dafnXvcp96nQzhgMgafOso0XUmGFo/view?usp=drive_link
pip install -r requirements.txt
-
Put the dataset in
XSum/data
; -
Set your openai api_key in
xsum_chatgpt.py
xsum_002.py
xsum_003.py
L9, your Claude userOAuthToken and channel_id intest_xsum.sh
L13, and your model-path intest_xsum.sh
L11-23; -
To get the evaluation scores of all models:
bash test_xsum.sh
-
All generated outputs are saved in
XSum/generation
.
https://drive.google.com/file/d/1fy3LorG0TkHNK22rlAYV1yftgU4CAipx/view?usp=drive_link
pip install -r requirements.txt
-
Put the dataset in
HumanEval/data
; -
Set your openai api_key in
model.py
L267util.py
L6, your Claude userOAuthToken and channel_id inmodel.py
L112-113, and your model-path intest_humaneval.sh
L10-22; -
To get the evaluation scores of all models:
bash test_humaneval.sh
-
All generated outputs are saved in
HumanEval/generation
.