https://drive.google.com/file/d/1HHDR1nMYoHCz5FNrppv3oee9E4CWZZUL/view?usp=drive_link
pip install -r requirements.txt
-
Put the dataset in
LAMBADA/data; -
Set your openai api_key in
lambada_chatgpt.pylambada_002.pylambada_003.pyL6, your Claude userOAuthToken and channel_id intest_lambada.shL6, and your model-path intest_lambada.shL7-13; -
To get the evaluation scores of all models:
bash test_lambada.sh -
All generated outputs are saved in
LAMBADA/generation.
https://drive.google.com/file/d/1p2jG-h8NTUOCh-X_rIZHjV_wjzfGWDE5/view?usp=drive_link
pip install -r requirements.txt
-
Put the dataset in
WMT22/data; -
Set your openai api_key in
wmt_chatgpt.pywmt_002.pywmt_003.pyL6, your Claude userOAuthToken and channel_id intest_wmt.shL9, and your model-path intest_wmt.shL11-23; -
To get the evaluation scores of all models:
bash test_wmt.sh -
All generated outputs are saved in
WMT22/generation.
https://drive.google.com/file/d/1SI7dafnXvcp96nQzhgMgafOso0XUmGFo/view?usp=drive_link
pip install -r requirements.txt
-
Put the dataset in
XSum/data; -
Set your openai api_key in
xsum_chatgpt.pyxsum_002.pyxsum_003.pyL9, your Claude userOAuthToken and channel_id intest_xsum.shL13, and your model-path intest_xsum.shL11-23; -
To get the evaluation scores of all models:
bash test_xsum.sh -
All generated outputs are saved in
XSum/generation.
https://drive.google.com/file/d/1fy3LorG0TkHNK22rlAYV1yftgU4CAipx/view?usp=drive_link
pip install -r requirements.txt
-
Put the dataset in
HumanEval/data; -
Set your openai api_key in
model.pyL267util.pyL6, your Claude userOAuthToken and channel_id inmodel.pyL112-113, and your model-path intest_humaneval.shL10-22; -
To get the evaluation scores of all models:
bash test_humaneval.sh -
All generated outputs are saved in
HumanEval/generation.