Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

add_bos_token causes very unstable results for quantized llama3-70B #2676

Open
wenhuach21 opened this issue Feb 7, 2025 · 2 comments
Open
Labels
asking questions For asking for clarification / support on library usage.

Comments

@wenhuach21
Copy link

wenhuach21 commented Feb 7, 2025

https://huggingface.co/OPEA/Llama-3.3-70B-Instruct-int3-sym-inc
arc_easy WO BOS 0.2643 vs with BOS 0.8523

https://huggingface.co/OPEA/Llama-3.3-70B-Instruct-int2-sym-inc
mmlu WO BOS 0.7142 vs with BOS 0.7606
lambada_openai WO BOS 0.7013 vs with BOS 0.7413

is it possible to align with the 16 bits model directly in lm-eval?

@baberabb
Copy link
Contributor

baberabb commented Feb 7, 2025

Hi! Is this with the chat template? generally with a template, you don't have to include the bos token manually as the chat template assures it is included

@baberabb baberabb added the asking questions For asking for clarification / support on library usage. label Feb 7, 2025
@wenhuach21
Copy link
Author

wenhuach21 commented Feb 8, 2025

Hi! Is this with the chat template? generally with a template, you don't have to include the bos token manually as the chat template assures it is included

we used the cmd recommended in the homepage, I don't know whether it used chat template or not
With BOS
lm-eval --model hf --model_args pretrained=OPEA/Llama-3.3-70B-Instruct-int3-sym-inc,add_bos_token=True --tasks mmlu --batch_size 16

Without BOS

lm-eval --model hf --model_args pretrained=OPEA/Llama-3.3-70B-Instruct-int3-sym-inc --tasks mmlu --batch_size 16

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
asking questions For asking for clarification / support on library usage.
Projects
None yet
Development

No branches or pull requests

2 participants