MMLU Performance #67
w013nad
started this conversation in
Show and tell
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I spent some time editing your MMLU test script to compare performance at different quantization levels. A full write-up may be found here.
https://www.reddit.com/r/LocalLLaMA/comments/16tgzzk/exllamav2_performance_with_different_quantization/
If you're interested, I can try to merge my script with yours (I'm unfamiliar with Git). Feel free to use this chart if it's useful.
Full results.
https://docs.google.com/spreadsheets/d/1MFmHDpqcf7CP_EYnwl1QsUP0KhS6jA8x1JYuwm3NH0U/edit?usp=sharing
Beta Was this translation helpful? Give feedback.
All reactions