You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I'm especially interested in the recently-introduced constitutional-ai tuning in this blog post. I've found the open-source SFT model and DPO model on huggingface. However, when I tried to launch the MT-Bench test with them, the returned results are significantly worse than those reported in the blog post, according to the figure below (which I've copied over here for reference):
The MT-Bench score I've collected are 5.33 / 6.39 for the SFT / DPO model respectively, where the reference figure shows (approximately) 6.5 / 7.2.
The text was updated successfully, but these errors were encountered:
Hi, thanks for your great work!
I'm especially interested in the recently-introduced constitutional-ai tuning in this blog post. I've found the open-source SFT model and DPO model on huggingface. However, when I tried to launch the MT-Bench test with them, the returned results are significantly worse than those reported in the blog post, according to the figure below (which I've copied over here for reference):
The MT-Bench score I've collected are 5.33 / 6.39 for the SFT / DPO model respectively, where the reference figure shows (approximately) 6.5 / 7.2.
The text was updated successfully, but these errors were encountered: