Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add Deepseek-R1-Distill-Qwen-32B results #3278

Open
wants to merge 2 commits into
base: main
Choose a base branch
from

Conversation

gcp
Copy link

@gcp gcp commented Feb 17, 2025

At 15.1% this is reasonable, much better than the LLama-70B Distill. This was with Q4(KL) and a 32k Q8 context, which all fits on a 24GB GPU and is fast to run. So it's also possible to use 64k context on this model with a bit more quantization on the context.

I might try edit-format: whole and maybe Q5_K_S with a quantized or smaller context to see if that is even better.

@CLAassistant
Copy link

CLAassistant commented Feb 17, 2025

CLA assistant check
All committers have signed the CLA.

@gcp
Copy link
Author

gcp commented Feb 18, 2025

A run with a Q5KS quant scored a tad higher, 16% exactly: gcp@1b054b0

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants