- cross-posted to:
- localllama
- cross-posted to:
- localllama
cross-posted from: https://lemmy.world/post/19925986
https://huggingface.co/collections/Qwen/qwen25-66e81a666513e518adb90d9e
Qwen 2.5 0.5B, 1.5B, 3B, 7B, 14B, 32B, and 72B just came out, with some variants in some sizes just for math or coding, and base models too.
All Apache licensed, all 128K context, and the 128K seems legit (unlike Mistral).
And it’s pretty sick, with a tokenizer that’s more efficient than Mistral’s or Cohere’s and benchmark scores even better than llama 3.1 or mistral in similar sizes, especially with newer metrics like MMLU-Pro and GPQA.
I am running 34B locally, and it seems super smart!
As long as the benchmarks aren’t straight up lies/trained, this is massive, and just made a whole bunch of models obsolete.
Get usable quants here:
GGUF: https://huggingface.co/bartowski?search_models=qwen2.5
EXL2: https://huggingface.co/models?sort=modified&search=exl2+qwen2.5