Hey, has anyone here used Qwen3.5-27B-NVFP4-GGUF with llama.cpp yet?

Hey, has anyone here used Qwen3.5-27B-NVFP4-GGUF with llama.cpp yet?

Hey!

I was wondering if anyone of you have used Qwen3.5-27B-NVFP4-GGUF on RTX5090 on llama.cpp? I have downloaded and tested today Freenixi/AxionML-Qwen3.5-27B-NVFP4-GGUF and it's quite impressive (quality of answers and deffinatelly beter in non-english langauges) Also what was your speed on llama.cpp? Just asking out of curiosity. Please share your experience. Thanks!

https://preview.redd.it/3r5f7r4ojevg1.png?width=4917&format=png&auto=webp&s=56489c69c0bfdee794aad6f909ee7679caf20cb3

submitted by /u/mossy_troll_84
[link] [comments]

Leave a Comment

Your email address will not be published. Required fields are marked *

Scroll to Top