| Code: https://github.com/fairydreaming/llama.cpp/tree/deepseek-dsa Supported GGUFs (Q4_K_M ~ 404GB, Q8_0 ~ 714GB):
Chat template to use: If you experience OOM errors in CUDA Let me know if you encounter any problems. [link] [comments] |
| Code: https://github.com/fairydreaming/llama.cpp/tree/deepseek-dsa Supported GGUFs (Q4_K_M ~ 404GB, Q8_0 ~ 714GB):
Chat template to use: If you experience OOM errors in CUDA Let me know if you encounter any problems. [link] [comments] |