opencode with gemma 26B

opencode with gemma 26B

I was testing OpenCode and Roo Code with Gemma 26B on llama.cpp yesterday for about 10 hours.

I was able to make progress on my project, both solutions work. But:

OpenCode is kind of fucked up at the moment, because of that there is often long prompt processing.. Roo Code works correctly, but it has different issues (thinking takes longer, probably OpenCode has better prompts).

The problem with OpenCode looks unsolvable on the llama.cpp side. I need to test it with other engines to confirm that, and then I will probably have to fix it on the OpenCode side. Maybe improving Roo Code’s prompts would be a better choice?

My current command (after lots of experimenting) is:

llama-server -c 200000 -m /mnt/models1/Google/gemma-4-26B-A4B-it-UD-Q8_K_XL.gguf --host 0.0.0.0 --jinja --temp 0.7 --top-p 0.95 --top-k 64 --repeat-penalty 1.15 --cache-ram 20000 --ctx-checkpoints 20 --checkpoint-every-n-tokens 16000 -b 8192 
submitted by /u/jacek2023
[link] [comments]

Leave a Comment

Your email address will not be published. Required fields are marked *

Scroll to Top