llama.cpp speculative checkpointing was merged
https://github.com/ggml-org/llama.cpp/pull/19493 Some prompts get a speedup, others don't (cases of low draft acceptance streak). Good working params depend on the task type and repetition patterns. For coding, I got some 0%~50% speedup with these …