LocalLLaMA

Simple to use vLLM Docker Container for Qwen3.6 27b with Lorbus AutoRound INT4 quant and MTP speculative decoding – 118 tokens/second on 2x 3090s

submitted by /u/tedivm [link] [comments]