Litespark Inference on Consumer CPUs: Custom SIMD Kernels for Ternary Neural Networks
arXiv:2605.06485v1 Announce Type: new
Abstract: Large language models (LLMs) have transformed artificial intelligence, but their computational requirements remain prohibitive for most users. Standard inference demands expensive datacenter GPUs or clou…