I finally sat down and did the math on my Cloud LLM bills… and I’m moving almost everything to a 4090. [R]
I used to be all-in on cloud APIs. For any side project, I’d just grab an OpenAI or Anthropic key and not think twice. It was convenient. No worrying about VRAM, super fast responses, and I could spin something up in minutes. But that “pay-as-you-go” c…