LocalLLaMA

What is the most capable model you can actually run on a single consumer GPU?

Not "what benchmarks the best" or "what has the most parameters." I mean in your actual daily use. If you had to pick one model to run locally on something like a 4090 or 3090 and use for real work, what is your go-to? I am curious …