Gemma4-31B worked in an iterative-correction loop (with a long-term memory bank) for 2 hours to solve a problem that baseline GPT-5.4-Pro couldn’t
submitted by /u/Ryoiki-Tokuiten [link] [comments]
submitted by /u/Ryoiki-Tokuiten [link] [comments]
Hey everyone, I’ve been looking into different platforms to access various AI models without breaking the bank, and I keep coming back to HuggingChat. It gives free web access to top-tier open-weight models without needing a $20/month subscription. Giv…
Found it on a USB drive in the parking lot. Should be interesting. Seriously tho, props to this guy and his cool Hermes Agent skins library here: https://github.com/joeynyc/hermes-skins submitted by /u/Porespellar [link] [co…
tl;dr: Fixes KV-cache rotation for hybrid-attention models like Gemma 4 (Not actually TurboQuant, but you can call it TurboQuant if that makes you feel better) submitted by /u/jacek2023 [link] [comments]
I lead a reserach lab at a university hospital and spent the last weeks configuring our internal LLM server. I put a lot of thought into the server config, software stack and model. Now I am at a point where I am happy, it actually holds up under load …
Ok so two things happened this week that made me appreciate my local setup way more tried to cancel cursor ($200/mo ultra plan) and they instantly threw 50% off at me before I could even confirm. no survey, no exit flow, just straight to "pl…
Here is the HF https://huggingface.co/zai-org/GLM-5.1-FP8 submitted by /u/dev_is_active [link] [comments]