cross-posted from: https://lemmy.ml/post/45766694
Hey :) For a while now I use gpt-oss-20b on my home lab for lightweight coding tasks and some automation. I’m not so up to date with the current self-hosted LLMs and since the model I’m using was released at the beginning of August 2025 (From an LLM development perspective, it feels like an eternity to me) I just wanted to use the collective wisdom of lemmy to maybe replace my model with something better out there.
Edit:
Specs:
GPU: RTX 3060 (12GB vRAM)
RAM: 64 GB
gpt-oss-20b does not fit into the vRAM completely but it partially offloaded and is reasonably fast (enough for me)


I think people are sleeping on GLM.
Tried it out recently and I like the results a lot so far.
GLM4.5 and 4.7 was good already, now they released 5 and 5.1 https://github.com/zai-org/GLM-5
It says it’s for vibecoding but I use it like I would use chatgpt and it gives useable ansers to all of my varied questions. (ofc. you always have to check for correctness, even if it’s correct most of the time, which I do cause I’m paranoid)
I guess the only downside is how frigging huge it is.
Yep :D
I saw 5.1 came out however it required a data centre to run :X
Hoping to see if they release smaller models how they do