

I just wanted to test if it was viable to run larger MoE LLMs on CPU, e.g. Qwen3-next-80B-A3B… Even if I got acceptable generation speeds I’d probably get bored with it after a few hours, as with other local models. Had I got it for €700 it was pretty low value for money anyway, since my current RAM is enough for everything else I use the computer for. On the positive side, I can put that money towards a Steam Frame instead.


https://github.com/resemble-ai/chatterbox is pretty good, and has both TTS and voice cloning. Main disadvantage for me was that even if the cloning gives a consistent voice, the generated samples can get random accents.
https://huggingface.co/zai-org/GLM-TTS also seemed pretty promising, but I haven’t had time to test it yet.