The Qwen3.5 models are still the best local models I’ve used, so I’m excited to see how this updated version performs.

  • TheCornCollector@piefed.zipOP
    link
    fedilink
    English
    arrow-up
    5
    ·
    4 hours ago

    I agree with the suggestion of the other commenters, just wanted to add that I personally run llama.cpp directly with the build in llama-server. For a single-user server this seems to work great and is almost always at the forefront of model support.