Running large language models (LLMs) on your local machine has become increasingly popular, offering privacy, offline access, and customization. Ollama is a ...
I got both IK and Croco built and running on Vulkan, but they both seem abysmally slow compared to Kobold, whether I run one of the GGUFs I already had or a trellis quant I just made. It’s using the GPU, so not sure why. I’ll keep an eye on the projects and try them again in the future.
I got both IK and Croco built and running on Vulkan, but they both seem abysmally slow compared to Kobold, whether I run one of the GGUFs I already had or a trellis quant I just made. It’s using the GPU, so not sure why. I’ll keep an eye on the projects and try them again in the future.