I found ollama too awkward to set up for models not pre-provisioned by the ollama team. I turned to #koboldcpp to run any model I want with single-binary installation (text gen, image gen, TTS) and super simple configuration. And it comes with webuis for text and stable diffusion; or use any ui that supports koboldcpp-api or openai-api or automatc1111-api. CUDA, ROCM, Vulkan, CPU-only all supported.
Login to reply
Replies (1)
I'll check it out. Thanks!