What are you using for inference if I can ask, just curiosity
Login to reply
Replies (2)
Which interface?
No for the text generation inference (the server that runs the llms), I was researching the different options available, some cool ones are huggingface tgi, llama.cpp, vllm