What are you using for inference if I can ask, just curiosity

Reply to this note

Please Login to reply.

Discussion

Which interface?

No for the text generation inference (the server that runs the llms), I was researching the different options available, some cool ones are huggingface tgi, llama.cpp, vllm