No for the text generation inference (the server that runs the llms), I was researching the different options available, some cool ones are huggingface tgi, llama.cpp, vllm

Reply to this note

Please Login to reply.

Discussion

No replies yet.