Ask HN: Is anybody using llama.cpp for production?
Considering the server version of llama.cpp for a commercial use case over bulkier options like vllm. But wondering if it's been battle-tested in production environments.
Considering the server version of llama.cpp for a commercial use case over bulkier options like vllm. But wondering if it's been battle-tested in production environments.
No, i think llama.cpp only using testing model
[dead]
[dead]