HACKER Q&A
📣 nancyp

Will it be a viable business model to offer llama.cpp as a service?


Wrapped in a nicer UI for b2b SaaS model.


  👤 fbnbr Accepted Answer ✓
Why would you need it as cpp? Do you mean just llama or specifically a version small enough running on edge?

👤 throw03172019
Do you mean just a non-fine tuned chat bot based on llama 70b? Probably not.

👤 quickthrower2
I suspect no.

👤 version_five
Short answer, possibly, especially if it was say part of a rag system or some other architecture like that. There is room for more. Nothing particularly special about llama.cpp as the llm back end though. It's optimized for running on lower-end hardware which matters less if you're serving models as a service. But it has many strengths.

Llama.cpp / ggml is the open core of ggml.ai founded by GG and funded by the guy from github, so they have some monetization plan for it.