HACKER Q&A
📣 JuanPosadas

Is anybody selling API access to HuggingFace models?


I've used the OpenAI and GCP chat-bison APIs with great success, but I'd like to put my money into more ethical companies with free-as-in-freedom-er models.

I don't want to fickle around with buying or renting GPUs or installing random LLM scripts that don't work 90% of the time.

I also don't want to pay $XX/month subscriptions, I want clean API billing like OpenAI and GCP.

What exists? Is anybody doing anything?


  👤 brucethemoose2 Accepted Answer ✓
This is not really an answer, but I really like the "AI horde" API setup.

https://aihorde.net/

Namely, you have a pool of gen AI volunteer workers constantly chewing through API requests. Host requests are prioritized, but the "downtime" of your AI worker instance is spent fulfilling other requests, which earns karma that can be used to prioritize requests from other hosts should you need more throughput.

Economically, this is way more efficient than self hosting or "renting" an instance that will be idle between requests, and you aren't paying a massive tax for a proprietary API.


👤 version_five
Sort of related, I'm surprised there are no "edge" GPUs available in the style of cloudflare workers or AWS lambda. I heard that cloudflare was going to do something like this but it never materialized. It would be great (as an alternative to API access) to be able to pay for a worker GPU that you only used when doing inference. I know there are technical challenges to that, I wonder of it's still too hard to make something responsive enough to be useful?

👤 smoldesu
Many of the "free as in freedom" models are not trained or licensed for commercial use. For example, selling access to a LLaMA API would be a direct violation of the license the model was provided with.