I don't want to use GPT since the project will be using personal information to train/fine tune the models.
These models perform slightly better than GPT-3 under some tasks[2], but they're still far from achieving the results from GPT-3.5 and GPT-4. This becomes evident when you try to use them in the real world; they're not "good enough" for general use cases, unlike ChatGPT models. However, if you can restrict your use case to one particular domain, you can achieve pretty good results by further fine-tuning these models.
[0]: https://huggingface.co/google/flan-t5-xxl
[1]: https://huggingface.co/google/flan-ul2
[2]: https://paperswithcode.com/sota/multi-task-language-understa...
Not yet mentioned:
* Pythia https://github.com/EleutherAI/pythia
* GLM-130B https://github.com/THUDM/GLM-130B - see also ChatGLM-6B https://github.com/THUDM/ChatGLM-6B
* GPT-NeoX-20B https://huggingface.co/EleutherAI/gpt-neox-20b
* GeoV-9B https://github.com/geov-ai/geov
* BLOOM https://huggingface.co/bigscience/bloom and BLOOMZ https://huggingface.co/bigscience/bloomz
LLMs take so much engineering effort, research, and compute that it's unlikely there will be good open source alternatives in the near future. Right now your only real option is OpenAI (or maybe Anthropic) and that seems unlikely to change anytime soon.
The only reason we have LLAMA is because Meta threw us a bone. They might not do that again.
A GPT has no training until you give it materials. I do believe Google released the code for theirs ages ago. Even without source, you can run a GPT against your own data locally, or on a cloud service setup for that purpose.
This is how Bloomberg, for example, created a financial LLM. They used a GPT to train on their own financial data.
Source: https://help.openai.com/en/articles/5722486-how-your-data-is...
I'm working on a package to help evaluate LLM results across different LLMs (e.g., GPT3.5 vs. GPT4 vs. Dolly 2 vs...); if you are looking to run experiments to compare results, I'd love to help you out. You can email me at w (at) phaseai (dot) com.
https://www.cerebras.net/blog/cerebras-gpt-a-family-of-open-...
Commercial product sure can be built on top of LLAMA, it's GPL-3. Your models are your own; just patches, modifications, and code you link to LLMA itself will be governed by the GPL as well.
This is almost certainly what you want since this way you can use patches, fixes, and improvements others make to LLMA. You won't have to do all that work yourself, or necessarily wait for Facebook.
https://www.heise.de/news/Open-source-AI-LAION-proposes-to-o...
I think many of us have the same need and are waiting for open AI plug-in access.
Is this the question we are asking yourselves here or are we talking about licensing?
Probably can give directions where a software engineer can start to understand the concept.
I thought the opt series can be used in production