I tried dumping the inventory information and basic BOM content as a system message for ChatGPT4-o model using their platform playground, and asking questions like “we are making 1000 power banks this month, so which components should I pre-order so that we don’t run out of them?”. This works as expected, but it took me a while to realize that each query used more than 30K tokens! That’s a quick way to lose money.
I am looking for a solution from someone who trained/tuned a decent LLM on custom data, and I’m pretty sure a lot of other small business owners are looking for something like this. Thank you!
https://huggingface.co/docs/transformers/en/model_doc/t5
https://medium.com/nlplanet/a-full-guide-to-finetuning-t5-fo...
Specifically you can show a T5 model input and output texts and it will try to learn the transformation between them. People tell me T5 models are relatively easy to train and they perform well on many tasks.
Note another approach to your problem is RAG
https://www.promptingguide.ai/techniques/rag
If you have some specific documentation on your topic you could use the embedding to find some text that is relevant to the query. In fact this stacks great with the fine-tuning because you could train the model to, given a question and a relevant document, give an answer. T5 is good at that kind of basically summarization task.