https://github.com/ExtensityAI/symbolicai?tab=readme-ov-file...
I think what you’ll find is that some applications are very capable locally, like Whisper.
A lot of plugins expect to work with the llama.cpp family. Nowadays, that’s HuggingFace TGI: https://huggingface.co/blog/tgi-messages-api
So your application could speak OpenAI api, and you’d run HuggingFace TGI on your hardware for testing and comparison.