I'm imagining making scripts/pipelines/etc which try to transform or write code, and I don't see how this can scale. Once LLMs are "good" I could see using it for small one-off tasks, but could you really make a _tool_ that you have to _maintain_ by gluing prompts together? What happens when the LLM gets "upgraded" and your finiky language now requires _different_ finiky language.
It feels like depending on a service who's API contract not only isn't upheld, but can never be upheld.
The whole AI field is revolving around tweaking and fine tuning, so I don’t think the finicky thing will go away any soon. Looking back at the whole automation/expert/AI history, we can see that specialization is always a more promising approach and I believe it’s the way to go for current AI development as well; not replacing humans but filling in the niche and automate the long and boring stuffs.