If not, why not?
EDIT:
Put another way,
1. If initial training runs some training loop 10 trillion times to modify a model by a lot
2. And if fine-tuning runs some training loop 10 thousand times to modify that pretrained model by a bit more
3. Can LLMs be architectured to take human feedback as input to run some training loop 10 times to nudge the actual model some more?
But you could create an LLM for which it wouldn't be the case.