Though the big question is whether learning should always happen offline or it would be useful to have models learn in production. Let me hear your thoughts.
For example, if a major event or discovery happened, the model would know about it once a critical mass of news stories and discussions had been generated online. You’d probably be looking at a few days before the content accumulated to the point where it would affect the model weights, which encode all human knowledge ever digitized, so some single news article in a training set of trillions of tokens is not enough.
If you want a long term memory of user interactions, long context and RAG seems to do the job nicely as a single fact can be pulled out of a context length of millions of tokens.