Hi there, fairly basic but important question, just looking for confirmation that models are not affected by anything you give them at inference time. Learning is done at pre-training and fine-tuning stages, but a model is effectively read-only once you start using it to generate text. Do I understand this correctly?
As I recall from walking through the Fine-tune a pretrained model example, all your work is lost, unless you save the updated model. If you used a model for inference for a few hours and then saved it, would it be any different?
Just a little context, I want to use data in a RAG app that should not end up affecting the model in any way, and, say, end up showing up in later conversations. That is, unless I intentionally implement a “memory” mechanism (which is just passing the conversation history back to the model every time). So for me, this “forgetfulness” of the model is a feature I’m counting on. Yes, I’ve probably watched too many episodes of Westworld.
One last question, is this forgetfulness behavior generally considered a feature or a bug. In other words, is this a behavior that might change in the future? Models that learn on the fly?
Thanks so much.