Your question doesn’t even make sense. The whole point of training on as large of a dataset as possible is so the model picks up on the general rules of text and doesn’t just overfit and regurgitate the training data.
If you have a model where you ask it for a book and it spits out the exact book you’ve trained a terrible model.