I am trying to understand the purpose of a 3rd split in the form of a validation dataset. I am not necessarily talking about cross-validation here.
In the scenario below, it would appear that the model is overfit to the training dataset.
Train dataset {acc: 97%, loss: 0.07}
Test dataset {acc: 90%, loss: 8.02}
However, in this scenario it appears much more balanced.
Train dataset {acc: 95%, loss: 1.14}
Test dataset {acc: 93%, loss: 1.83}
Do I need validation data if my train and test accuracy/loss is consistent? Is the purpose of setting a validation split of 10% to ensure this kind of balance before evaluating the model on the training set? What does it prove?
model.evaluate()
after every training when tuning? – Kermit Jun 16 '20 at 01:49