Solved – Can we say that k fold cross validation reduces bias and variance as compared to train-validation-test split procedure

biascross-validationvariance

I have to compare the two strategies to choose optimal parameters of my NN. I have the following reasoning. Please tell if it is correct!

In train-validation-test split case, the validation dataset is being used to select model hyper-parameters and not to train the model hence reducing the training data might lead to higher bias. Also the validation error deoends a lot on which data points end up in the validation set and hence different validation sets might give different optimal parameters i.e. the evaluation results in higher variance.
In k fold cv , which is a more progressive procedure, each subset and hence every data point is used for validation exactly once. Since the RMSE is averaged over k subsets, the evaluation is less sensitive to the partitioning of data and variance of the resulting estimate is significantly reduced. Also since all the data points are used for training bias is also reduced.

Best Answer

While your argumentation why cross validation (or resampling validation in general) is better than a single random split is valid, for the optimization you need nested or double cross validation. That is the analogue to train-valiation-test, i.e. 2 splits producing 3 subsets of your data.