From one session to the next the only thing that changes is the previous trained model
There is one other thing I see that does change - the validation set. Each training session will have a new validation set and some new training data. Each time you run NetTrain 20% of your data is randomly selected to be the validation set. So there's new data in each training session.
Under these conditions, I would expect somewhat of a zigzag pattern. I'm not sure however if that justifies what we see in your example.
Or worse, maybe each round is somewhat ..overfitting.. to the validation set? I'm not sure what the correct word choice is there.