As I run the exact same data science GPU code on identical hardware software configuration except for the change from Mathematica 11.2 to Mathematica 11.3 and see my neural network performance go from 295 seconds on 11.2 to 2038 seconds on 11.3.
When training the net, could you set the BatchSize
and MaxTrainingRounds
options to the same value in both versions, and then report the training time again? (we changed the heuristic for choosing a value of MaxTrainingRounds
which sometimes uses more training rounds for extra accuracy. And: the inputs/s
is a better indicator of speed than the total training time as it doesn't depend on MaxTrainingRounds
). If there is still a speed difference, could you give information about the GPU you are using, the output of $Version
and if possible the training script to reproduce this training?
And then I see that Mathematica 11.3 still does not support current XCode LLVM/GCC compiler or NVIDIA for CUDA tools
I am not sure exactly what you mean by this. Could you expand a bit about which functionality you are referring to that doesn't support "current XCode LLVM/GCC compiler or NVIDIA"?