statikk wrote:
1. Could you please provide a baseline model inference tests on DirectX vs. CUDA?
Are you asking about a performance comparison between DirectX and CUDA? Performance depends on a number of different factors (the model being evaluated, input types, device hardware, graphics drivers, etc.) so results tend to be specific to a developer's unique scenario. However, the developers behind DirectX and the Windows AI stack (WinML, DirectML, and related technologies) work extremely closely with hardware vendors to ensure consistent results and performance across the broad range of Windows devices and GPUs.
statikk wrote:
2. Our code is all in `TensorFlow`, running inside `nvidia-docker`. Can you please elaborate on how hard would it be to port over the models?
Again, this depends on the details of your unique situation, but Microsoft does provide tools for porting models to ONNX, the Open Neural Network Exchange format. You can learn more about model conversion here: Convert ML models to ONNX with WinMLTools.