convert DeepFaceLab XSegNet's *.npy weights to onnx file. Inference time is optimized from 500ms to 90ms.
- put DeepFaceLab XSegNet weights to weights folder, such as
weights/XSeg_256.npy. Set XSegNet class default augmentexport_weights=True.- Then run
python test_seg.pyto generate tensorflow SavedModel format checkpoint file tosaved_modeldirectory. - install tf2onnx library, run
pip install tf2onnx. - convert model to onnx file,
python -m tf2onnx.convert --saved-model ./saved_model/ --output xseg.onnx --tag serve. - (optinal) install onnxsim
pip install onnxsimand runonnxsim ./xseg.onnx ./xseg.sim.onnx.
see test_seg_onnx.py.
Because of Conv2d_transpose requires asymmetric padding which the CUDA EP currently does not support #11312, XSegNet OnnxRuntime Conv2d_transpose layer does not support CudaExcuation.