Tianlei Wu
Tianlei Wu
/azp run iOS CI Pipeline,ONNX Runtime React Native CI Pipeline,CoreML CI Pipeline,Linux DNNL CI Pipeline,Linux MIGraphX CI Pipeline,Linux ROCm CI Pipeline
/azp run Windows CPU CI Pipeline, ONNX Runtime Web CI Pipeline, Windows GPU CUDA CI Pipeline, Linux ROCm CI Pipeline, Linux OpenVINO CI Pipeline
/azp run Windows CPU CI Pipeline, ONNX Runtime Web CI Pipeline, Windows GPU CUDA CI Pipeline, Linux ROCm CI Pipeline, Linux OpenVINO CI Pipeline
/azp run Windows ARM64 QNN CI Pipeline,Windows x64 QNN CI Pipeline,Windows CPU CI Pipeline,Windows GPU CUDA CI Pipeline,Windows GPU DML CI Pipeline,Windows GPU Doc Gen CI Pipeline,Windows GPU TensorRT CI...
/azp run Linux GPU CI Pipeline,Linux GPU TensorRT CI Pipeline,Linux OpenVINO CI Pipeline,Linux QNN CI Pipeline,MacOS CI Pipeline,orttraining-amd-gpu-ci-pipeline,orttraining-linux-gpu-ci-pipeline,orttraining-ortmodule-distributed,onnxruntime-binary-size-checks-ci-pipeline,Big Models
@phamkhactu, sorry for late response. I might miss the thread during vacation. The topological error means float16 conversion script might have a bug and cannot deal with the model. Did...
Note that there are cudnn 8 and cudnn 9. The command to install onnxruntime for cuda 11 and 12 are different. See the following for detail: https://onnxruntime.ai/docs/execution-providers/CUDA-ExecutionProvider.html#requirements onnxruntime-gpu for cuda...
@JiayuanWang-JW, Where operator is supported by onnxruntime and it is common in transformer model to use Where for padding mask. It is likely that your onnx model is not valid....
@JiayuanWang-JW, you can save the model like the following, and attach the model.onnx only (not need model.onnx.data). ``` onnx.save_model(model, "model.onnx", save_as_external_data=True, all_tensors_to_one_file=True, location="model.onnx.data", size_threshold=128, convert_attribute=False) ```
For CUDA, Where operator only supports uint8, int32, int64, float, double, float16, bfloat16. It seems that the model uses bool as X type. It is valid for ONNX, however not...