Tianlei Wu

Results 108 comments of Tianlei Wu

/azp run iOS CI Pipeline,ONNX Runtime React Native CI Pipeline,CoreML CI Pipeline,Linux DNNL CI Pipeline,Linux MIGraphX CI Pipeline,Linux ROCm CI Pipeline

/azp run Windows CPU CI Pipeline, ONNX Runtime Web CI Pipeline, Windows GPU CUDA CI Pipeline, Linux ROCm CI Pipeline, Linux OpenVINO CI Pipeline

/azp run Windows CPU CI Pipeline, ONNX Runtime Web CI Pipeline, Windows GPU CUDA CI Pipeline, Linux ROCm CI Pipeline, Linux OpenVINO CI Pipeline

/azp run Windows ARM64 QNN CI Pipeline,Windows x64 QNN CI Pipeline,Windows CPU CI Pipeline,Windows GPU CUDA CI Pipeline,Windows GPU DML CI Pipeline,Windows GPU Doc Gen CI Pipeline,Windows GPU TensorRT CI...

/azp run Linux GPU CI Pipeline,Linux GPU TensorRT CI Pipeline,Linux OpenVINO CI Pipeline,Linux QNN CI Pipeline,MacOS CI Pipeline,orttraining-amd-gpu-ci-pipeline,orttraining-linux-gpu-ci-pipeline,orttraining-ortmodule-distributed,onnxruntime-binary-size-checks-ci-pipeline,Big Models

@phamkhactu, sorry for late response. I might miss the thread during vacation. The topological error means float16 conversion script might have a bug and cannot deal with the model. Did...

Note that there are cudnn 8 and cudnn 9. The command to install onnxruntime for cuda 11 and 12 are different. See the following for detail: https://onnxruntime.ai/docs/execution-providers/CUDA-ExecutionProvider.html#requirements onnxruntime-gpu for cuda...

@JiayuanWang-JW, Where operator is supported by onnxruntime and it is common in transformer model to use Where for padding mask. It is likely that your onnx model is not valid....

@JiayuanWang-JW, you can save the model like the following, and attach the model.onnx only (not need model.onnx.data). ``` onnx.save_model(model, "model.onnx", save_as_external_data=True, all_tensors_to_one_file=True, location="model.onnx.data", size_threshold=128, convert_attribute=False) ```

For CUDA, Where operator only supports uint8, int32, int64, float, double, float16, bfloat16. It seems that the model uses bool as X type. It is valid for ONNX, however not...