Results 1 issues of WX

### Describe the issue The model weight is quantified per channel: - weight_scale.shape=[64,], - zero_point.shape=[64]. When using onnxruntime-train to do QAT, the following error is reported. Does onnxruntime-train support per_channel...

quantization
training
ep:CUDA
stale