Pegah Salehi
Pegah Salehi
I have the same problem, but it was not solved by changing the version of torchvision? Did you find a solution? Thanks in advance
Thank you very much for responding so quickly I did that. but I have a same error again. ``` RuntimeError: Error(s) in loading state_dict for NeRFNetwork: size mismatch for audio_net.encoder_conv.0.weight:...
Both `wav2vec `and `deepspeech `methods are used in the files required for training. But only the `wav2vec `method is used during the test. What is the reason behind this?
Thanks a lot @ashawkey !
If `deepspeech `is to be used in tests, what changes need to be made? I change the default name and use the corresponding audio features. But it did not work!...
> Could you provide the full error log? ``` Traceback (most recent call last): File "C:\Users\---\anaconda3\envs\rad-nerf\lib\argparse.py", line 2484, in _get_value result = type_func(arg_string) ValueError: invalid literal for int() with base...
> It says ` argument --data_range: invalid int value: '{Pose_start}'`, what's the command line you are running? ``` %run test.py -O --torso \ --pose data/pose.json \ --data_range {Pose_start} {Pose_end} \...
Hi @ashawkey , Does the `model.pth `relate to `deepspeech` and `wav2vec` generated during training? According to my assumptions, the file named `ngp` is for `deepspeech`. A question. Don't we need...
Thanks a lot @ashawkey for your response! The question is whether the trained model will be generated for both (wav2wec and deepspeech) during training, or should it be trained separately...
The problem has been solved! [https://github.com/NVlabs/nvdiffrec/issues/35#issuecomment-1147749506](url)