audio2photoreal
audio2photoreal copied to clipboard
gpu
(a2p_env) C:\Users\kit\audio2photoreal>python -m train.train_diffusion --save_dir checkpoints/diffusion/c1_face_test --data_root ./dataset/RLW104/ --batch_size 4 --dataset social --data_format face --layers 8 --heads 8 --timestep_respacing '' --max_seq_length 600
using 0 gpus
Traceback (most recent call last):
File "C:\Users\kit\miniconda3\envs\a2p_env\lib\runpy.py", line 197, in _run_module_as_main
return _run_code(code, main_globals, None,
File "C:\Users\kit\miniconda3\envs\a2p_env\lib\runpy.py", line 87, in _run_code
exec(code, run_globals)
File "C:\Users\kit\audio2photoreal\train\train_diffusion.py", line 83, in
(a2p_env) C:\Users\kit\audio2photoreal>python -m train.train_diffusion --save_dir checkpoints/diffusion/c1_face_test --data_root ./dataset/RLW104/ --batch_size 4 --dataset social --data_format face --layers 8 --heads 8 --timestep_respacing '' --max_seq_length 600
using 0 gpus
creating data loader...
[dataset.py] training face only model
['[dataset.py] sequences of 600']
C:\Users\kit\miniconda3\envs\a2p_env\lib\site-packages\numpy\core\fromnumeric.py:43: VisibleDeprecationWarning: Creating an ndarray from ragged nested sequences (which is a list-or-tuple of lists-or-tuples-or ndarrays with different lengths or shapes) is deprecated. If you meant to do this, you must specify 'dtype=object' when creating the ndarray.
result = getattr(asarray(obj), method)(*args, **kwds)
C:\Users\kit\miniconda3\envs\a2p_env\lib\site-packages\numpy\core\fromnumeric.py:43: FutureWarning: The input object of type 'Tensor' is an array-like implementing one of the corresponding protocols (__array__
, __array_interface__
or __array_struct__
); but not a sequence (or 0-D). In the future, this object will be coerced as if it was first converted using np.array(obj)
. To retain the old behaviour, you have to either modify the type 'Tensor', or assign to an empty array created with np.empty(correct_shape, dtype=object)
.
result = getattr(asarray(obj), method)(*args, **kwds)
[dataset.py] loading from... ./dataset/RLW104/data_stats.pth
[dataset.py] train | 18 sequences ((8989, 256)) | total len 160523
creating logger...
creating model and diffusion...
Traceback (most recent call last):
File "C:\Users\kit\miniconda3\envs\a2p_env\lib\runpy.py", line 197, in _run_module_as_main
return _run_code(code, main_globals, None,
File "C:\Users\kit\miniconda3\envs\a2p_env\lib\runpy.py", line 87, in _run_code
exec(code, run_globals)
File "C:\Users\kit\audio2photoreal\train\train_diffusion.py", line 83, in
Hmm it seems that you are attempting to load the model onto a non-gpu supported device (just the cpu). And i see that when loading the model, the map_location is already specified re: cp = torch.load(cp_path, map_location=torch.device(self.device))
(post)
Could you double check to see if self.device == 'cpu'
in your case please?
Closing for now due to inactivity, but please feel free to reopen as needed!