SkimCaffe icon indicating copy to clipboard operation
SkimCaffe copied to clipboard

ResNet-50-train-val-gesl.prototxt

Open newwhitecheng opened this issue 8 years ago • 6 comments

Hi, Can you provide models/resnet/ResNet-50-train-val-gesl.prototxt?

newwhitecheng avatar Jun 30 '17 03:06 newwhitecheng

Oops! I forgot to add those files. Thanks for pointing out. Now the necessary files to reproduce resnet-50 pruning results should be there. Please let me know if you encounter other issues.

jspark1105 avatar Jun 30 '17 16:06 jspark1105

I'm wondering why the sparse network has the same size as original one. Which are both 102Mb? have you ever try to sparsify and compress model? And When I test the speed on GPU, their speed performance are similar. Thanks!

newwhitecheng avatar Jul 13 '17 18:07 newwhitecheng

Hi, good question. Even if pruned model has lots of zeros we didn't take advantage of that when storing to files. This is because our main goal was reducing inference time, so the model stored in memory during inference is small. If model size stored in disk (or transferring over network) is important, we need to change the code to store the model in a different format taking advantage of the zeros. I'm not that familiar with hdf5 and so on, so I'd need some help here.

jspark1105 avatar Jul 13 '17 21:07 jspark1105

Regarding speed up in GPU, I didn't optimize for GPU inference speed using the sparsity. In Xeon and Xeon Phi, you should be able to see speedups.

jspark1105 avatar Jul 14 '17 05:07 jspark1105

Thank you for you reply. Another question which is not related to this thread. Is it possible to use it on ARM cpu?

newwhitecheng avatar Jul 14 '17 05:07 newwhitecheng

Technically, yes. ARM also should be able to take advantage of sparsity to speed up inference but you'd need to implement an optimized sparse convolution similar to the one for x86 in sconv.hpp .

jspark1105 avatar Jul 14 '17 05:07 jspark1105