caffe-escoin
caffe-escoin copied to clipboard
Escoin: Efficient Sparse Convolutional Neural Network Inference on GPUs
Escoin
Copyright 2018 Xuhao Chen, National University of Defense Technology
This is a Caffe branch for training sparse CNN on GPUs (tested with AlexNet, GoogLeNet-v1, and Resnet-50). Please let us know if you're interested in this experimental feature. More details are described in the following paper:
https://arxiv.org/pdf/1802.10280, Escort: Efficient Sparse Convolutional Neural Networks on GPUs, Xuhao Chen
For datasets, compilation and exection instructions, please got to SkimCaffe.
Contact
Caffe
Caffe is a deep learning framework made with expression, speed, and modularity in mind. It is developed by Berkeley AI Research (BAIR)/The Berkeley Vision and Learning Center (BVLC) and community contributors.
Check out the project site for all the details like
- DIY Deep Learning for Vision with Caffe
- Tutorial Documentation
- BAIR reference models and the community model zoo
- Installation instructions
and step-by-step examples.
Custom distributions
- Intel Caffe (Optimized for CPU and support for multi-node), in particular Xeon processors (HSW, BDW, SKX, Xeon Phi).
- OpenCL Caffe e.g. for AMD or Intel devices.
- Windows Caffe
Community
Please join the caffe-users group or gitter chat to ask questions and talk about methods and models. Framework development discussions and thorough bug reports are collected on Issues.
Happy brewing!
License and Citation
Caffe is released under the BSD 2-Clause license. The BAIR/BVLC reference models are released for unrestricted use.
Please cite Caffe in your publications if it helps your research:
@article{jia2014caffe,
Author = {Jia, Yangqing and Shelhamer, Evan and Donahue, Jeff and Karayev, Sergey and Long, Jonathan and Girshick, Ross and Guadarrama, Sergio and Darrell, Trevor},
Journal = {arXiv preprint arXiv:1408.5093},
Title = {Caffe: Convolutional Architecture for Fast Feature Embedding},
Year = {2014}
}