MobiRnn
MobiRnn copied to clipboard
Efficient LSTM parallelization on smartphone GPU
MobiRnn on Android
Intro
This repo is for running LSTM model on mobile devices. Currently we support the following mode:
- [x] Plain CPU (Java)
- [x] TensorFlow CPU (Java)
- [x] Native CPU (C)
- [x] Eigen CPU (C++)
- [x] GPU (RenderScript)
- [ ] GPU (OpenCL)
- [ ] GPU (Vulkan)
Usage
Just run ./gradlew iR to install MobiRNN on your connected phone.
You can run ./gradlew pu to generate apk file or simply download the blob/mobile-release.apk in the repo.