DeepRec
DeepRec copied to clipboard
DeepRec is a high-performance recommendation deep learning framework based on TensorFlow. It is hosted in incubation in LF AI & Data Foundation.
Introduction
DeepRec is a recommendation engine based on TensorFlow 1.15, Intel-TensorFlow and NVIDIA-TensorFlow.
Background
Sparse model is a type of deep learning model that accounts for a relatively high proportion of discrete feature calculation logic in the model structure. Discrete features are usually expressed as non-numeric features that cannot be directly processed by algorithms such as id, tag, text, and phrases. They are widely used in high-value businesses such as search, advertising, and recommendation.
DeepRec has been deeply cultivated since 2016, which supports core businesses such as Taobao Search, recommendation and advertising. It precipitates a list of features on basic frameworks and has excellent performance in sparse models training. Facing a wide variety of external needs and the environment of deep learning framework embracing open source, DeepeRec open source is conducive to establishing standardized interfaces, cultivating user habits, greatly reducing the cost of external customers working on cloud and establishing the brand value.
Key Features
DeepRec has super large-scale distributed training capability, supporting model training of trillion samples and 100 billion Embedding Processing. For sparse model scenarios, in-depth performance optimization has been conducted across CPU and GPU platform. It contains 3 kinds of features to improve usability and performance for super-scale scenarios.
Sparse Functions
- Embedding Variable.
- Dynamic Dimension Embedding Variable.
- Adaptive Embedding Variable.
- Multiple Hash Embedding Variable.
- Multi-tier Hybrid Embedding Storage
Performance Optimization
- Distributed Training Framework Optimization, such as grpc+seastar, FuseRecv, StarServer, HybridBackend etc.
- Runtime Optimization, such as CPU memory allocator (PRMalloc), GPU memory allocator, Cost based and critical path first Executor etc.
- Operator level optimization, such as BF16 mixed precision optimization, sparse operator optimization and EmbeddingVariable on PMEM and GPU, new hardware feature enabling, etc.
- Graph level optimization, such as AutoGraphFusion, SmartStage, AutoPipeline, StrutureFeature, MicroBatch etc.
- Compilation optimization, support BladeDISC, XLA etc.
Deploy and Serving
- Incremental model loading and exporting.
- Super-scale sparse model distributed serving.
- Multi-tier hybrid storage and multi backend supported.
- Online deep learning with low latency.
- High performance processor with SessionGroup supported.
Installation
Prepare for installation
CPU Platform
alideeprec/deeprec-build:deeprec-dev-cpu-py36-ubuntu18.04
GPU Platform
alideeprec/deeprec-build:deeprec-dev-gpu-py36-cu110-ubuntu18.04
How to Build
Configure
$ ./configure
Compile for CPU and GPU defaultly
$ bazel build -c opt --config=opt //tensorflow/tools/pip_package:build_pip_package
Compile for CPU and GPU: ABI=0
$ bazel build --cxxopt="-D_GLIBCXX_USE_CXX11_ABI=0" --host_cxxopt="-D_GLIBCXX_USE_CXX11_ABI=0" -c opt --config=opt //tensorflow/tools/pip_package:build_pip_package
Compile for CPU optimization: oneDNN + Unified Eigen Thread pool
$ bazel build -c opt --config=opt --config=mkl_threadpool //tensorflow/tools/pip_package:build_pip_package
Compile for CPU optimization and ABI=0
$ bazel build --cxxopt="-D_GLIBCXX_USE_CXX11_ABI=0" --host_cxxopt="-D_GLIBCXX_USE_CXX11_ABI=0" -c opt --config=opt --config=mkl_threadpool //tensorflow/tools/pip_package:build_pip_package
Create whl package
$ ./bazel-bin/tensorflow/tools/pip_package/build_pip_package /tmp/tensorflow_pkg
Install whl package
$ pip3 install /tmp/tensorflow_pkg/tensorflow-1.15.5+${version}-cp36-cp36m-linux_x86_64.whl
Latest Release Images
Image for CPU
alideeprec/deeprec-release:deeprec2206-cpu-py36-ubuntu18.04
Image for GPU CUDA11.0
alideeprec/deeprec-release:deeprec2206-gpu-py36-cu110-ubuntu18.04
Continuous Build Status
Official Build
Build Type | Status |
---|---|
Linux CPU | |
Linux GPU | |
Linux CPU Serving |
Official Unit Tests
Unit Test Type | Status |
---|---|
Linux CPU C | |
Linux CPU CC | |
Linux CPU Contrib | |
Linux CPU Core | |
Linux CPU Examples | |
Linux CPU Java | |
Linux CPU JS | |
Linux CPU Python | |
Linux CPU Stream Executor | |
Linux GPU C | |
Linux GPU CC | |
Linux GPU Contrib | |
Linux GPU Core | |
Linux GPU Examples | |
Linux GPU Java | |
Linux GPU JS | |
Linux GPU Python | |
Linux GPU Stream Executor | |
Linux CPU Serving UT |
User Document (Chinese)
Contact Us
Join the Official Discussion Group on DingTalk

License
Apache License 2.0