BiBERT
BiBERT copied to clipboard
This project is the official implementation of our accepted ICLR 2022 paper BiBERT: Accurate Fully Binarized BERT.
Results
3
BiBERT issues
Sort by
recently updated
recently updated
newest added
is there trick can solve the problem or it is a mistake?
Was Two Stage Knowledge Distillation used as in BinaryBERT in Table 7 (https://arxiv.org/pdf/2012.15701.pdf) to get these results?
您好,想问一个问题,论文里量化 1-1-1 模型的存储空间占用是13.4M,为什么我跑出来还是418M呢?怎样才能实现全1 bit位的模型,将模型最终的大小压缩到13.4M呢?