pretrain topic

List pretrain repositories

What-I-Have-Read

163
Stars
16
Forks
Watchers

Paper Lists, Notes and Slides, Focus on NLP. For summarization, please refer to https://github.com/xcfcode/Summarization-Papers

nlp_chinese_corpus

9.2k
Stars
1.5k
Forks
Watchers

大规模中文自然语言处理语料 Large Scale Chinese Corpus for NLP

UniVL

330
Stars
54
Forks
Watchers

An official implementation for " UniVL: A Unified Video and Language Pre-Training Model for Multimodal Understanding and Generation"

BERT-CCPoem

136
Stars
19
Forks
Watchers

BERT-CCPoem is an BERT-based pre-trained model particularly for Chinese classical poetry

CLUECorpus2020

888
Stars
80
Forks
Watchers

Large-scale Pre-training Corpus for Chinese 100G 中文预训练语料

MatDGL

65
Stars
12
Forks
Watchers

MatDGL is a neural network package that allows researchers to train custom models for crystal modeling tasks. It aims to accelerate the research and application of material science.

RE-Context-or-Names

100
Stars
22
Forks
Watchers

Bert-based models(BERT, MTB, CP) for relation extraction.

albert-mongolian

18
Stars
2
Forks
Watchers

ALBERT trained on Mongolian text corpus

SparK

1.4k
Stars
80
Forks
Watchers

[ICLR'23 Spotlight🔥] The first successful BERT/MAE-style pretraining on any convolutional network; Pytorch impl. of "Designing BERT for Convolutional Networks: Sparse and Hierarchical Masked Modeling...

MFM

23
Stars
2
Forks
Watchers

code for paper "Masked Frequency Modeling for Self-Supervised Visual Pre-Training" (https://arxiv.org/pdf/2206.07706.pdf)