Seq2CNN
Seq2CNN copied to clipboard
Word Embedding Annealing Using Sequence-to-sequence Model
Word Embedding Annealing Using Sequence-to-sequence Model
We propose a new technique to improve text classification accuracy by annealing the word embedding layer using Seq2seq Neural Networks. Our model Sequence-to-convolution Neural Networks(Seq2CNN) consists of two blocks: Sequential Block that summarizes input texts and Convolution Block that classifies the original text to a label.
We also present Gradual Weight Shift(GWS) method that stabilize training. GWS is applied to our model's loss function. We compared our model with word-based TextCNN trained with different data preprocessing methods.
We obtained some improvement in classification accuracy over word-based TextCNN without any ensemble or data augmentation.
Here's the overview of Seq2CNN model.