simpleT5
simpleT5 copied to clipboard
simpleT5 is built on top of PyTorch-lightning⚡️ and Transformers🤗 that lets you quickly train your T5 models.
I tried to train My model for translating English to Bengali. After Training when I run the code, The output is not Unicode Bengali character. I Eat Rice (eng)=> আমি...
import soundfile as sf from scipy.io import wavfile from IPython.display import Audio from transformers import Wav2Vec2ForCTC, Wav2Vec2CTCTokenizer import speech_recognition as sr import io from pydub import AudioSegment tokenizer = Wav2Vec2CTCTokenizer.from_pretrained("facebook/wav2vec2-base-960h")...
` def training_epoch_end(self, training_step_outputs): """ save tokenizer and model on epoch end """ self.average_training_loss = np.round( torch.mean(torch.stack([x["loss"] for x in training_step_outputs])).item(), 4, ) path = f"{self.outputdir}/simplet5-epoch-{self.current_epoch}-train-loss-{str(self.average_training_loss)}-val-loss-{str(self.average_validation_loss)}"` Will be very helpful...
I want to train my t5 model from scratch and with bpe tokenizer, is there a example ?
I'd like to save the model. It has a load method, but if i save using model.model.save_pretrained and then use model.load I get: OSError: Can't load tokenizer for 't5.model'. If...
I don't find the parameter for multi-gpus training
hi im trying to create a new language model for tamil downtask is abstractive question Answering how to use simple t5 to build a new language model i have dataset...
Thanks for outsourcing nice repo @Shivanandroy did you also develop functionality to predict multiple strings as a single vector like batch execution. Its computationally expensive to predict one by one
Any chance you can add TPU support in the Colab? I think this is supported more of less out of the box now in the newest pyTorch Lightning versions.