Sebastian Raschka

Results 180 issues of Sebastian Raschka

Should we automatically infer and download the tokenizer config when no `--tokenizer_dir` is provided in `litgpt pretrain`? I.e., ```bash litgpt download \ --repo_id EleutherAI/pythia-14m \ --tokenizer_only true litgpt pretrain \...

These PR does 2 things 1) Users were confused about the initial n/a in the validation loss. One idea I had was that we can actually use the initial validation...

enhancement

If a user doesn't set `--data.val_split_fraction` in ```bash litgpt finetune lora \ --data JSON \ --data.json_path ....json \ --checkpoint_dir checkpoints/$REPO_NAME ``` It raises an error ```bash train_data, test_data = self.get_splits()...

enhancement
good first issue

Ideally we only want the SFT finetuning datasets to show up in `litgpt finetune --help` and only the pretraining datasets to show up in `litgpt pretrain --help`. I believe we...

enhancement

I think that all models should all work out of the box with the defaults. Maybe we need to do some automatic clipping on the inputs and raise a warning...

enhancement

LLMs are commonly pretrained using gradient clipping, which we also implement for pretraining here in LitGPT. In addition, it might be useful to add a gradient clipping option to the...

enhancement

I talked to some users who were a bit confused / surprised that the validation loss didn't get calculated during a run. I think that's a valid issue if you...

enhancement

Is there anything that speaks against listing Python 3.12 support on the Readme? ![PyPI - Python Version](https://img.shields.io/pypi/pyversions/pytorch-lightning)

As far as I know, one can pretrain fine using TinyStories: ```python litgpt/pretrain.py --data litgpt.data.TinyStories ``` Should we add this to the documentation? Right now, we only have a `pretrain_tinyllama.md`...

documentation