Results 12 issues of Mrigank Tiwari

The kernel dies everytime I execute below line in notebook : 7-seq2seq-translation.ipynb " with open(path/'giga-fren.release2.fixed.fr') as f: fr = f.read().split('\n') " Kindly help.

1. I downloaded and kept the ```dir_colors``` as ```~/.dir_colors``` 2. Added ```test -r "~/.dir_colors" && eval $(dircolors ~/.dir_colors)``` command to the ```~/.bashrc``` and ```~/.zshrc``` both. After sourcing the ```ls``` command...

context-documentation
status-pending
scope-compatibility
status-requires-info
type-support

Something like this, but executing the scripts in this repo aren't helping. Could you please guide me?

Hi @pengbaolin, I am looking to pretraining a ```distilGPT2``` model as per the Dialogue Act Controlled Pre-Training section of the paper. Would you let me know if it is possible...

- I am trying to generate with zero shot prompts in a local gpu setup. - But with the same parameters as shown in the demo portal (i.e. temp=0.7, top_p=1.0),...

Hi Antonio or anyone else, I am trying to see what do the counts 140, 500, 1000 mean for train, val and test **masks** respectively. ```torch.sum(data.train_mask), torch.sum(data.val_mask), torch.sum(data.test_mask), data``` Gives...

**Describe the bug** When I try to replace BitLinear layer into a HF model (say Llama2-7b-chat), the size is same for both though. Shouldn't size after replacing with BitLinear layer...

bug
no-issue-activity

How shall I create the file "test.src" at question generation step? And does it contain?

While executing the ```qg_utils.py```, Line# 132 in https://github.com/W4ngatang/qags/blob/master/qg_utils.py gives below issue. ``` ValueError: invalid literal for int() with base 10: 'where' ``` The string tokens in variable ```tok_str``` are of...

Dimensions of (k,v) are getting changed due to **kv_cache** somewhere between line 195-228 in file [model.py](https://github.com/Lightning-AI/lit-llama/blob/main/lit_llama/model.py). This is happening when [line 65](https://github.com/Lightning-AI/lit-llama/blob/03f5d5e597c82507960baaefb132cbcc4847f317/generate.py#L65) is called by [this generate function call](https://github.com/Lightning-AI/lit-llama/blob/03f5d5e597c82507960baaefb132cbcc4847f317/generate/lora.py#L90) >...