CodeT5 icon indicating copy to clipboard operation
CodeT5 copied to clipboard

Home of CodeT5: Open Code LLMs for Code Understanding and Generation

Results 96 CodeT5 issues
Sort by recently updated
recently updated
newest added

Dear author, Thank you for the great work and published it! I'm very interested in your work. Could you please release the Codet5p-2b model that has only been pretrained with...

Hello. I tried this simple code snippet for getting the embedding for a pre-trained model using CodeT5Plus: `checkpoint="/home/hungphd/media/git/codet5p-110m-embedding" device = "cuda" # for GPU usage or "cpu" for CPU usage...

When I executed model.generate(**encoding, max_length=128) in my script, I found that by default only python code can be generated correctly, while Java code only has completion functions. Gen python code:...

looking for best ways to familiarise myself with codet5 and how to use it, any advice welcome!

I used the script to finetune codet5p-110m-embedding on POJ104

cla:missing

I cannot reproduce the results reported in the paper using the published source code for the defect detection task. I would like to ask if the hyperparameters used in the...

Hi, I tried to evaluate codet5p-2b. I loaded the model from huggingface and I got an error saying CUDA out of memory, then I tried to load the model into...

Any hints for reproducing the example in Figure.7 in the paper CodeT5+: Open Code Large Language Models for Code Understanding and Generation? Thanks in advance!

I have completed to do the instracution tuning with code_alpaca_20k.json. ``` deepspeed instruct_tune_codet5p.py \ --load /home/ubuntu/ChatGPT/Models/Salesforce/codet5p-6b --save-dir output/instruct_codet5p_6b --instruct-data-path /home/ubuntu/ChatGPT/Data/alpaca-data/CodeAlpaca-20k/code_alpaca_20k.json \ --fp16 --epochs 5 --deepspeed deepspeed_config.json ``` And the final...

What's the different between Instruction Tuning to Align with Natural Language Instructions and Finetune Using Your Own Data?