Neo LLama

Results 13 issues of Neo LLama

model = RobertaModel.from_pretrained('../model/RoBERTa_zh_Large_PyTorch/') tokenizer=RobertaTokenizer.from_pretrained('../model/RoBERTa_zh_Large_PyTorch/') 求问我下载后加载,model可以,但是tokenzier加载失败,但是用BertTokenizer可以加载,求回答谢谢

url4 = f"https://api.twitter.com/2/users/by/username/{self.UserByScreenName}?&expansions=pinned_tweet_id&tweet.fields=created_at" i use this url,but didn't return the tweet text ![image](https://user-images.githubusercontent.com/55910045/210207080-7e2231f1-8de8-4bc8-b1e8-a2916014679d.png) but your document case return

bug

when i run the example code in readme,when exec nx.draw_networkx_nodes the program just return ,i don' know why,can you help me ? ![image](https://github.com/taynaud/python-louvain/assets/55910045/b19780f2-a1d7-4000-8138-695f5f5af5f7)

What is the difference between your package's bleu implementation and sacrebleu implementation? I calculated the result differently in the two ways, Chinese expected, passed sacrebleu's zh tokenizer

i think your py file have problem to process the raw dataset,such as : 当 O 希 O 望 O 工 O 程 O 救 O 助 O 的 O...

Compare to deepseek-moe, your model add an additional learnable parameter `self.shared_expert_gate = torch.nn.Linear(config.hidden_size, 1, bias=False)`, to control the proportion of the results of shared experts. There are somethings i want...

### Non-ChatGPT bug - [X] This issue does not occur on chat.openai.com and only occurs on this app. ### Version v1.1.0 ### Bug description The Windows version 1.1.0 is unable...

bug

**Describe the bug** ``` self.param_groups_fp16_flat_cpu_memory.append(get_accelerator().pin_memory( File "/usr/local/python/lib/python3.8/site-packages/deepspeed/accelerator/cuda_accelerator.py", line 291, in pin_memory self.param_groups_fp16_flat_cpu_memory.append(get_accelerator().pin_memory( File "/usr/local/python/lib/python3.8/site-packages/deepspeed/accelerator/cuda_accelerator.py", line 291, in pin_memory return tensor.pin_memory() RuntimeError: cannot pin 'CUDABFloat16Type' only dense CPU tensors can be...

bug
training

### 提交前必须检查以下项目 - [X] 请确保使用的是仓库最新代码(git pull),一些问题已被解决和修复。 - [X] 由于相关依赖频繁更新,请确保按照[Wiki](https://github.com/ymcui/Chinese-LLaMA-Alpaca/wiki)中的相关步骤执行 - [X] 我已阅读[FAQ章节](https://github.com/ymcui/Chinese-LLaMA-Alpaca/wiki/常见问题)并且已在Issue中对问题进行了搜索,没有找到相似问题和解决方案 - [X] 第三方插件问题:例如[llama.cpp](https://github.com/ggerganov/llama.cpp)、[text-generation-webui](https://github.com/oobabooga/text-generation-webui)、[LlamaChat](https://github.com/alexrozanski/LlamaChat)等,同时建议到对应的项目中查找解决方案 - [X] 模型正确性检查:务必检查模型的[SHA256.md](https://github.com/ymcui/Chinese-LLaMA-Alpaca/blob/main/SHA256.md),模型不对的情况下无法保证效果和正常运行 ### 问题类型 None ### 基础模型 None ### 操作系统 None ### 详细描述问题 ```...

stale

https://github.com/cooper12121/llama3-8x8b-MoE/