MagicSource

Results 1144 comments of MagicSource

@LinB203 Shouldn't be 1.8 MoE? Does the model open? Does the pretrained (no finetune) model able to do simple image caption ?

@LinB203 1. When will the model release? Looks like better than currect model, is it also 1.8bx4 moe? 2. Which pretrained data you were using? sharegpt4v_instruct_gpt4-vision_cap100k.json or pt part? Does...

@LinB203 I think the pretrain loss hard to be 0.2, the official pretrain loss of llava is about 1.9: ![image](https://github.com/PKU-YuanGroup/MoE-LLaVA/assets/21303438/9912da19-92a7-4fd3-b788-65d40eeb5c8c) How did u guys manageed trained pretrain loss so small?

@LinB203 I found the sharegpt4v also lack of Chinese part data. Do u think any hight quality Chinese pretrain image-text pair can be used to enhance Chinese ability?

@LinB203 Do u think raw ocr image-text paris can be used in pretrain data?

openllama版本啥时候发布呢

@peakji Ithink its not related about model. For model am simple using Llama. The reason is that when we decode same id, compare with decode ids in a sentence, tokenizers...

Or maybe these is something missed inside your StreamTokenizer? (like ignored some ids). Can u try get decode ids one by one and print it? ``` outputs = [] for...

@peakji Thanks, I just using tokenizer of StreamModel and the Chinese decoding error problems still exist. ![image](https://github.com/hyperonym/basaran/assets/21303438/8ba6fadb-34a6-47d7-b644-f419543a453a) And I still can not get the spaces between engliesh words . I...

I got no space and Chinese were wrong either (try print(word, end='')) I don't want change line in every word and I don't want unexpect spaces in un-English characters.