MiniCPM icon indicating copy to clipboard operation
MiniCPM copied to clipboard

MiniCPM4: Ultra-Efficient LLMs on End Devices, achieving 5+ speedup on typical end-side chips

Results 152 MiniCPM issues
Sort by recently updated
recently updated
newest added

### Is there an existing issue ? / 是否已有相关的 issue ? - [X] I have searched, and there is no existing issue. / 我已经搜索过了,没有相关的 issue。 ### Describe the bug /...

bug
triage

### Description / 描述 error loading model architecture: unknown model architecture: 'minicpm3' time=2024-10-12T20:12:01.292+08:00 level=ERROR source=sched.go:456 msg="error loading llama server" error="llama runner process has terminated: this model is not supported by...

badcase

### Feature request / 功能建议 大部分人在微调2b模型用的是单卡,建议默认参数为--include localhost:0

feature

### Feature request / 功能建议 如果使用CPU的话,要求的CPU核数和内存大概是多大,运行4B的模型。

feature

### Feature request / 功能建议 后续有开源Minicpm量化模型相关代码的计划吗。

feature

### Is there an existing issue ? / 是否已有相关的 issue ? - [x] I have searched, and there is no existing issue. / 我已经搜索过了,没有相关的 issue。 ### Describe the bug /...

bug
triage

Thank you for releasing the excellent model and work. The paper appears to state that the 8B model was pre-trained natively with a 32K sequence length. I would like to...

请问你们开源的代码中,在哪里能找到sparse attention的torch实现吗

### Description / 描述 你好! 我注意到minicpm4-0.5B中描述无法支持sparse attention,但是我看modeling中有就修改为可用,但报错: ``` topk_idx[topk_idx >= q_idx[None, :, None]] = -1 RuntimeError: The size of tensor a (355) must match the size of tensor b (711)...

badcase