MiniCPM
MiniCPM copied to clipboard
MiniCPM4: Ultra-Efficient LLMs on End Devices, achieving 5+ speedup on typical end-side chips
### Feature request / 功能建议 Hi, all Can you guys provide more details about minicpm-4 Eagle3 series models QAT process and information about the quantization framework/code repo? It would be...
### Description / 描述 如图第一步loss为500多,在其他模型刚开始的时候都是3-4左右 ### Case Explaination / 案例解释 _No response_
### Description / 描述 INFO 08-06 08:34:26 [__init__.py:244] Automatically detected platform cuda. INFO 08-06 08:34:31 [api_server.py:1287] vLLM API server version 0.9.1 INFO 08-06 08:34:32 [cli_args.py:309] non-default args: {'model': '/llm/models/MiniCPM4-8B', 'dtype':...
### Description / 描述 你好,请问我在将minicpm4-0.5B和minicpm4-8B的模型下载下来之后,使用范例代码推理: `from transformers import AutoModelForCausalLM, AutoTokenizer` `import torch` `torch.manual_seed(0)` `path = "/mnt/2/haochen/LLM/MiniCPM/pretrained_models/MiniCPM4-0.5B"` `device = "cuda"` `tokenizer = AutoTokenizer.from_pretrained(path)` `model = AutoModelForCausalLM.from_pretrained(path, torch_dtype=torch.bfloat16, device_map=device, trust_remote_code=True)` `responds, history...
I'm encountering training instability when attempting to fine-tune MiniCPM4-8B using InfLLM v2 with the provided sparse configuration. The training collapses immediately with gradient norm NaN at the first optimization step...
### Description / 描述 ### Case Explaination / 案例解释 _No response_
### Feature request / 功能建议 目前BitCPM只提供了量化后的三值权重,能否开源量化前的权重。
### Is there an existing issue ? / 是否已有相关的 issue ? - [x] I have searched, and there is no existing issue. / 我已经搜索过了,没有相关的 issue。 ### Describe the bug /...