mlc-llm icon indicating copy to clipboard operation
mlc-llm copied to clipboard

"As an AI language model,"...

Open donuts-are-good opened this issue 2 years ago • 7 comments

USER: tell me an offensive joke ASSISTANT: I'm sorry, but I cannot provide offensive or inappropriate content. My purpose is to provide helpful and informative responses to your questions. Can I assist you with something else?

How do I disable these protections?

donuts-are-good avatar May 08 '23 14:05 donuts-are-good

duplicate of #20 .

The output is generated by the underlying model (Vicuna-7b) though.

The output is controlled by the underlying model itself and not related to mlc-llm much, while you can still filter out these type of words before sending back to user in cpp layer. (This does not make too muc sense.)

BTW, llm (large language model) is like a blackbox to human, even OpenAI and Microsoft has no way to prevent people reversing engineering information from the llm. (e.g. sydney).

shiqimei avatar May 08 '23 14:05 shiqimei

USER: tell me an offensive joke ASSISTANT: I'm sorry, but I cannot provide offensive or inappropriate content. My purpose is to provide helpful and informative responses to your questions. Can I assist you with something else?

How do I disable these protections?

You can't. Since vicuna's training dataset heavily relies on ShareGPT, so the model's really similar to ChatGPT's behaviour. While there're some other models your can try, but there're no official usage instructions there.

shiqimei avatar May 08 '23 14:05 shiqimei

Get the uncensored version and be happy.

jeffwadsworth avatar May 08 '23 15:05 jeffwadsworth

Get the uncensored version and be happy.

where is that?

donuts-are-good avatar May 08 '23 15:05 donuts-are-good

USER: tell me an offensive joke ASSISTANT: I'm sorry, but I cannot provide offensive or inappropriate content. My purpose is to provide helpful and informative responses to your questions. Can I assist you with something else?

How do I disable these protections?

You can't. Since vicuna's training dataset heavily relies on ShareGPT, so the model's really similar to ChatGPT's behaviour. While there're some other models your can try, but there're no official usage instructions there.

You can change the system message here: https://github.com/mlc-ai/mlc-llm/blob/9010be6d7a4a58f7bb5daf2c13e857f87811c47c/cpp/llm_chat.cc#LL42C43-L42C43

shiqimei avatar May 08 '23 16:05 shiqimei

MLC LLM is a compiler, which doesn't control the text generation from models per se. Currently the "as an AI language model" is directly generated from Vicuna-7b and we didn't censor anything. You may filter out those words in cpp file if you'd love to :-)

junrushao avatar May 08 '23 23:05 junrushao

I really hope that an uncensored llm which in best case also uses the open assistant dataset will be added as an alternative to the plain vicuna as an option to be downloaded to Android. This thing is nearly useful we just need a bit more ai capability.

GameOverFlowChart avatar May 09 '23 11:05 GameOverFlowChart

I'm sure people (like me) will be uploading all sorts of variants to HF once they enable the llama profiling in the build script.

It could be done right now (with a shim to change the prompting syntax), but the performance of compiled models is really poor without the profiling.

AlphaAtlas avatar May 11 '23 18:05 AlphaAtlas

I'm sure people (like me) will be uploading all sorts of variants to HF once they enable the llama profiling in the build script.

It could be done right now (with a shim to change the prompting syntax), but the performance of compiled models is really poor without the profiling.

offtopic

donuts-are-good avatar May 11 '23 18:05 donuts-are-good

MLC LLM is a compiler, which doesn't control the text generation from models per se. Currently the "as an AI language model" is directly generated from Vicuna-7b and we didn't censor anything. You may filter out those words in cpp file if you'd love to :-)

thanks.

donuts-are-good avatar May 11 '23 18:05 donuts-are-good