awesome-llm-security
awesome-llm-security copied to clipboard
A curation of awesome tools, documents and projects about LLM Security.
Hello! I would like to add our completed paper from MSFT Research about defense against adversarial attacks.
Operationalizing a Threat Model for Red-Teaming LLMs
Yu, Zhiyuan et al. “Don't Listen To Me: Understanding and Exploring Jailbreak Prompts of Large Language Models.” ArXiv abs/2403.17336 (2024): n. pag.
added Machine_Learning_CTF_Challenges from https://github.com/alexdevassy/Machine_Learning_CTF_Challenges
Thank you for the wonderful paper collection. We have a line of research on harmful fine-tuning for LLMs. Could you please include this line of work into the repo? |...