llm-security-prompt-injection icon indicating copy to clipboard operation
llm-security-prompt-injection copied to clipboard

This project investigates the security of large language models by performing binary classification of a set of input prompts to discover malicious prompts. Several approaches have been analyzed using...

Results 0 llm-security-prompt-injection issues
Sort by recently updated
recently updated
newest added