langtest.modelhandler.promptguard.PromptGuard#
- class PromptGuard(model_name: str = 'meta-llama/Prompt-Guard-86M', device='cpu')#
Bases:
object
Methods
__init__
([model_name, device])get_indirect_injection_score
(text[, ...])Get indirect injection score for a single input text.
Get indirect injection scores for a batch of texts.
get_jailbreak_score
(text[, temperature, ...])Get jailbreak score for a single input text.
get_jailbreak_scores_for_texts
(texts[, ...])Get jailbreak scores for a batch of texts.
- get_indirect_injection_score(text, temperature=1.0, preprocess=True)#
Get indirect injection score for a single input text.
- get_indirect_injection_scores_for_texts(texts, temperature=1.0, max_batch_size=16, preprocess=True)#
Get indirect injection scores for a batch of texts.
- get_jailbreak_score(text, temperature=1.0, preprocess=True)#
Get jailbreak score for a single input text.
- get_jailbreak_scores_for_texts(texts, temperature=1.0, max_batch_size=16, preprocess=True)#
Get jailbreak scores for a batch of texts.