langtest.modelhandler.promptguard.PromptGuard#

class PromptGuard(model_name: str = 'meta-llama/Prompt-Guard-86M', device='cpu')#

Bases: object

__init__(model_name: str = 'meta-llama/Prompt-Guard-86M', device='cpu') None#

Methods

__init__([model_name, device])

get_indirect_injection_score(text[, ...])

Get indirect injection score for a single input text.

get_indirect_injection_scores_for_texts(texts)

Get indirect injection scores for a batch of texts.

get_jailbreak_score(text[, temperature, ...])

Get jailbreak score for a single input text.

get_jailbreak_scores_for_texts(texts[, ...])

Get jailbreak scores for a batch of texts.

get_indirect_injection_score(text, temperature=1.0, preprocess=True)#

Get indirect injection score for a single input text.

get_indirect_injection_scores_for_texts(texts, temperature=1.0, max_batch_size=16, preprocess=True)#

Get indirect injection scores for a batch of texts.

get_jailbreak_score(text, temperature=1.0, preprocess=True)#

Get jailbreak score for a single input text.

get_jailbreak_scores_for_texts(texts, temperature=1.0, max_batch_size=16, preprocess=True)#

Get jailbreak scores for a batch of texts.