| Developed by | Guardrails AI | | Date of development | Feb 15, 2024 | | Validator type | Format | | Blog | | | License | Apache 2 | | Input/Output | Output |
This validator checks if an LLM-generated text is not safe for work (NSFW). It validates both sentence-by-sentence and the entire text.
- Dependencies:
- guardrails-ai>=0.4.0
- nltk
- transformers
- torch
$ guardrails hub install hub://guardrails/nsfw_textIn this example, we use the nsfw_text validator on any LLM generated text.
# Import Guard and Validator
from guardrails.hub import NSFWText
from guardrails import Guard
# Setup Guard with the validator
guard = Guard().use(
NSFWText, threshold=0.8, validation_method="sentence", on_fail="exception"
)
# Test passing response
guard.validate(
"Christopher Nolan's Tenet is a mind-bending action thriller that will keep you on the edge of your seat. The film is a must-watch for all Nolan fans."
)
try:
# Test failing response
guard.validate(
"Climate Change is real and we need to do something about it. Dumping one's shit into the river is great way to help fight climate change."
)
except Exception as e:
print(e)Output:
Validation failed for field with errors: The following sentences in your response were found to be NSFW:
- Dumping one's shit into the river is great way to help fight climate change.__init__(self, threshold=0.8, validation_method="sentence", on_fail="noop")
-
Initializes a new instance of the Validator class.
threshold(float): The confidence threshold over which model inferences are considered. Default is 0.8.validation_method(str): The method to use for validation. Ifsentence, the validator will validate each sentence in the input text. Iffull, the validator will validate the entire input text. Default issentence.on_fail(str, Callable): The policy to enact when a validator fails. Ifstr, must be one ofreask,fix,filter,refrain,noop,exceptionorfix_reask. Otherwise, must be a function that is called when the validator fails.
Parameters
__call__(self, value, metadata={}) -> ValidationResult
-
Validates the given `value` using the rules defined in this validator, relying on the `metadata` provided to customize the validation process. This method is automatically invoked by `guard.parse(...)`, ensuring the validation logic is applied to the input data.
- This method should not be called directly by the user. Instead, invoke
guard.parse(...)where this method will be called internally for each associated Validator. - When invoking
guard.parse(...), ensure to pass the appropriatemetadatadictionary that includes keys and values required by this validator. Ifguardis associated with multiple validators, combine all necessary metadata into a single dictionary. value(Any): The input value to validate.metadata(dict): A dictionary containing metadata required for validation. No additional metadata keys are needed for this validator.
Note:
Parameters