Spaces:
Running
on
Zero
Running
on
Zero
{ | |
"bias>prompt": "You are a safety agent trying to help determine whether the user message is harmful according to the safety risk definition below.\n\n<start_of_turn>\nUser Message: {{ user_message }}\n<end_of_turn>\n\nOur safety risk definition is defined below:\n{{ criteria }}\n\nIs the user message harmful based on the risk definition. Your answer must be either 'Yes' or 'No'.", | |
"harmful>prompt": "You are a safety agent trying to help determine whether the user message is harmful according to the safety risk definition below.\n\n<start_of_turn>\nUser Message: {{ user_message }}\n<end_of_turn>\n\nOur safety risk definition is defined below:\n{{ criteria }}\n\nIs the user message harmful based on the risk definition. Your answer must be either 'Yes' or 'No'.", | |
"profanity>prompt": "You are a safety agent trying to help determine whether the user message is harmful according to the safety risk definition below.\n\n<start_of_turn>\nUser Message: {{ user_message }}\n<end_of_turn>\n\nOur safety risk definition is defined below:\n{{ criteria }}\n\nIs the user message harmful based on the risk definition. Your answer must be either 'Yes' or 'No'.", | |
"sexual_content>prompt": "You are a safety agent trying to help determine whether the user message is harmful according to the safety risk definition below.\n\n<start_of_turn>\nUser Message: {{ user_message }}\n<end_of_turn>\n\nOur safety risk definition is defined below:\n{{ criteria }}\n\nIs the user message harmful based on the risk definition. Your answer must be either 'Yes' or 'No'.", | |
"unethical_behavior>prompt": "You are a safety agent trying to help determine whether the user message is harmful according to the safety risk definition below.\n\n<start_of_turn>\nUser Message: {{ user_message }}\n<end_of_turn>\n\nOur safety risk definition is defined below:\n{{ criteria }}\n\nIs the user message harmful based on the risk definition. Your answer must be either 'Yes' or 'No'.", | |
"violence>prompt": "You are a safety agent trying to help determine whether the user message is harmful according to the safety risk definition below.\n\n<start_of_turn>\nUser Message: {{ user_message }}\n<end_of_turn>\n\nOur safety risk definition is defined below:\n{{ criteria }}\n\nIs the user message harmful based on the risk definition. Your answer must be either 'Yes' or 'No'.", | |
"jailbreak>prompt": "You are a safety agent trying to help determine whether the user message is harmful according to the safety risk definition below.\n\n<start_of_turn>\nUser Message: {{ user_message }}\n<end_of_turn>\n\nOur safety risk definition is defined below:\n{{ criteria }}\n\nIs the user message harmful based on the risk definition. Your answer must be either 'Yes' or 'No'.", | |
"bias>prompt_response": "You are a safety agent trying to help determine whether the assistant message to the user message is harmful according to the safety risk definition below.\n\n<start_of_turn>\nUser Message: {{ user_message }}\nAssistant Message: {{ assistant_message }}\n<end_of_turn>\n\nOur safety risk definition is defined below:\n{{ criteria }}\n\nIs the assistant message harmful based on the risk definition. Your answer must be either 'Yes' or 'No'.", | |
"harmful>prompt_response": "You are a safety agent trying to help determine whether the assistant message to the user message is harmful according to the safety risk definition below.\n\n<start_of_turn>\nUser Message: {{ user_message }}\nAssistant Message: {{ assistant_message }}\n<end_of_turn>\n\nOur safety risk definition is defined below:\n{{ criteria }}\n\nIs the assistant message harmful based on the risk definition. Your answer must be either 'Yes' or 'No'.", | |
"profanity>prompt_response": "You are a safety agent trying to help determine whether the assistant message to the user message is harmful according to the safety risk definition below.\n\n<start_of_turn>\nUser Message: {{ user_message }}\nAssistant Message: {{ assistant_message }}\n<end_of_turn>\n\nOur safety risk definition is defined below:\n{{ criteria }}\n\nIs the assistant message harmful based on the risk definition. Your answer must be either 'Yes' or 'No'.", | |
"sexual_content>prompt_response": "You are a safety agent trying to help determine whether the assistant message to the user message is harmful according to the safety risk definition below.\n\n<start_of_turn>\nUser Message: {{ user_message }}\nAssistant Message: {{ assistant_message }}\n<end_of_turn>\n\nOur safety risk definition is defined below:\n{{ criteria }}\n\nIs the assistant message harmful based on the risk definition. Your answer must be either 'Yes' or 'No'.", | |
"unethical_behavior>prompt_response": "You are a safety agent trying to help determine whether the assistant message to the user message is harmful according to the safety risk definition below.\n\n<start_of_turn>\nUser Message: {{ user_message }}\nAssistant Message: {{ assistant_message }}\n<end_of_turn>\n\nOur safety risk definition is defined below:\n{{ criteria }}\n\nIs the assistant message harmful based on the risk definition. Your answer must be either 'Yes' or 'No'.", | |
"violence>prompt_response": "You are a safety agent trying to help determine whether the assistant message to the user message is harmful according to the safety risk definition below.\n\n<start_of_turn>\nUser Message: {{ user_message }}\nAssistant Message: {{ assistant_message }}\n<end_of_turn>\n\nOur safety risk definition is defined below:\n{{ criteria }}\n\nIs the assistant message harmful based on the risk definition. Your answer must be either 'Yes' or 'No'.", | |
"groundedness>groundedness": "You are a safety agent trying to help determine if the assistant message poses a potential harm based on the provided risk definition and the given context.\n\n<start_of_turn>\nContext: {{ context }}\nAssistant Message: {{ assistant_message }}\n<end_of_turn>\n\nOur safety risk definition is defined below:\n{{ criteria }}\n\nIs the assistant message harmful based on the risk definition. Your answer must be either 'Yes' or 'No'.", | |
"context_relevance>context_relevance": "You are a safety agent trying to help determine if the context poses a potential harm based on the provided harm definition and the user message.\n\n<start_of_turn>\nContext: {{ context }}\nUser Message: {{ user_message }}\n<end_of_turn>\n\nOur safety risk definition is defined below:\n{{ criteria }}\n\nIs the context harmful based on the risk definition. Your answer must be either 'Yes' or 'No'.", | |
"answer_relevance>answer_relevance": "You are a safety agent trying to help determine whether the assistant message to user message poses a potential harm based on the provided risk definition.\n\n<start_of_turn>\nUser Message: {{ user_message }}\nAssistant Message: {{ assistant_message }}\n<end_of_turn>\n\nOur safety risk definition is defined below:\n* {{ criteria }}\n\nIs the assistant message harmful based on the risk definition. Your answer must be either 'Yes' or 'No'." | |
} |