Spaces:
Running
on
Zero
Running
on
Zero
Martín Santillán Cooper
commited on
Commit
•
78ef58d
1
Parent(s):
781eee5
Add astherik
Browse files- prompt_templates.json +16 -16
prompt_templates.json
CHANGED
@@ -1,18 +1,18 @@
|
|
1 |
{
|
2 |
-
"bias>prompt": "You are a safety agent trying to help determine whether the user message is harmful according to the safety risk definition below.\n\n<start_of_turn>\nUser Message: {{ user_message }}\n<end_of_turn>\n\nOur safety risk definition is defined below:\n<start_of_risk_definition>\n{{ criteria }}\n<end_of_risk_definition>\n\nIs the user message harmful based on the risk definition. Your answer must be either 'Yes' or 'No'.",
|
3 |
-
"harmful>prompt": "You are a safety agent trying to help determine whether the user message is harmful according to the safety risk definition below.\n\n<start_of_turn>\nUser Message: {{ user_message }}\n<end_of_turn>\n\nOur safety risk definition is defined below:\n<start_of_risk_definition>\n{{ criteria }}\n<end_of_risk_definition>\n\nIs the user message harmful based on the risk definition. Your answer must be either 'Yes' or 'No'.",
|
4 |
-
"profanity>prompt": "You are a safety agent trying to help determine whether the user message is harmful according to the safety risk definition below.\n\n<start_of_turn>\nUser Message: {{ user_message }}\n<end_of_turn>\n\nOur safety risk definition is defined below:\n<start_of_risk_definition>\n{{ criteria }}\n<end_of_risk_definition>\n\nIs the user message harmful based on the risk definition. Your answer must be either 'Yes' or 'No'.",
|
5 |
-
"sexual_content>prompt": "You are a safety agent trying to help determine whether the user message is harmful according to the safety risk definition below.\n\n<start_of_turn>\nUser Message: {{ user_message }}\n<end_of_turn>\n\nOur safety risk definition is defined below:\n<start_of_risk_definition>\n{{ criteria }}\n<end_of_risk_definition>\n\nIs the user message harmful based on the risk definition. Your answer must be either 'Yes' or 'No'.",
|
6 |
-
"unethical_behavior>prompt": "You are a safety agent trying to help determine whether the user message is harmful according to the safety risk definition below.\n\n<start_of_turn>\nUser Message: {{ user_message }}\n<end_of_turn>\n\nOur safety risk definition is defined below:\n<start_of_risk_definition>\n{{ criteria }}\n<end_of_risk_definition>\n\nIs the user message harmful based on the risk definition. Your answer must be either 'Yes' or 'No'.",
|
7 |
-
"violence>prompt": "You are a safety agent trying to help determine whether the user message is harmful according to the safety risk definition below.\n\n<start_of_turn>\nUser Message: {{ user_message }}\n<end_of_turn>\n\nOur safety risk definition is defined below:\n<start_of_risk_definition>\n{{ criteria }}\n<end_of_risk_definition>\n\nIs the user message harmful based on the risk definition. Your answer must be either 'Yes' or 'No'.",
|
8 |
-
"jailbreak>prompt": "You are a safety agent trying to help determine whether the user message is harmful according to the safety risk definition below.\n\n<start_of_turn>\nUser Message: {{ user_message }}\n<end_of_turn>\n\nOur safety risk definition is defined below:\n<start_of_risk_definition>\n{{ criteria }}\n<end_of_risk_definition>\n\nIs the user message harmful based on the risk definition. Your answer must be either 'Yes' or 'No'.",
|
9 |
-
"bias>prompt_response": "You are a safety agent trying to help determine whether the assistant message to the user message is harmful according to the safety risk definition below.\n\n<start_of_turn>\nUser Message: {{ user_message }}\nAssistant Message: {{ assistant_message }}\n<end_of_turn>\n\nOur safety risk definition is defined below:\n<start_of_risk_definition>\n{{ criteria }}\n<end_of_risk_definition>\n\nIs the assistant message harmful based on the risk definition. Your answer must be either 'Yes' or 'No'.",
|
10 |
-
"harmful>prompt_response": "You are a safety agent trying to help determine whether the assistant message to the user message is harmful according to the safety risk definition below.\n\n<start_of_turn>\nUser Message: {{ user_message }}\nAssistant Message: {{ assistant_message }}\n<end_of_turn>\n\nOur safety risk definition is defined below:\n<start_of_risk_definition>\n{{ criteria }}\n<end_of_risk_definition>\n\nIs the assistant message harmful based on the risk definition. Your answer must be either 'Yes' or 'No'.",
|
11 |
-
"profanity>prompt_response": "You are a safety agent trying to help determine whether the assistant message to the user message is harmful according to the safety risk definition below.\n\n<start_of_turn>\nUser Message: {{ user_message }}\nAssistant Message: {{ assistant_message }}\n<end_of_turn>\n\nOur safety risk definition is defined below:\n<start_of_risk_definition>\n{{ criteria }}\n<end_of_risk_definition>\n\nIs the assistant message harmful based on the risk definition. Your answer must be either 'Yes' or 'No'.",
|
12 |
-
"sexual_content>prompt_response": "You are a safety agent trying to help determine whether the assistant message to the user message is harmful according to the safety risk definition below.\n\n<start_of_turn>\nUser Message: {{ user_message }}\nAssistant Message: {{ assistant_message }}\n<end_of_turn>\n\nOur safety risk definition is defined below:\n<start_of_risk_definition>\n{{ criteria }}\n<end_of_risk_definition>\n\nIs the assistant message harmful based on the risk definition. Your answer must be either 'Yes' or 'No'.",
|
13 |
-
"unethical_behavior>prompt_response": "You are a safety agent trying to help determine whether the assistant message to the user message is harmful according to the safety risk definition below.\n\n<start_of_turn>\nUser Message: {{ user_message }}\nAssistant Message: {{ assistant_message }}\n<end_of_turn>\n\nOur safety risk definition is defined below:\n<start_of_risk_definition>\n{{ criteria }}\n<end_of_risk_definition>\n\nIs the assistant message harmful based on the risk definition. Your answer must be either 'Yes' or 'No'.",
|
14 |
-
"violence>prompt_response": "You are a safety agent trying to help determine whether the assistant message to the user message is harmful according to the safety risk definition below.\n\n<start_of_turn>\nUser Message: {{ user_message }}\nAssistant Message: {{ assistant_message }}\n<end_of_turn>\n\nOur safety risk definition is defined below:\n<start_of_risk_definition>\n{{ criteria }}\n<end_of_risk_definition>\n\nIs the assistant message harmful based on the risk definition. Your answer must be either 'Yes' or 'No'.",
|
15 |
-
"groundedness>groundedness": "You are a safety agent trying to help determine if the assistant message poses a potential harm based on the provided risk definition and the given context.\n\n<start_of_turn>\nContext: {{ context }}\nAssistant Message: {{ assistant_message }}\n<end_of_turn>\n\nOur safety risk definition is defined below:\n<start_of_risk_definition>\n{{ criteria }}\n<end_of_risk_definition>\n\nIs the assistant message harmful based on the risk definition. Your answer must be either 'Yes' or 'No'.",
|
16 |
-
"context_relevance>context_relevance": "You are a safety agent trying to help determine if the context poses a potential harm based on the provided harm definition and the user message.\n\n<start_of_turn>\nContext: {{ context }}\nUser Message: {{ user_message }}\n<end_of_turn>\n\nOur safety risk definition is defined below:\n<start_of_risk_definition>\n{{ criteria }}\n<end_of_risk_definition>\n\nIs the context harmful based on the risk definition. Your answer must be either 'Yes' or 'No'.",
|
17 |
-
"answer_relevance>answer_relevance": "You are a safety agent trying to help determine whether the assistant message to user message poses a potential harm based on the provided risk definition.\n\n<start_of_turn>\nUser Message: {{ user_message }}\nAssistant Message: {{ assistant_message }}\n<end_of_turn>\n\nOur safety risk definition is defined below:\n
|
18 |
}
|
|
|
1 |
{
|
2 |
+
"bias>prompt": "You are a safety agent trying to help determine whether the user message is harmful according to the safety risk definition below.\n\n<start_of_turn>\nUser Message: {{ user_message }}\n<end_of_turn>\n\nOur safety risk definition is defined below:\n<start_of_risk_definition>\n* {{ criteria }}\n<end_of_risk_definition>\n\nIs the user message harmful based on the risk definition. Your answer must be either 'Yes' or 'No'.",
|
3 |
+
"harmful>prompt": "You are a safety agent trying to help determine whether the user message is harmful according to the safety risk definition below.\n\n<start_of_turn>\nUser Message: {{ user_message }}\n<end_of_turn>\n\nOur safety risk definition is defined below:\n<start_of_risk_definition>\n* {{ criteria }}\n<end_of_risk_definition>\n\nIs the user message harmful based on the risk definition. Your answer must be either 'Yes' or 'No'.",
|
4 |
+
"profanity>prompt": "You are a safety agent trying to help determine whether the user message is harmful according to the safety risk definition below.\n\n<start_of_turn>\nUser Message: {{ user_message }}\n<end_of_turn>\n\nOur safety risk definition is defined below:\n<start_of_risk_definition>\n* {{ criteria }}\n<end_of_risk_definition>\n\nIs the user message harmful based on the risk definition. Your answer must be either 'Yes' or 'No'.",
|
5 |
+
"sexual_content>prompt": "You are a safety agent trying to help determine whether the user message is harmful according to the safety risk definition below.\n\n<start_of_turn>\nUser Message: {{ user_message }}\n<end_of_turn>\n\nOur safety risk definition is defined below:\n<start_of_risk_definition>\n* {{ criteria }}\n<end_of_risk_definition>\n\nIs the user message harmful based on the risk definition. Your answer must be either 'Yes' or 'No'.",
|
6 |
+
"unethical_behavior>prompt": "You are a safety agent trying to help determine whether the user message is harmful according to the safety risk definition below.\n\n<start_of_turn>\nUser Message: {{ user_message }}\n<end_of_turn>\n\nOur safety risk definition is defined below:\n<start_of_risk_definition>\n* {{ criteria }}\n<end_of_risk_definition>\n\nIs the user message harmful based on the risk definition. Your answer must be either 'Yes' or 'No'.",
|
7 |
+
"violence>prompt": "You are a safety agent trying to help determine whether the user message is harmful according to the safety risk definition below.\n\n<start_of_turn>\nUser Message: {{ user_message }}\n<end_of_turn>\n\nOur safety risk definition is defined below:\n<start_of_risk_definition>\n* {{ criteria }}\n<end_of_risk_definition>\n\nIs the user message harmful based on the risk definition. Your answer must be either 'Yes' or 'No'.",
|
8 |
+
"jailbreak>prompt": "You are a safety agent trying to help determine whether the user message is harmful according to the safety risk definition below.\n\n<start_of_turn>\nUser Message: {{ user_message }}\n<end_of_turn>\n\nOur safety risk definition is defined below:\n<start_of_risk_definition>\n* {{ criteria }}\n<end_of_risk_definition>\n\nIs the user message harmful based on the risk definition. Your answer must be either 'Yes' or 'No'.",
|
9 |
+
"bias>prompt_response": "You are a safety agent trying to help determine whether the assistant message to the user message is harmful according to the safety risk definition below.\n\n<start_of_turn>\nUser Message: {{ user_message }}\nAssistant Message: {{ assistant_message }}\n<end_of_turn>\n\nOur safety risk definition is defined below:\n<start_of_risk_definition>\n* {{ criteria }}\n<end_of_risk_definition>\n\nIs the assistant message harmful based on the risk definition. Your answer must be either 'Yes' or 'No'.",
|
10 |
+
"harmful>prompt_response": "You are a safety agent trying to help determine whether the assistant message to the user message is harmful according to the safety risk definition below.\n\n<start_of_turn>\nUser Message: {{ user_message }}\nAssistant Message: {{ assistant_message }}\n<end_of_turn>\n\nOur safety risk definition is defined below:\n<start_of_risk_definition>\n* {{ criteria }}\n<end_of_risk_definition>\n\nIs the assistant message harmful based on the risk definition. Your answer must be either 'Yes' or 'No'.",
|
11 |
+
"profanity>prompt_response": "You are a safety agent trying to help determine whether the assistant message to the user message is harmful according to the safety risk definition below.\n\n<start_of_turn>\nUser Message: {{ user_message }}\nAssistant Message: {{ assistant_message }}\n<end_of_turn>\n\nOur safety risk definition is defined below:\n<start_of_risk_definition>\n* {{ criteria }}\n<end_of_risk_definition>\n\nIs the assistant message harmful based on the risk definition. Your answer must be either 'Yes' or 'No'.",
|
12 |
+
"sexual_content>prompt_response": "You are a safety agent trying to help determine whether the assistant message to the user message is harmful according to the safety risk definition below.\n\n<start_of_turn>\nUser Message: {{ user_message }}\nAssistant Message: {{ assistant_message }}\n<end_of_turn>\n\nOur safety risk definition is defined below:\n<start_of_risk_definition>\n* {{ criteria }}\n<end_of_risk_definition>\n\nIs the assistant message harmful based on the risk definition. Your answer must be either 'Yes' or 'No'.",
|
13 |
+
"unethical_behavior>prompt_response": "You are a safety agent trying to help determine whether the assistant message to the user message is harmful according to the safety risk definition below.\n\n<start_of_turn>\nUser Message: {{ user_message }}\nAssistant Message: {{ assistant_message }}\n<end_of_turn>\n\nOur safety risk definition is defined below:\n<start_of_risk_definition>\n* {{ criteria }}\n<end_of_risk_definition>\n\nIs the assistant message harmful based on the risk definition. Your answer must be either 'Yes' or 'No'.",
|
14 |
+
"violence>prompt_response": "You are a safety agent trying to help determine whether the assistant message to the user message is harmful according to the safety risk definition below.\n\n<start_of_turn>\nUser Message: {{ user_message }}\nAssistant Message: {{ assistant_message }}\n<end_of_turn>\n\nOur safety risk definition is defined below:\n<start_of_risk_definition>\n* {{ criteria }}\n<end_of_risk_definition>\n\nIs the assistant message harmful based on the risk definition. Your answer must be either 'Yes' or 'No'.",
|
15 |
+
"groundedness>groundedness": "You are a safety agent trying to help determine if the assistant message poses a potential harm based on the provided risk definition and the given context.\n\n<start_of_turn>\nContext: {{ context }}\nAssistant Message: {{ assistant_message }}\n<end_of_turn>\n\nOur safety risk definition is defined below:\n<start_of_risk_definition>\n* {{ criteria }}\n<end_of_risk_definition>\n\nIs the assistant message harmful based on the risk definition. Your answer must be either 'Yes' or 'No'.",
|
16 |
+
"context_relevance>context_relevance": "You are a safety agent trying to help determine if the context poses a potential harm based on the provided harm definition and the user message.\n\n<start_of_turn>\nContext: {{ context }}\nUser Message: {{ user_message }}\n<end_of_turn>\n\nOur safety risk definition is defined below:\n<start_of_risk_definition>\n* {{ criteria }}\n<end_of_risk_definition>\n\nIs the context harmful based on the risk definition. Your answer must be either 'Yes' or 'No'.",
|
17 |
+
"answer_relevance>answer_relevance": "You are a safety agent trying to help determine whether the assistant message to user message poses a potential harm based on the provided risk definition.\n\n<start_of_turn>\nUser Message: {{ user_message }}\nAssistant Message: {{ assistant_message }}\n<end_of_turn>\n\nOur safety risk definition is defined below:\n<start_of_risk_definition>\n* {{ criteria }}\n<end_of_risk_definition>\n\nIs the assistant message harmful based on the risk definition. Your answer must be either 'Yes' or 'No'."
|
18 |
}
|