Skip to content

Commit 83cc533

Browse files
committed
documentation
1 parent 9f0cd91 commit 83cc533

File tree

2 files changed

+54
-25
lines changed

2 files changed

+54
-25
lines changed

evaluations/qa_quality_eval.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -93,7 +93,7 @@ def main():
9393
output_path="./qa_flow_quality_eval.json"
9494
)
9595
except Exception as e:
96-
print(f"An error occurred during evaluation: {e}\n Retrying without Azure AI Project.")
96+
print(f"An error occurred during evaluation: {e}\n Retrying without reporting results in Azure AI Project.")
9797
result = evaluate(
9898
evaluation_name=f"{prefix} Quality Evaluation",
9999
data=data,

evaluations/safety_eval.py

+53-24
Original file line numberDiff line numberDiff line change
@@ -83,19 +83,34 @@ async def main():
8383
adversarial_conversation_result = outputs.to_eval_qa_json_lines()
8484
print(f"Adversarial conversation results: {adversarial_conversation_result}.")
8585

86-
azure_ai_project["credential"] = ""
87-
adversarial_eval_result = evaluate(
88-
evaluation_name=f"{prefix} Adversarial Tests",
89-
data=adversarial_conversation_result,
90-
evaluators={
91-
"sexual": sexual_evaluator,
92-
"self_harm": self_harm_evaluator,
93-
"hate_unfairness": hate_unfairness_evaluator,
94-
"violence": violence_evaluator
95-
},
96-
azure_ai_project= azure_ai_project,
97-
output_path="./adversarial_test.json"
86+
try:
87+
azure_ai_project["credential"] = ""
88+
adversarial_eval_result = evaluate(
89+
evaluation_name=f"{prefix} Adversarial Tests",
90+
data=adversarial_conversation_result,
91+
evaluators={
92+
"sexual": sexual_evaluator,
93+
"self_harm": self_harm_evaluator,
94+
"hate_unfairness": hate_unfairness_evaluator,
95+
"violence": violence_evaluator
96+
},
97+
azure_ai_project= azure_ai_project,
98+
output_path="./adversarial_test.json"
9899
)
100+
except Exception as e:
101+
print(f"An error occurred during evaluation: {e}\n Retrying without reporting results in Azure AI Project.")
102+
adversarial_eval_result = evaluate(
103+
evaluation_name=f"{prefix} Adversarial Tests",
104+
data=adversarial_conversation_result,
105+
evaluators={
106+
"sexual": sexual_evaluator,
107+
"self_harm": self_harm_evaluator,
108+
"hate_unfairness": hate_unfairness_evaluator,
109+
"violence": violence_evaluator
110+
},
111+
output_path="./adversarial_test.json"
112+
)
113+
99114

100115
jb_outputs = await simulator(
101116
scenario=scenario,
@@ -106,18 +121,32 @@ async def main():
106121
adversarial_conversation_result_w_jailbreak = jb_outputs.to_eval_qa_json_lines()
107122
print(f"Adversarial conversation w/ jailbreak results: {adversarial_conversation_result_w_jailbreak}.")
108123

109-
adversarial_eval_w_jailbreak_result = evaluate(
110-
evaluation_name=f"{prefix} Adversarial Tests w/ Jailbreak",
111-
data=adversarial_conversation_result_w_jailbreak,
112-
evaluators={
113-
"sexual": sexual_evaluator,
114-
"self_harm": self_harm_evaluator,
115-
"hate_unfairness": hate_unfairness_evaluator,
116-
"violence": violence_evaluator
117-
},
118-
azure_ai_project=azure_ai_project,
119-
output_path="./adversarial_test_w_jailbreak.json"
120-
)
124+
try:
125+
adversarial_eval_w_jailbreak_result = evaluate(
126+
evaluation_name=f"{prefix} Adversarial Tests w/ Jailbreak",
127+
data=adversarial_conversation_result_w_jailbreak,
128+
evaluators={
129+
"sexual": sexual_evaluator,
130+
"self_harm": self_harm_evaluator,
131+
"hate_unfairness": hate_unfairness_evaluator,
132+
"violence": violence_evaluator
133+
},
134+
azure_ai_project=azure_ai_project,
135+
output_path="./adversarial_test_w_jailbreak.json"
136+
)
137+
except Exception as e:
138+
print(f"An error occurred during evaluation: {e}\n Retrying without reporting results in Azure AI Project.")
139+
adversarial_eval_w_jailbreak_result = evaluate(
140+
evaluation_name=f"{prefix} Adversarial Tests w/ Jailbreak",
141+
data=adversarial_conversation_result_w_jailbreak,
142+
evaluators={
143+
"sexual": sexual_evaluator,
144+
"self_harm": self_harm_evaluator,
145+
"hate_unfairness": hate_unfairness_evaluator,
146+
"violence": violence_evaluator
147+
},
148+
output_path="./adversarial_test_w_jailbreak.json"
149+
)
121150

122151
if __name__ == '__main__':
123152
import promptflow as pf

0 commit comments

Comments
 (0)