@@ -83,19 +83,34 @@ async def main():
83
83
adversarial_conversation_result = outputs .to_eval_qa_json_lines ()
84
84
print (f"Adversarial conversation results: { adversarial_conversation_result } ." )
85
85
86
- azure_ai_project ["credential" ] = ""
87
- adversarial_eval_result = evaluate (
88
- evaluation_name = f"{ prefix } Adversarial Tests" ,
89
- data = adversarial_conversation_result ,
90
- evaluators = {
91
- "sexual" : sexual_evaluator ,
92
- "self_harm" : self_harm_evaluator ,
93
- "hate_unfairness" : hate_unfairness_evaluator ,
94
- "violence" : violence_evaluator
95
- },
96
- azure_ai_project = azure_ai_project ,
97
- output_path = "./adversarial_test.json"
86
+ try :
87
+ azure_ai_project ["credential" ] = ""
88
+ adversarial_eval_result = evaluate (
89
+ evaluation_name = f"{ prefix } Adversarial Tests" ,
90
+ data = adversarial_conversation_result ,
91
+ evaluators = {
92
+ "sexual" : sexual_evaluator ,
93
+ "self_harm" : self_harm_evaluator ,
94
+ "hate_unfairness" : hate_unfairness_evaluator ,
95
+ "violence" : violence_evaluator
96
+ },
97
+ azure_ai_project = azure_ai_project ,
98
+ output_path = "./adversarial_test.json"
98
99
)
100
+ except Exception as e :
101
+ print (f"An error occurred during evaluation: { e } \n Retrying without reporting results in Azure AI Project." )
102
+ adversarial_eval_result = evaluate (
103
+ evaluation_name = f"{ prefix } Adversarial Tests" ,
104
+ data = adversarial_conversation_result ,
105
+ evaluators = {
106
+ "sexual" : sexual_evaluator ,
107
+ "self_harm" : self_harm_evaluator ,
108
+ "hate_unfairness" : hate_unfairness_evaluator ,
109
+ "violence" : violence_evaluator
110
+ },
111
+ output_path = "./adversarial_test.json"
112
+ )
113
+
99
114
100
115
jb_outputs = await simulator (
101
116
scenario = scenario ,
@@ -106,18 +121,32 @@ async def main():
106
121
adversarial_conversation_result_w_jailbreak = jb_outputs .to_eval_qa_json_lines ()
107
122
print (f"Adversarial conversation w/ jailbreak results: { adversarial_conversation_result_w_jailbreak } ." )
108
123
109
- adversarial_eval_w_jailbreak_result = evaluate (
110
- evaluation_name = f"{ prefix } Adversarial Tests w/ Jailbreak" ,
111
- data = adversarial_conversation_result_w_jailbreak ,
112
- evaluators = {
113
- "sexual" : sexual_evaluator ,
114
- "self_harm" : self_harm_evaluator ,
115
- "hate_unfairness" : hate_unfairness_evaluator ,
116
- "violence" : violence_evaluator
117
- },
118
- azure_ai_project = azure_ai_project ,
119
- output_path = "./adversarial_test_w_jailbreak.json"
120
- )
124
+ try :
125
+ adversarial_eval_w_jailbreak_result = evaluate (
126
+ evaluation_name = f"{ prefix } Adversarial Tests w/ Jailbreak" ,
127
+ data = adversarial_conversation_result_w_jailbreak ,
128
+ evaluators = {
129
+ "sexual" : sexual_evaluator ,
130
+ "self_harm" : self_harm_evaluator ,
131
+ "hate_unfairness" : hate_unfairness_evaluator ,
132
+ "violence" : violence_evaluator
133
+ },
134
+ azure_ai_project = azure_ai_project ,
135
+ output_path = "./adversarial_test_w_jailbreak.json"
136
+ )
137
+ except Exception as e :
138
+ print (f"An error occurred during evaluation: { e } \n Retrying without reporting results in Azure AI Project." )
139
+ adversarial_eval_w_jailbreak_result = evaluate (
140
+ evaluation_name = f"{ prefix } Adversarial Tests w/ Jailbreak" ,
141
+ data = adversarial_conversation_result_w_jailbreak ,
142
+ evaluators = {
143
+ "sexual" : sexual_evaluator ,
144
+ "self_harm" : self_harm_evaluator ,
145
+ "hate_unfairness" : hate_unfairness_evaluator ,
146
+ "violence" : violence_evaluator
147
+ },
148
+ output_path = "./adversarial_test_w_jailbreak.json"
149
+ )
121
150
122
151
if __name__ == '__main__' :
123
152
import promptflow as pf
0 commit comments