@@ -94,14 +94,13 @@ def evaluate(self, data: Dict) -> bool:
9494
9595 except Exception as e :
9696 Log .error (
97- "[SYNC EVALUATION ERROR]\n "
98- "------------------------------\n "
99- f"[INPUT DATA]: { data } \n \n "
100- f"[PROMPT]: { self .evaluator_config .evaluator_params .prompt } \n \n "
101- f"[FORMATTED MESSAGE]: { formatted_message } \n \n "
102- f"[LLM RESPONSE]: { llm_response } \n \n "
103- f"[JSON RESPONSE]: { json_response } \n \n "
104- f"[DECISION]: { decision } \n \n "
97+ "[SYNC EVALUATION ERROR]\t \t ->"
98+ f"[INPUT DATA]: { data } \t \t ->"
99+ f"[PROMPT]: { self .evaluator_config .evaluator_params .prompt } \t \t ->"
100+ f"[FORMATTED MESSAGE]: { formatted_message } \t \t ->"
101+ f"[LLM RESPONSE]: { llm_response } \t \t ->"
102+ f"[JSON RESPONSE]: { json_response } \t \t ->"
103+ f"[DECISION]: { decision } \t \t ->"
105104 f"[ERROR]: { e } "
106105 )
107106 raise
@@ -131,14 +130,13 @@ async def aevaluate(self, data: Dict) -> bool:
131130
132131 except Exception as e :
133132 Log .error (
134- "[ASYNC EVALUATION ERROR]\n "
135- "--------------------------------\n "
136- f"[INPUT DATA]: { data } \n \n "
137- f"[PROMPT]: { self .evaluator_config .evaluator_params .prompt } \n \n "
138- f"[FORMATTED MESSAGE]: { formatted_message } \n \n "
139- f"[LLM RESPONSE]: { llm_response } \n \n "
140- f"[JSON RESPONSE]: { json_response } \n \n "
141- f"[DECISION]: { decision } \n \n "
133+ "[ASYNC EVALUATION ERROR]\t \t ->"
134+ f"[INPUT DATA]: { data } \t \t ->"
135+ f"[PROMPT]: { self .evaluator_config .evaluator_params .prompt } \t \t ->"
136+ f"[FORMATTED MESSAGE]: { formatted_message } \t \t ->"
137+ f"[LLM RESPONSE]: { llm_response } \t \t ->"
138+ f"[JSON RESPONSE]: { json_response } \t \t ->"
139+ f"[DECISION]: { decision } \t \t ->"
142140 f"[ERROR]: { e } "
143141 )
144142 raise
0 commit comments