@@ -176,7 +176,7 @@ async def _get_llm_response(self, evaluation_prompt: str) -> LLMResponse:
176176 )
177177
178178 # Prepare the request
179- request_data = {
179+ request_data : dict [ str , Any ] = {
180180 "model" : model ,
181181 "messages" : [
182182 {"role" : "system" , "content" : system_prompt },
@@ -217,20 +217,34 @@ async def _get_llm_response(self, evaluation_prompt: str) -> LLMResponse:
217217 logger .error ("❌ LLM REQUEST FAILED" )
218218 logger .error ("=" * 80 )
219219 logger .error (f"Model: { model } " )
220- logger .error (f"API Endpoint: { 'Normalized API (/llm/api/chat/completions)' if not is_openai_model else 'OpenAI Passthrough (/llm/openai/deployments/...)' } " )
220+ logger .error (
221+ f"API Endpoint: { 'Normalized API (/llm/api/chat/completions)' if not is_openai_model else 'OpenAI Passthrough (/llm/openai/deployments/...)' } "
222+ )
221223 logger .error (f"Error Type: { type (e ).__name__ } " )
222224 logger .error (f"Error Message: { str (e )} " )
223225
224226 # Try to extract HTTP error details if available
225- if hasattr (e , 'response' ):
226- logger .error (f"HTTP Status Code: { e .response .status_code if hasattr (e .response , 'status_code' ) else 'N/A' } " )
227+ if hasattr (e , "response" ):
228+ logger .error (
229+ f"HTTP Status Code: { e .response .status_code if hasattr (e .response , 'status_code' ) else 'N/A' } "
230+ )
227231 try :
228- error_body = e .response .json () if hasattr (e .response , 'json' ) else str (e .response .content )
229- logger .error (f"Response Body: { json .dumps (error_body , indent = 2 ) if isinstance (error_body , dict ) else error_body } " )
230- except :
231- logger .error (f"Response Body: { str (e .response .content ) if hasattr (e .response , 'content' ) else 'N/A' } " )
232-
233- logger .error (f"Request Details:\n { json .dumps ({k : v for k , v in request_data .items () if k != 'messages' }, indent = 2 )} " )
232+ error_body = (
233+ e .response .json ()
234+ if hasattr (e .response , "json" )
235+ else str (e .response .content )
236+ )
237+ logger .error (
238+ f"Response Body: { json .dumps (error_body , indent = 2 ) if isinstance (error_body , dict ) else error_body } "
239+ )
240+ except Exception :
241+ logger .error (
242+ f"Response Body: { str (e .response .content ) if hasattr (e .response , 'content' ) else 'N/A' } "
243+ )
244+
245+ logger .error (
246+ f"Request Details:\n { json .dumps ({k : v for k , v in request_data .items () if k != 'messages' }, indent = 2 )} "
247+ )
234248 logger .error ("=" * 80 )
235249
236250 raise UiPathEvaluationError (
@@ -284,7 +298,9 @@ async def _get_llm_response(self, evaluation_prompt: str) -> LLMResponse:
284298 content_str = content_str [start :end ]
285299
286300 parsed_response = json .loads (content_str )
287- logger .info (f"📊 Parsed evaluation score: { parsed_response .get ('score' , 'N/A' )} " )
301+ logger .info (
302+ f"📊 Parsed evaluation score: { parsed_response .get ('score' , 'N/A' )} "
303+ )
288304 except Exception as e :
289305 # Enhanced JSON parsing error logging
290306 logger .error ("=" * 80 )
0 commit comments