claude bugpass
This commit is contained in:
@@ -61,7 +61,7 @@ def _call_llm_sync(system_prompt, user_prompt):
|
||||
return extracted
|
||||
return None
|
||||
except Exception as e:
|
||||
print(f"LLM error: {type(e).__name__}: {e}", flush=True)
|
||||
print(f"LLM error ({AI_CONFIG['model']}): {type(e).__name__}: {e}", flush=True)
|
||||
return None
|
||||
|
||||
|
||||
@@ -119,7 +119,7 @@ async def parse(user_input, interaction_type, retry_count=0, errors=None, histor
|
||||
|
||||
response_text = await _call_llm(prompt_config["system"], user_prompt)
|
||||
if not response_text:
|
||||
return {"error": "AI service unavailable", "user_input": user_input}
|
||||
return {"error": f"AI service unavailable (model: {AI_CONFIG['model']})", "user_input": user_input}
|
||||
|
||||
try:
|
||||
parsed = json.loads(response_text)
|
||||
|
||||
Reference in New Issue
Block a user