From c226fbd34b5a994c85a25e5b11fdf88ca16bd02f Mon Sep 17 00:00:00 2001 From: ronalds Date: Sun, 30 Nov 2025 23:35:39 -0300 Subject: [PATCH] =?UTF-8?q?=E2=9C=85=20Corregido=20Error=20de=20Chat=20IA?= =?UTF-8?q?=20(Backend=20v1.0.86)=20Problema=20Causa=20El=20modelo=20de=20?= =?UTF-8?q?base=20de=20datos=20usa=20model=5Fname=20pero=20el=20c=C3=B3dig?= =?UTF-8?q?o=20intentaba=20acceder=20a=20model.?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Solución ✅ Cambiado ai_config.model → ai_config.model_name en: Llamada a OpenAI: model=ai_config.model_name or "gpt-4" Llamada a Gemini: GenerativeModel(ai_config.model_name or 'gemini-pro') Response del endpoint: "model": ai_config.model_name --- backend/app/main.py | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-) diff --git a/backend/app/main.py b/backend/app/main.py index 65b0c2d..11a0693 100644 --- a/backend/app/main.py +++ b/backend/app/main.py @@ -204,7 +204,7 @@ def send_completed_inspection_to_n8n(inspection, db): # No lanzamos excepción para no interrumpir el flujo normal -BACKEND_VERSION = "1.0.85" +BACKEND_VERSION = "1.0.86" app = FastAPI(title="Checklist Inteligente API", version=BACKEND_VERSION) # S3/MinIO configuration @@ -2965,7 +2965,7 @@ FORMATO DE RESPUESTA: client = OpenAI(api_key=ai_config.api_key) response = client.chat.completions.create( - model=ai_config.model or "gpt-4", + model=ai_config.model_name or "gpt-4", messages=messages, max_tokens=max_tokens, temperature=0.7 @@ -2978,7 +2978,7 @@ FORMATO DE RESPUESTA: import google.generativeai as genai genai.configure(api_key=ai_config.api_key) - model = genai.GenerativeModel(ai_config.model or 'gemini-pro') + model = genai.GenerativeModel(ai_config.model_name or 'gemini-pro') # Gemini maneja el chat diferente # Convertir mensajes al formato de Gemini @@ -3003,7 +3003,7 @@ FORMATO DE RESPUESTA: "response": ai_response, "confidence": confidence, "provider": ai_config.provider, - "model": ai_config.model + "model": ai_config.model_name } except Exception as e: