diff --git a/app/backend/InputOutputHandler.tsx b/app/backend/InputOutputHandler.tsx index 1addec9..b630d9c 100644 --- a/app/backend/InputOutputHandler.tsx +++ b/app/backend/InputOutputHandler.tsx @@ -254,7 +254,7 @@ const InputOutputBackend: React.FC = () => { } setInputMessage("") const windowname = window.location.hostname - postWorkerRef.current.postMessage({ messages: [...messages, { role: "user", content: inputValue }], ai_model: "llama3.2", model_type: type, access_token: accessToken, api_key: api_key, windowname }) + postWorkerRef.current.postMessage({ messages: [...messages, { role: "user", content: inputValue }], ai_model: localStorage.getItem("model") || "llama3.2", model_type: type, access_token: accessToken, api_key: api_key, windowname }) startGetWorker() } } diff --git a/py/api.py b/py/api.py index f9d321a..70c848e 100644 --- a/py/api.py +++ b/py/api.py @@ -43,6 +43,8 @@ class API: model_type = data.get("model_type") ai_model = data.get("ai_model") access_token = data.get("access_token") + + print(model_type) if access_token not in self.ai_response: return jsonify({"status": 401, "error": "Invalid access token"}) @@ -56,6 +58,7 @@ class API: sleep(0.5) return jsonify({"status": 200}) elif model_type == "mistral": + print(model_type) api_key = data.get("api_key") thread = threading.Thread( target=self.ai.process_mistralai,