forked from React-Group/interstellar_ai
Compare commits
2 commits
2650fb8205
...
8ee87b88e9
Author | SHA1 | Date | |
---|---|---|---|
8ee87b88e9 | |||
|
32edaaef42 |
3 changed files with 11 additions and 2 deletions
|
@ -49,7 +49,7 @@ const InputOutputBackend: React.FC = () => {
|
||||||
Additionally, under no circumstances ever translate your answer into multiple languages.`,
|
Additionally, under no circumstances ever translate your answer into multiple languages.`,
|
||||||
},
|
},
|
||||||
{ role: "assistant", content: "Hello! How may I help you?" },
|
{ role: "assistant", content: "Hello! How may I help you?" },
|
||||||
]);
|
]);
|
||||||
}
|
}
|
||||||
}, [preferredCurrency, preferredLanguage, timeFormat, dateFormat, preferredMeasurement, timeZone]);
|
}, [preferredCurrency, preferredLanguage, timeFormat, dateFormat, preferredMeasurement, timeZone]);
|
||||||
|
|
||||||
|
|
|
@ -162,6 +162,8 @@ const Models: React.FC = () => {
|
||||||
setRadioSelection(localStorage.getItem('radioSelection'))
|
setRadioSelection(localStorage.getItem('radioSelection'))
|
||||||
const handleStorageChange = () => {
|
const handleStorageChange = () => {
|
||||||
setSelectedModel(localStorage.getItem('selectedModel') || '');
|
setSelectedModel(localStorage.getItem('selectedModel') || '');
|
||||||
|
console.log("Changed the selectedModel")
|
||||||
|
console.log(selectedModel)
|
||||||
};
|
};
|
||||||
handleStorageChange();
|
handleStorageChange();
|
||||||
|
|
||||||
|
@ -177,7 +179,7 @@ const Models: React.FC = () => {
|
||||||
const handleModelChange = (event: React.ChangeEvent<HTMLSelectElement>) => {
|
const handleModelChange = (event: React.ChangeEvent<HTMLSelectElement>) => {
|
||||||
const newModel = event.target.value;
|
const newModel = event.target.value;
|
||||||
setSelectedModel(newModel);
|
setSelectedModel(newModel);
|
||||||
localStorage.setItem('radioSelection', newModel); // Update localStorage directly
|
localStorage.setItem('selectedModel', newModel); // Update localStorage directly
|
||||||
};
|
};
|
||||||
|
|
||||||
// Determine the filtered models based on current radioSelection
|
// Determine the filtered models based on current radioSelection
|
||||||
|
|
|
@ -1,3 +1,5 @@
|
||||||
|
from time import sleep
|
||||||
|
|
||||||
from flask import Flask, request, jsonify
|
from flask import Flask, request, jsonify
|
||||||
from flask_cors import CORS
|
from flask_cors import CORS
|
||||||
import secrets
|
import secrets
|
||||||
|
@ -48,6 +50,7 @@ class API:
|
||||||
thread = threading.Thread(target=self.ai.process_local, args=(ai_model, messages, self, access_token))
|
thread = threading.Thread(target=self.ai.process_local, args=(ai_model, messages, self, access_token))
|
||||||
thread.start()
|
thread.start()
|
||||||
thread.join()
|
thread.join()
|
||||||
|
sleep(0.5)
|
||||||
return jsonify({'status': 200})
|
return jsonify({'status': 200})
|
||||||
elif model_type == "mistral":
|
elif model_type == "mistral":
|
||||||
api_key = data.get('api_key')
|
api_key = data.get('api_key')
|
||||||
|
@ -55,6 +58,7 @@ class API:
|
||||||
args=(ai_model, messages, self, access_token, api_key))
|
args=(ai_model, messages, self, access_token, api_key))
|
||||||
thread.start()
|
thread.start()
|
||||||
thread.join()
|
thread.join()
|
||||||
|
sleep(0.5)
|
||||||
return jsonify({'status': 200})
|
return jsonify({'status': 200})
|
||||||
elif model_type == "openai":
|
elif model_type == "openai":
|
||||||
api_key = data.get('api_key')
|
api_key = data.get('api_key')
|
||||||
|
@ -62,6 +66,7 @@ class API:
|
||||||
args=(ai_model, messages, self, access_token, api_key))
|
args=(ai_model, messages, self, access_token, api_key))
|
||||||
thread.start()
|
thread.start()
|
||||||
thread.join()
|
thread.join()
|
||||||
|
sleep(0.5)
|
||||||
return jsonify({'status': 200})
|
return jsonify({'status': 200})
|
||||||
elif model_type == "anthropic":
|
elif model_type == "anthropic":
|
||||||
api_key = data.get('api_key')
|
api_key = data.get('api_key')
|
||||||
|
@ -69,6 +74,7 @@ class API:
|
||||||
args=(ai_model, messages, self, access_token, api_key))
|
args=(ai_model, messages, self, access_token, api_key))
|
||||||
thread.start()
|
thread.start()
|
||||||
thread.join()
|
thread.join()
|
||||||
|
sleep(0.5)
|
||||||
return jsonify({'status': 200})
|
return jsonify({'status': 200})
|
||||||
elif model_type == "google":
|
elif model_type == "google":
|
||||||
api_key = data.get('api_key')
|
api_key = data.get('api_key')
|
||||||
|
@ -76,6 +82,7 @@ class API:
|
||||||
args=(ai_model, messages, self, access_token, api_key))
|
args=(ai_model, messages, self, access_token, api_key))
|
||||||
thread.start()
|
thread.start()
|
||||||
thread.join()
|
thread.join()
|
||||||
|
sleep(0.5)
|
||||||
return jsonify({'status': 200})
|
return jsonify({'status': 200})
|
||||||
|
|
||||||
return jsonify({'status': 401, 'error': 'Invalid AI model type'})
|
return jsonify({'status': 401, 'error': 'Invalid AI model type'})
|
||||||
|
|
Loading…
Reference in a new issue