Merge pull request 'main' (#59) from React-Group/interstellar_ai:main into main
Reviewed-on: https://interstellardevelopment.org/code/code/YasinOnm08/interstellar_ai/pulls/59
This commit is contained in:
commit
1e4dda33ad
5 changed files with 79 additions and 185 deletions
|
@ -1,8 +1,21 @@
|
||||||
"use client";
|
"use client";
|
||||||
import React, { useState, useEffect } from 'react';
|
import React, { useState, useEffect } from 'react';
|
||||||
|
|
||||||
// Define all models that should be available.
|
// Define the types for ModelType and ModelListType
|
||||||
const modelList = {
|
type ModelType = {
|
||||||
|
model_type: string;
|
||||||
|
Math: string;
|
||||||
|
Code: string;
|
||||||
|
Language: string;
|
||||||
|
Weather: string;
|
||||||
|
};
|
||||||
|
|
||||||
|
type ModelListType = {
|
||||||
|
[key: string]: ModelType;
|
||||||
|
};
|
||||||
|
|
||||||
|
// Define the AI models list
|
||||||
|
const modelList: ModelListType = {
|
||||||
'Offline Fast': {
|
'Offline Fast': {
|
||||||
model_type: 'local',
|
model_type: 'local',
|
||||||
Math: 'qwen2-math:1.5b',
|
Math: 'qwen2-math:1.5b',
|
||||||
|
@ -86,161 +99,47 @@ const modelList = {
|
||||||
Code: 'open-codestral-mamba',
|
Code: 'open-codestral-mamba',
|
||||||
Language: 'open-mistral-nemo',
|
Language: 'open-mistral-nemo',
|
||||||
Weather: 'open-mistral-nemo',
|
Weather: 'open-mistral-nemo',
|
||||||
}
|
},
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
// Define the available selectedAIFunction options
|
|
||||||
const modelDropdown = {
|
|
||||||
offlineNonFoss: ['Offline Fast', 'Offline Slow'],
|
|
||||||
offlineFoss: ['Offline Fast (FOSS)', 'Offline Slow (FOSS)'],
|
|
||||||
onlineNonFoss: [
|
|
||||||
'Online Cheap (OpenAI)',
|
|
||||||
'Online Expensive (OpenAI)',
|
|
||||||
'Online Cheap (Anthropic)',
|
|
||||||
'Online Expensive (Anthropic)',
|
|
||||||
'Online Cheap (Google)',
|
|
||||||
'Online Expensive (Google)',
|
|
||||||
'Online (La Plateforme)'
|
|
||||||
],
|
|
||||||
onlineFoss: ['Online (FOSS) (La Plateforme)'],
|
|
||||||
};
|
};
|
||||||
|
|
||||||
const selectedAIFunction = [
|
// AI Functions list
|
||||||
'Code',
|
const aiFunctions = ['Math', 'Code', 'Language', 'Weather'] as const;
|
||||||
'Math',
|
type AIFunction = typeof aiFunctions[number]; // Restrict to these exact function names
|
||||||
'Language',
|
|
||||||
'Weather'
|
|
||||||
]
|
|
||||||
|
|
||||||
const ModelSection: React.FC = () => {
|
const ModelSection: React.FC = () => {
|
||||||
// Initialize state with value from localStorage or default to ''
|
const [selectedModelDropdown, setSelectedModelDropdown] = useState(() => "Offline Fast");
|
||||||
const [selectedModelDropdown, setSelectedModelDropdown] = useState('');
|
const [activeSelectedAIFunction, setActiveSelectedAIFunction] = useState(() => "Code");
|
||||||
const [radioSelection, setRadioSelection] = useState<string | null>("")
|
const [, setSelectedModel] = useState<string>("");
|
||||||
const [activeSelectedAIFunction, setActiveSelectedAIFunction] = useState('');
|
const [, setSelectedModelType] = useState<string>("");
|
||||||
const [currentSelectedAIFunction, setCurrentSelectedAIFunction] = useState<string | null>("");
|
|
||||||
const [isOpenSourceMode, setIsOpenSourceMode] = useState<string|null>("false")
|
|
||||||
|
|
||||||
useEffect(() => {
|
useEffect(() => {
|
||||||
if (typeof localStorage !== 'undefined') {
|
setSelectedModelDropdown(localStorage.getItem("selectedModelDropdown")|| "Offline Fast");
|
||||||
const defaultValues = {
|
setActiveSelectedAIFunction(localStorage.getItem("activeSelectedAIFunction") || "Code");
|
||||||
selectedModelDropdown: 'Offline Fast',
|
|
||||||
activeSelectedAIFunction: 'Code',
|
|
||||||
model: 'starcoder2',
|
|
||||||
radioSelection: 'None',
|
|
||||||
type: 'local',
|
|
||||||
};
|
|
||||||
|
|
||||||
Object.entries(defaultValues).forEach(([key, value]) => {
|
|
||||||
if (!localStorage.getItem(key)) {
|
|
||||||
localStorage.setItem(key, value);
|
|
||||||
}
|
|
||||||
});
|
|
||||||
|
|
||||||
setIsOpenSourceMode(localStorage.getItem("openSourceMode"));
|
|
||||||
setActiveSelectedAIFunction(localStorage.getItem("activeSelectedAIFunction") || '');
|
|
||||||
setRadioSelection(localStorage.getItem("radioSelection") || '');
|
|
||||||
setSelectedModelDropdown(localStorage.getItem('selectedModelDropdown') || '');
|
|
||||||
|
|
||||||
const handleStorageChange = () => {
|
|
||||||
setSelectedModelDropdown(localStorage.getItem('selectedModelDropdown') || '');
|
|
||||||
};
|
|
||||||
|
|
||||||
// Update immediately when localStorage changes
|
|
||||||
window.addEventListener('storage', handleStorageChange);
|
|
||||||
|
|
||||||
// Cleanup listener on component unmount
|
|
||||||
return () => {
|
|
||||||
window.removeEventListener('storage', handleStorageChange);
|
|
||||||
};
|
|
||||||
}
|
|
||||||
}, []);
|
}, []);
|
||||||
|
|
||||||
|
|
||||||
|
// Update the model and type when the dropdown or function changes
|
||||||
useEffect(() => {
|
useEffect(() => {
|
||||||
if (typeof localStorage !== 'undefined') {
|
const aiFunctionsActiveSelectedAIFunction = activeSelectedAIFunction as AIFunction
|
||||||
const storedActiveSelectedAIFunction = localStorage.getItem("activeSelectedAIFunction") || "";
|
const newSelectedModel = modelList[selectedModelDropdown]?.[aiFunctionsActiveSelectedAIFunction] || "";
|
||||||
if (storedActiveSelectedAIFunction !== currentSelectedAIFunction) {
|
const newModelType = modelList[selectedModelDropdown]?.model_type || "";
|
||||||
setCurrentSelectedAIFunction(storedActiveSelectedAIFunction);
|
|
||||||
}
|
setSelectedModel(newSelectedModel);
|
||||||
}
|
setSelectedModelType(newModelType);
|
||||||
}, [activeSelectedAIFunction]);
|
|
||||||
|
localStorage.setItem("model", newSelectedModel);
|
||||||
|
localStorage.setItem("type", newModelType);
|
||||||
|
}, [selectedModelDropdown, activeSelectedAIFunction]);
|
||||||
|
|
||||||
const handleModelChange = (event: React.ChangeEvent<HTMLSelectElement>) => {
|
const handleModelChange = (event: React.ChangeEvent<HTMLSelectElement>) => {
|
||||||
const newModel = event.target.value;
|
const newModel = event.target.value;
|
||||||
setSelectedModelDropdown(newModel);
|
setSelectedModelDropdown(newModel);
|
||||||
if (typeof localStorage !== 'undefined') {
|
localStorage.setItem('selectedModelDropdown', newModel);
|
||||||
localStorage.setItem('selectedModelDropdown', newModel); // Update localStorage directly
|
|
||||||
const model = localStorage.getItem('activeSelectedAIFunction') || "Code"
|
|
||||||
modelClicked(model)
|
|
||||||
}
|
|
||||||
};
|
};
|
||||||
|
|
||||||
// Determine the filtered models based on current radioSelection
|
const modelClicked = (functionName: AIFunction) => {
|
||||||
const filteredModels = (() => {
|
setActiveSelectedAIFunction(functionName);
|
||||||
let models = [];
|
localStorage.setItem('activeSelectedAIFunction', functionName);
|
||||||
switch (radioSelection) {
|
};
|
||||||
case 'Offline':
|
|
||||||
models = [
|
|
||||||
...modelDropdown.onlineNonFoss,
|
|
||||||
...modelDropdown.onlineFoss,
|
|
||||||
];
|
|
||||||
if (isOpenSourceMode == "true") {
|
|
||||||
models = [
|
|
||||||
...modelDropdown.onlineFoss,
|
|
||||||
];
|
|
||||||
} // Show only offline models without FOSS
|
|
||||||
break;
|
|
||||||
case 'Online':
|
|
||||||
models = [
|
|
||||||
...modelDropdown.offlineNonFoss,
|
|
||||||
...modelDropdown.offlineFoss,
|
|
||||||
];
|
|
||||||
if (isOpenSourceMode == "true") {
|
|
||||||
models = [
|
|
||||||
...modelDropdown.offlineFoss,
|
|
||||||
];
|
|
||||||
} // Show only online models without FOSS
|
|
||||||
break;
|
|
||||||
case 'None':
|
|
||||||
models = [
|
|
||||||
...modelDropdown.offlineNonFoss,
|
|
||||||
...modelDropdown.offlineFoss,
|
|
||||||
...modelDropdown.onlineNonFoss,
|
|
||||||
...modelDropdown.onlineFoss,
|
|
||||||
];
|
|
||||||
if (isOpenSourceMode == "true") {
|
|
||||||
models = [
|
|
||||||
...modelDropdown.offlineFoss,
|
|
||||||
...modelDropdown.onlineFoss,
|
|
||||||
];
|
|
||||||
} // Show all models if nothing matches
|
|
||||||
break;
|
|
||||||
default:
|
|
||||||
models = [
|
|
||||||
...modelDropdown.offlineNonFoss,
|
|
||||||
...modelDropdown.offlineFoss,
|
|
||||||
...modelDropdown.onlineNonFoss,
|
|
||||||
...modelDropdown.onlineFoss,
|
|
||||||
]; // Show all models if nothing matches
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
return Array.from(new Set(models)); // Remove duplicates using Set
|
|
||||||
})();
|
|
||||||
|
|
||||||
const isOfflineModel = (model: string) =>
|
|
||||||
modelDropdown.offlineNonFoss.includes(model) || modelDropdown.offlineFoss.includes(model);
|
|
||||||
|
|
||||||
const modelClicked = (model: string) => {
|
|
||||||
if (typeof localStorage !== 'undefined') {
|
|
||||||
localStorage.setItem('activeSelectedAIFunction', model)
|
|
||||||
setActiveSelectedAIFunction(model)
|
|
||||||
const modelDropdown = localStorage.getItem('selectedModelDropdown') || 'Offline Fast'
|
|
||||||
const selectedAIFunction = modelDropdown as keyof typeof modelList;
|
|
||||||
localStorage.setItem("model", modelList[selectedAIFunction][model as keyof typeof modelList[typeof selectedAIFunction]])
|
|
||||||
localStorage.setItem("type", modelList[selectedAIFunction]['model_type' as keyof typeof modelList[typeof selectedAIFunction]])
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
return (
|
return (
|
||||||
<div className="model-background">
|
<div className="model-background">
|
||||||
|
@ -249,33 +148,29 @@ const ModelSection: React.FC = () => {
|
||||||
<h3>Different AI Models</h3>
|
<h3>Different AI Models</h3>
|
||||||
</div>
|
</div>
|
||||||
|
|
||||||
{/* Model Selection Dropdown */}
|
|
||||||
<div className="model-dropdown">
|
<div className="model-dropdown">
|
||||||
<label htmlFor="model-select">Select AI Model:</label>
|
<label htmlFor="model-select">Select AI Model:</label>
|
||||||
<select id="model-select" value={selectedModelDropdown} onChange={handleModelChange}>
|
<select id="model-select" value={selectedModelDropdown} onChange={handleModelChange}>
|
||||||
{filteredModels.map((model) => (
|
{Object.keys(modelList).map((model) => (
|
||||||
<option key={model} value={model}>
|
<option key={model} value={model}>
|
||||||
{model}
|
{model}
|
||||||
</option>
|
</option>
|
||||||
))}
|
))}
|
||||||
</select>
|
</select>
|
||||||
</div>
|
</div>
|
||||||
{/* Model Grid with Cards */}
|
|
||||||
<div className="grid">
|
<div className="grid">
|
||||||
{selectedAIFunction.map(
|
{aiFunctions.map((functionName) => (
|
||||||
(displayedCategory) => (
|
<button
|
||||||
<button
|
key={functionName}
|
||||||
key={displayedCategory}
|
className={`${functionName.toLowerCase()}-model model-box ${activeSelectedAIFunction === functionName ? 'selected' : ''}`}
|
||||||
className={`${displayedCategory.toLowerCase()}-model model-box ${currentSelectedAIFunction === displayedCategory ? 'selected' : ''}`}
|
onClick={() => modelClicked(functionName)}
|
||||||
onClick={() => modelClicked(displayedCategory)}
|
>
|
||||||
>
|
<div className="overlay">
|
||||||
<div className="overlay">
|
<h3>{functionName}</h3>
|
||||||
<h3>{displayedCategory}</h3>
|
</div>
|
||||||
{isOfflineModel(selectedModelDropdown) && <img src="/img/nowifi.svg" alt="No Wi-Fi" />}
|
</button>
|
||||||
</div>
|
))}
|
||||||
</button>
|
|
||||||
)
|
|
||||||
)}
|
|
||||||
</div>
|
</div>
|
||||||
</div>
|
</div>
|
||||||
</div>
|
</div>
|
||||||
|
|
|
@ -12,9 +12,11 @@ import PrivacySettings from './PrivacySettings';
|
||||||
import FontSizeSetting from './FontSize';
|
import FontSizeSetting from './FontSize';
|
||||||
import OpenSourceModeToggle from './OpenSourceToggle';
|
import OpenSourceModeToggle from './OpenSourceToggle';
|
||||||
import {
|
import {
|
||||||
|
changeHistory,
|
||||||
changeSettings,
|
changeSettings,
|
||||||
createAccount,
|
createAccount,
|
||||||
deleteAccount,
|
deleteAccount,
|
||||||
|
getHistory,
|
||||||
} from '../../backend/database';
|
} from '../../backend/database';
|
||||||
import ThemeDropdown from './DropDownTheme';
|
import ThemeDropdown from './DropDownTheme';
|
||||||
|
|
||||||
|
@ -372,7 +374,11 @@ const Settings: React.FC<{ closeSettings: () => void; accountName: string }> = (
|
||||||
localStorage.setItem("currentEmail", useEmail)
|
localStorage.setItem("currentEmail", useEmail)
|
||||||
alert('Account successfully changed!')
|
alert('Account successfully changed!')
|
||||||
window.location.reload()
|
window.location.reload()
|
||||||
|
} else {
|
||||||
|
alert("failed to send settings")
|
||||||
}
|
}
|
||||||
|
} else {
|
||||||
|
alert("failed to create account")
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
|
@ -409,7 +415,7 @@ const Settings: React.FC<{ closeSettings: () => void; accountName: string }> = (
|
||||||
checked={myBoolean}
|
checked={myBoolean}
|
||||||
setChecked={setMyBoolean}
|
setChecked={setMyBoolean}
|
||||||
/>
|
/>
|
||||||
<TextSettings
|
<TextSettings
|
||||||
label="Nearest City"
|
label="Nearest City"
|
||||||
value={weatherInfo}
|
value={weatherInfo}
|
||||||
type='text'
|
type='text'
|
||||||
|
@ -624,6 +630,7 @@ const Settings: React.FC<{ closeSettings: () => void; accountName: string }> = (
|
||||||
onClick={handleLogout} // Function to call on click
|
onClick={handleLogout} // Function to call on click
|
||||||
className="update-credentials-button" // Custom styling class
|
className="update-credentials-button" // Custom styling class
|
||||||
/>
|
/>
|
||||||
|
<p>WARNING: Will delete your chat history.</p>
|
||||||
<ButtonSetting
|
<ButtonSetting
|
||||||
label="Update Credentials" // Button label
|
label="Update Credentials" // Button label
|
||||||
onClick={handleUpdateCredentials} // Function to call on click
|
onClick={handleUpdateCredentials} // Function to call on click
|
||||||
|
|
|
@ -49,8 +49,8 @@
|
||||||
border-radius: 5%;
|
border-radius: 5%;
|
||||||
overflow: hidden;
|
overflow: hidden;
|
||||||
position: relative;
|
position: relative;
|
||||||
height: 18vh;
|
height: 8vw;
|
||||||
width: 18vh;
|
width: 8vw;
|
||||||
margin: auto; /* Center each model box in the grid cell */
|
margin: auto; /* Center each model box in the grid cell */
|
||||||
}
|
}
|
||||||
.model-box.selected {
|
.model-box.selected {
|
||||||
|
|
|
@ -57,8 +57,8 @@
|
||||||
}
|
}
|
||||||
/* Model box styles */
|
/* Model box styles */
|
||||||
.model-box {
|
.model-box {
|
||||||
max-width: none; /* Remove max-width */
|
width: 50vw;
|
||||||
margin: 0 auto; /* Center each model-box */
|
height: 50vw;
|
||||||
}
|
}
|
||||||
/* Input styles */
|
/* Input styles */
|
||||||
.input {
|
.input {
|
||||||
|
|
34
py/ai.py
34
py/ai.py
|
@ -20,34 +20,29 @@ class AI:
|
||||||
|
|
||||||
for chunk in stream:
|
for chunk in stream:
|
||||||
with return_class.ai_response_lock:
|
with return_class.ai_response_lock:
|
||||||
return_class.ai_response[access_token] += chunk['message']['content']
|
return_class.ai_response[access_token] += chunk["message"]["content"]
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def process_mistralai(model, messages, return_class, access_token, api_key):
|
def process_mistralai(model, messages, return_class, access_token, api_key):
|
||||||
|
|
||||||
client = Mistral(api_key=api_key)
|
client = Mistral(api_key=api_key)
|
||||||
|
|
||||||
stream_response = client.chat.stream(
|
stream_response = client.chat.stream(model=model, messages=messages)
|
||||||
model=model,
|
|
||||||
messages=messages
|
|
||||||
)
|
|
||||||
|
|
||||||
with return_class.ai_response_lock:
|
with return_class.ai_response_lock:
|
||||||
return_class.ai_response[access_token] = ""
|
return_class.ai_response[access_token] = ""
|
||||||
|
|
||||||
for chunk in stream_response:
|
for chunk in stream_response:
|
||||||
with return_class.ai_response_lock:
|
with return_class.ai_response_lock:
|
||||||
return_class.ai_response[access_token] += chunk.data.choices[0].delta.content
|
return_class.ai_response[access_token] += chunk.data.choices[
|
||||||
|
0
|
||||||
|
].delta.content
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def process_openai(model, messages, return_class, access_token, api_key):
|
def process_openai(model, messages, return_class, access_token, api_key):
|
||||||
|
|
||||||
client = OpenAI(api_key=api_key)
|
client = OpenAI(api_key=api_key)
|
||||||
|
|
||||||
stream_response = client.chat.completions.create(
|
stream_response = client.chat.completions.create(
|
||||||
model=model,
|
model=model, messages=messages, stream=True
|
||||||
messages=messages,
|
|
||||||
stream=True
|
|
||||||
)
|
)
|
||||||
|
|
||||||
with return_class.ai_response_lock:
|
with return_class.ai_response_lock:
|
||||||
|
@ -59,16 +54,15 @@ class AI:
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def process_anthropic(model, messages, return_class, access_token, api_key):
|
def process_anthropic(model, messages, return_class, access_token, api_key):
|
||||||
|
|
||||||
client = anthropic.Anthropic(api_key=api_key)
|
client = anthropic.Anthropic(api_key=api_key)
|
||||||
|
|
||||||
with return_class.ai_response_lock:
|
with return_class.ai_response_lock:
|
||||||
return_class.ai_response[access_token] = ""
|
return_class.ai_response[access_token] = ""
|
||||||
|
|
||||||
with client.messages.stream(
|
with client.messages.stream(
|
||||||
max_tokens=1024,
|
max_tokens=1024,
|
||||||
model=model,
|
model=model,
|
||||||
messages=messages,
|
messages=messages,
|
||||||
) as stream:
|
) as stream:
|
||||||
for text in stream.text_stream:
|
for text in stream.text_stream:
|
||||||
with return_class.ai_response_lock:
|
with return_class.ai_response_lock:
|
||||||
|
@ -76,16 +70,15 @@ class AI:
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def process_google(model, messages, return_class, access_token, api_key):
|
def process_google(model, messages, return_class, access_token, api_key):
|
||||||
|
message = messages[-1]["content"]
|
||||||
message = messages[-1]['content']
|
|
||||||
messages.pop()
|
messages.pop()
|
||||||
|
|
||||||
for msg in messages:
|
for msg in messages:
|
||||||
msg['parts'] = msg.pop('content')
|
msg["parts"] = msg.pop()["content"]
|
||||||
|
|
||||||
for msg in messages:
|
for msg in messages:
|
||||||
if msg['role'] == 'assistant':
|
if msg["role"] == "assistant":
|
||||||
msg['role'] = 'model'
|
msg["role"] = "model"
|
||||||
|
|
||||||
genai.configure(api_key=api_key)
|
genai.configure(api_key=api_key)
|
||||||
|
|
||||||
|
@ -93,7 +86,6 @@ class AI:
|
||||||
|
|
||||||
chat = model.start_chat(
|
chat = model.start_chat(
|
||||||
history=messages,
|
history=messages,
|
||||||
|
|
||||||
)
|
)
|
||||||
|
|
||||||
response = chat.send_message(message, stream=True)
|
response = chat.send_message(message, stream=True)
|
||||||
|
|
Loading…
Reference in a new issue