Compare commits
No commits in common. "d0f27295d763307ca148473a4bc5365f404839e7" and "a2254fcc7f37380dbb398cf5f1ab33027443a7df" have entirely different histories.
d0f27295d7
...
a2254fcc7f
3 changed files with 13 additions and 44 deletions
26
py/ai.py
26
py/ai.py
|
@ -1,5 +1,4 @@
|
||||||
from mistralai import Mistral
|
from mistralai import Mistral
|
||||||
from openai import OpenAI
|
|
||||||
import ollama
|
import ollama
|
||||||
|
|
||||||
|
|
||||||
|
@ -13,15 +12,15 @@ class AI:
|
||||||
options={"temperature": 0.5},
|
options={"temperature": 0.5},
|
||||||
)
|
)
|
||||||
|
|
||||||
with return_class.ai_response_lock:
|
|
||||||
return_class.ai_response[access_token] = ""
|
return_class.ai_response[access_token] = ""
|
||||||
|
|
||||||
for chunk in stream:
|
for chunk in stream:
|
||||||
with return_class.ai_response_lock:
|
|
||||||
return_class.ai_response[access_token] += chunk['message']['content']
|
return_class.ai_response[access_token] += chunk['message']['content']
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def process_mistralai(model, messages, return_class, access_token, api_key):
|
def process_mistralai(model, messages, return_class, access_token):
|
||||||
|
with open("api_key.txt", 'r') as f:
|
||||||
|
api_key = f.read().strip()
|
||||||
|
|
||||||
client = Mistral(api_key=api_key)
|
client = Mistral(api_key=api_key)
|
||||||
|
|
||||||
|
@ -30,26 +29,7 @@ class AI:
|
||||||
messages=messages
|
messages=messages
|
||||||
)
|
)
|
||||||
|
|
||||||
with return_class.ai_response_lock:
|
|
||||||
return_class.ai_response[access_token] = ""
|
return_class.ai_response[access_token] = ""
|
||||||
|
|
||||||
for chunk in stream_response:
|
for chunk in stream_response:
|
||||||
with return_class.ai_response_lock:
|
|
||||||
return_class.ai_response[access_token] += chunk.data.choices[0].delta.content
|
return_class.ai_response[access_token] += chunk.data.choices[0].delta.content
|
||||||
|
|
||||||
@staticmethod
|
|
||||||
def process_openai(model, messages, return_class, access_token, api_key):
|
|
||||||
|
|
||||||
client = OpenAI(api_key=api_key)
|
|
||||||
|
|
||||||
stream_response = client.chat.completions.create(
|
|
||||||
model=model,
|
|
||||||
messages=messages
|
|
||||||
)
|
|
||||||
|
|
||||||
with return_class.ai_response_lock:
|
|
||||||
return_class.ai_response[access_token] = ""
|
|
||||||
|
|
||||||
for chunk in stream_response:
|
|
||||||
with return_class.ai_response_lock:
|
|
||||||
return_class.ai_response[access_token] += chunk.choices[0].delta.content
|
|
20
py/api.py
20
py/api.py
|
@ -1,7 +1,6 @@
|
||||||
from flask import Flask, request, jsonify
|
from flask import Flask, request, jsonify
|
||||||
from flask_cors import CORS
|
from flask_cors import CORS
|
||||||
import secrets
|
import secrets
|
||||||
import threading
|
|
||||||
from ai import AI
|
from ai import AI
|
||||||
from db import DB
|
from db import DB
|
||||||
from OpenSSL import crypto
|
from OpenSSL import crypto
|
||||||
|
@ -9,13 +8,12 @@ from OpenSSL import crypto
|
||||||
|
|
||||||
class API:
|
class API:
|
||||||
def __init__(self):
|
def __init__(self):
|
||||||
self.crypt_size = 1
|
self.crypt_size = 4096
|
||||||
self.app = Flask(__name__)
|
self.app = Flask(__name__)
|
||||||
self.ai_response = {}
|
self.ai_response = {}
|
||||||
self.ai = AI()
|
self.ai = AI()
|
||||||
self.db = DB()
|
self.db = DB()
|
||||||
self.db.load_database()
|
self.db.load_database()
|
||||||
self.ai_response_lock = threading.Lock()
|
|
||||||
CORS(self.app)
|
CORS(self.app)
|
||||||
|
|
||||||
def run(self):
|
def run(self):
|
||||||
|
@ -36,18 +34,10 @@ class API:
|
||||||
return jsonify({'status': 401, 'error': 'Invalid access token'})
|
return jsonify({'status': 401, 'error': 'Invalid access token'})
|
||||||
|
|
||||||
if model_type == "local":
|
if model_type == "local":
|
||||||
thread = threading.Thread(target=self.ai.process_local, args=(ai_model, messages, self, access_token))
|
self.ai.process_local(ai_model, messages, self, access_token)
|
||||||
thread.start()
|
if model_type == "mistral":
|
||||||
thread.join()
|
self.ai.process_mistralai(ai_model, messages, self, access_token)
|
||||||
return jsonify({'status': 200})
|
return jsonify({'status': 200})
|
||||||
elif model_type == "mistral":
|
|
||||||
api_key = data.get('api_key')
|
|
||||||
thread = threading.Thread(target=self.ai.process_mistralai, args=(ai_model, messages, self, access_token, api_key))
|
|
||||||
thread.start()
|
|
||||||
thread.join()
|
|
||||||
return jsonify({'status': 200})
|
|
||||||
|
|
||||||
return jsonify({'status': 401, 'error': 'Invalid AI model type'})
|
|
||||||
|
|
||||||
@self.app.route('/interstellar/api/ai_get', methods=['GET'])
|
@self.app.route('/interstellar/api/ai_get', methods=['GET'])
|
||||||
def get_ai():
|
def get_ai():
|
||||||
|
@ -105,7 +95,7 @@ class API:
|
||||||
f.write(crypto.dump_privatekey(crypto.FILETYPE_PEM, k).decode("utf-8"))
|
f.write(crypto.dump_privatekey(crypto.FILETYPE_PEM, k).decode("utf-8"))
|
||||||
|
|
||||||
ssl_context = ("cert.pem", "key.pem")
|
ssl_context = ("cert.pem", "key.pem")
|
||||||
self.app.run(debug=True, host='0.0.0.0', port=5000)
|
self.app.run(debug=True, host='0.0.0.0', port=5000, ssl_context=ssl_context)
|
||||||
|
|
||||||
|
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
|
|
|
@ -2,5 +2,4 @@ flask
|
||||||
flask-cors
|
flask-cors
|
||||||
ollama
|
ollama
|
||||||
mistralai
|
mistralai
|
||||||
openai
|
|
||||||
pyOpenSSL
|
pyOpenSSL
|
Loading…
Reference in a new issue