diff --git a/g4f/gui/client/html/index.html b/g4f/gui/client/index.html similarity index 91% rename from g4f/gui/client/html/index.html rename to g4f/gui/client/index.html index d70f87a1..9ce6b66a 100644 --- a/g4f/gui/client/html/index.html +++ b/g4f/gui/client/index.html @@ -1,5 +1,5 @@ - + @@ -10,14 +10,14 @@ - - - - - - - - + + + + + + + + @@ -38,8 +38,8 @@ - - + + g4f - gui diff --git a/g4f/gui/client/css/style.css b/g4f/gui/client/static/css/style.css similarity index 100% rename from g4f/gui/client/css/style.css rename to g4f/gui/client/static/css/style.css diff --git a/g4f/gui/client/img/android-chrome-192x192.png b/g4f/gui/client/static/img/android-chrome-192x192.png similarity index 100% rename from g4f/gui/client/img/android-chrome-192x192.png rename to g4f/gui/client/static/img/android-chrome-192x192.png diff --git a/g4f/gui/client/img/android-chrome-512x512.png b/g4f/gui/client/static/img/android-chrome-512x512.png similarity index 100% rename from g4f/gui/client/img/android-chrome-512x512.png rename to g4f/gui/client/static/img/android-chrome-512x512.png diff --git a/g4f/gui/client/img/apple-touch-icon.png b/g4f/gui/client/static/img/apple-touch-icon.png similarity index 100% rename from g4f/gui/client/img/apple-touch-icon.png rename to g4f/gui/client/static/img/apple-touch-icon.png diff --git a/g4f/gui/client/img/favicon-16x16.png b/g4f/gui/client/static/img/favicon-16x16.png similarity index 100% rename from g4f/gui/client/img/favicon-16x16.png rename to g4f/gui/client/static/img/favicon-16x16.png diff --git a/g4f/gui/client/img/favicon-32x32.png b/g4f/gui/client/static/img/favicon-32x32.png similarity index 100% rename from g4f/gui/client/img/favicon-32x32.png rename to g4f/gui/client/static/img/favicon-32x32.png diff --git a/g4f/gui/client/img/gpt.png b/g4f/gui/client/static/img/gpt.png similarity index 100% rename from g4f/gui/client/img/gpt.png rename to g4f/gui/client/static/img/gpt.png diff --git a/g4f/gui/client/img/site.webmanifest b/g4f/gui/client/static/img/site.webmanifest similarity index 100% rename from g4f/gui/client/img/site.webmanifest rename to g4f/gui/client/static/img/site.webmanifest diff --git a/g4f/gui/client/img/user.png b/g4f/gui/client/static/img/user.png similarity index 100% rename from g4f/gui/client/img/user.png rename to g4f/gui/client/static/img/user.png diff --git a/g4f/gui/client/js/chat.v1.js b/g4f/gui/client/static/js/chat.v1.js similarity index 80% rename from g4f/gui/client/js/chat.v1.js rename to g4f/gui/client/static/js/chat.v1.js index c0d0cb6d..bfe641d5 100644 --- a/g4f/gui/client/js/chat.v1.js +++ b/g4f/gui/client/static/js/chat.v1.js @@ -1,5 +1,4 @@ const colorThemes = document.querySelectorAll('[name="theme"]'); -const markdown = window.markdownit(); const message_box = document.getElementById(`messages`); const messageInput = document.getElementById(`message-input`); const box_conversations = document.querySelector(`.top`); @@ -22,8 +21,6 @@ let prompt_lock = false; const options = ["switch", "model", "model2", "jailbreak", "patch", "provider", "history"]; -hljs.addPlugin(new CopyButtonPlugin()); - messageInput.addEventListener("blur", () => { window.scrollTo(0, 0); }); @@ -39,31 +36,29 @@ appStorage = window.localStorage || { length: 0 } +const markdown = window.markdownit(); const markdown_render = (content) => { return markdown.render(content - .replaceAll(//gm, "") + .replaceAll(/[\s\S]+/gm, "") .replaceAll(//gm, "") ) .replaceAll("', '') } +hljs.addPlugin(new CopyButtonPlugin()); let typesetPromise = Promise.resolve(); -let timeoutHighlightId; const highlight = (container) => { - if (timeoutHighlightId) clearTimeout(timeoutHighlightId); - timeoutHighlightId = setTimeout(() => { - container.querySelectorAll('code:not(.hljs').forEach((el) => { - if (el.className != "hljs") { - hljs.highlightElement(el); - } - }); - typesetPromise = typesetPromise.then( - () => MathJax.typesetPromise([container]) - ).catch( - (err) => console.log('Typeset failed: ' + err.message) - ); - }, 100); + container.querySelectorAll('code:not(.hljs').forEach((el) => { + if (el.className != "hljs") { + hljs.highlightElement(el); + } + }); + typesetPromise = typesetPromise.then( + () => MathJax.typesetPromise([container]) + ).catch( + (err) => console.log('Typeset failed: ' + err.message) + ); } const register_remove_message = async () => { @@ -99,47 +94,48 @@ const handle_ask = async () => { window.scrollTo(0, 0); message = messageInput.value - if (message.length > 0) { - messageInput.value = ""; - prompt_lock = true; - count_input() - await add_conversation(window.conversation_id, message); - if ("text" in fileInput.dataset) { - message += '\n```' + fileInput.dataset.type + '\n'; - message += fileInput.dataset.text; - message += '\n```' - } - let message_index = await add_message(window.conversation_id, "user", message); - window.token = message_id(); - - if (imageInput.dataset.src) URL.revokeObjectURL(imageInput.dataset.src); - const input = imageInput && imageInput.files.length > 0 ? imageInput : cameraInput - if (input.files.length > 0) imageInput.dataset.src = URL.createObjectURL(input.files[0]); - else delete imageInput.dataset.src - - message_box.innerHTML += ` -
-
- ${user_image} - - -
-
-
- ${markdown_render(message)} - ${imageInput.dataset.src - ? 'Image upload' - : '' - } -
-
${count_words_and_tokens(message, get_selected_model())}
-
-
- `; - await register_remove_message(); - highlight(message_box); - await ask_gpt(); + if (message.length <= 0) { + return; } + messageInput.value = ""; + prompt_lock = true; + count_input() + await add_conversation(window.conversation_id, message); + + if ("text" in fileInput.dataset) { + message += '\n```' + fileInput.dataset.type + '\n'; + message += fileInput.dataset.text; + message += '\n```' + } + let message_index = await add_message(window.conversation_id, "user", message); + window.token = message_id(); + + if (imageInput.dataset.src) URL.revokeObjectURL(imageInput.dataset.src); + const input = imageInput && imageInput.files.length > 0 ? imageInput : cameraInput + if (input.files.length > 0) imageInput.dataset.src = URL.createObjectURL(input.files[0]); + else delete imageInput.dataset.src + + message_box.innerHTML += ` +
+
+ ${user_image} + + +
+
+
+ ${markdown_render(message)} + ${imageInput.dataset.src + ? 'Image upload' + : '' + } +
+
${count_words_and_tokens(message, get_selected_model())}
+
+
+ `; + highlight(message_box); + await ask_gpt(); }; const remove_cancel_button = async () => { @@ -151,7 +147,7 @@ const remove_cancel_button = async () => { }, 300); }; -const prepare_messages = (messages, filter_last_message = true) => { +const prepare_messages = (messages, filter_last_message=true) => { // Removes none user messages at end if (filter_last_message) { let last_message; @@ -201,17 +197,52 @@ const prepare_messages = (messages, filter_last_message = true) => { return new_messages; } +async function add_message_chunk(message) { + if (message.type == "provider") { + window.provider_result = message.provider; + window.content.querySelector('.provider').innerHTML = ` +
+ ${message.provider.name} + + ${message.provider.model ? ' with ' + message.provider.model : ''} + ` + } else if (message.type == "message") { + console.error(messag.message) + return; + } else if (message.type == "error") { + console.error(message.error); + window.content_inner.innerHTML += `

An error occured: ${message.error}

`; + } else if (message.type == "content") { + window.text += message.content; + html = markdown_render(window.text); + let lastElement, lastIndex = null; + for (element of ['

', '
', '

\n\n', '\n', '\n']) { + const index = html.lastIndexOf(element) + if (index - element.length > lastIndex) { + lastElement = element; + lastIndex = index; + } + } + if (lastIndex) { + html = html.substring(0, lastIndex) + '' + lastElement; + } + window.content_inner.innerHTML = html; + window.content_count.innerText = count_words_and_tokens(text, window.provider_result?.model); + highlight(window.content_inner); + } + + window.scrollTo(0, 0); + if (message_box.scrollTop >= message_box.scrollHeight - message_box.clientHeight - 100) { + message_box.scrollTo({ top: message_box.scrollHeight, behavior: "auto" }); + } +} + const ask_gpt = async () => { regenerate.classList.add(`regenerate-hidden`); messages = await get_messages(window.conversation_id); total_messages = messages.length; - messages = prepare_messages(messages); - window.scrollTo(0, 0); - window.controller = new AbortController(); - window.text = ""; - stop_generating.classList.remove(`stop_generating-hidden`); message_box.scrollTop = message_box.scrollHeight; @@ -234,103 +265,31 @@ const ask_gpt = async () => { `; - content = document.getElementById(`gpt_${window.token}`); - content_inner = content.querySelector('.content_inner'); - content_count = content.querySelector('.count'); + + window.controller = new AbortController(); + window.text = ""; + window.error = null; + window.provider_result = null; + + window.content = document.getElementById(`gpt_${window.token}`); + window.content_inner = content.querySelector('.content_inner'); + window.content_count = content.querySelector('.count'); message_box.scrollTop = message_box.scrollHeight; window.scrollTo(0, 0); - - error = provider_result = null; try { - let body = JSON.stringify({ + const input = imageInput && imageInput.files.length > 0 ? imageInput : cameraInput; + const file = input && input.files.length > 0 ? input.files[0] : null; + await api("conversation", { id: window.token, conversation_id: window.conversation_id, model: get_selected_model(), jailbreak: jailbreak?.options[jailbreak.selectedIndex].value, - web_search: document.getElementById(`switch`).checked, + web_search: document.getElementById("switch").checked, provider: providerSelect.options[providerSelect.selectedIndex].value, - patch_provider: document.getElementById('patch')?.checked, + patch_provider: document.getElementById("patch")?.checked, messages: messages - }); - const headers = { - accept: 'text/event-stream' - } - const input = imageInput && imageInput.files.length > 0 ? imageInput : cameraInput - if (input && input.files.length > 0) { - const formData = new FormData(); - formData.append('image', input.files[0]); - formData.append('json', body); - body = formData; - } else { - headers['content-type'] = 'application/json'; - } - - const response = await fetch(`/backend-api/v2/conversation`, { - method: 'POST', - signal: window.controller.signal, - headers: headers, - body: body - }); - const reader = response.body.pipeThrough(new TextDecoderStream()).getReader(); - let buffer = "" - while (true) { - const { value, done } = await reader.read(); - if (done) break; - for (const line of value.split("\n")) { - if (!line) { - continue; - } - let message; - try { - message = JSON.parse(buffer + line); - buffer = ""; - } catch { - buffer += line - continue; - } - if (message.type == "content") { - text += message.content; - } else if (message.type == "provider") { - provider_result = message.provider - content.querySelector('.provider').innerHTML = ` - - ${provider_result.name} - - ${provider_result.model ? ' with ' + provider_result.model : ''} - ` - } else if (message.type == "error") { - error = message.error; - } else if (messag.type == "message") { - console.error(messag.message) - } - } - if (error) { - console.error(error); - content_inner.innerHTML += `

An error occured: ${error}

`; - } else { - html = markdown_render(text); - let lastElement, lastIndex = null; - for (element of ['

', '', '

\n\n', '\n', '\n']) { - const index = html.lastIndexOf(element) - if (index - element.length > lastIndex) { - lastElement = element; - lastIndex = index; - } - } - if (lastIndex) { - html = html.substring(0, lastIndex) + '' + lastElement; - } - content_inner.innerHTML = html; - content_count.innerText = count_words_and_tokens(text, provider_result?.model); - highlight(content_inner); - } - - window.scrollTo(0, 0); - if (message_box.scrollTop >= message_box.scrollHeight - message_box.clientHeight - 100) { - message_box.scrollTo({ top: message_box.scrollHeight, behavior: "auto" }); - } - } + }, file); if (!error) { html = markdown_render(text); content_inner.innerHTML = html; @@ -355,7 +314,7 @@ const ask_gpt = async () => { await add_message(window.conversation_id, "assistant", text, provider_result); await load_conversation(window.conversation_id); } else { - let cursorDiv = document.getElementById(`cursor`); + let cursorDiv = document.getElementById("cursor"); if (cursorDiv) cursorDiv.parentNode.removeChild(cursorDiv); } window.scrollTo(0, 0); @@ -444,7 +403,7 @@ const new_conversation = async () => { say_hello(); }; -const load_conversation = async (conversation_id, scroll = true) => { +const load_conversation = async (conversation_id, scroll=true) => { let conversation = await get_conversation(conversation_id); let messages = conversation?.items || []; @@ -459,7 +418,6 @@ const load_conversation = async (conversation_id, scroll = true) => { last_model = item.provider?.model; let next_i = parseInt(i) + 1; let next_provider = item.provider ? item.provider : (messages.length > next_i ? messages[next_i].provider : null); - let provider_link = item.provider?.name ? `${item.provider.name}` : ""; let provider = provider_link ? `
@@ -496,7 +454,6 @@ const load_conversation = async (conversation_id, scroll = true) => { } message_box.innerHTML = elements; - register_remove_message(); highlight(message_box); @@ -548,7 +505,9 @@ async function add_conversation(conversation_id, content) { } async function save_system_message() { - if (!window.conversation_id) return; + if (!window.conversation_id) { + return; + } const conversation = await get_conversation(window.conversation_id); conversation.system = systemPrompt?.value; await save_conversation(window.conversation_id, conversation); @@ -585,7 +544,6 @@ const remove_message = async (conversation_id, index) => { const add_message = async (conversation_id, role, content, provider) => { const conversation = await get_conversation(conversation_id); - conversation.items.push({ role: role, content: content, @@ -667,11 +625,10 @@ sidebar_button.addEventListener("click", (event) => { sidebar.classList.add("shown"); sidebar_button.classList.add("rotated"); } - window.scrollTo(0, 0); }); -const register_settings_localstorage = async () => { +const register_settings_storage = async () => { options.forEach((id) => { element = document.getElementById(id); if (!element) { @@ -692,10 +649,10 @@ const register_settings_localstorage = async () => { }); } -const load_settings_localstorage = async () => { +const load_settings_storage = async () => { options.forEach((id) => { element = document.getElementById(id); - if (!element || !(value = appStorage.getItem(element.id))) { + if (!element || !(value = appStorage.getItem(id))) { return; } if (value) { @@ -807,9 +764,21 @@ systemPrompt.addEventListener("blur", function() { count_input(); }); -window.onload = async () => { - setTheme(); +window.addEventListener('load', async function() { + await on_load(); + if (window.conversation_id == "{{chat_id}}") { + window.conversation_id = uuid(); + } else { + await on_api(); + } +}); +window.addEventListener('pywebviewready', async function() { + await on_api(); +}); + +async function on_load() { + setTheme(); count_input(); if (/\/chat\/.+/.test(window.location.href)) { @@ -817,9 +786,10 @@ window.onload = async () => { } else { say_hello() } - load_conversations(); +} +async function on_api() { messageInput.addEventListener("keydown", async (evt) => { if (prompt_lock) return; @@ -832,48 +802,17 @@ window.onload = async () => { messageInput.style.height = messageInput.scrollHeight + "px"; } }); - sendButton.addEventListener(`click`, async () => { console.log("clicked send"); if (prompt_lock) return; await handle_ask(); }); - messageInput.focus(); - register_settings_localstorage(); -}; + register_settings_storage(); -(async () => { - response = await fetch('/backend-api/v2/models') - models = await response.json() - - for (model of models) { - let option = document.createElement('option'); - option.value = option.text = model; - modelSelect.appendChild(option); - } - - response = await fetch('/backend-api/v2/providers') - providers = await response.json() - select = document.getElementById('provider'); - - providers.forEach((provider) => { - let option = document.createElement('option'); - option.value = option.text = provider; - select.appendChild(option); - }) - - await load_provider_models(); - - await load_settings_localstorage() -})(); - -(async () => { - response = await fetch('/backend-api/v2/version') - versions = await response.json() - - document.title = 'g4f - gui - ' + versions["version"]; + versions = await api('version') + document.title = 'g4f - ' + versions["version"]; let text = "version ~ " if (versions["version"] != versions["latest_version"]) { let release_url = 'https://github.com/xtekky/gpt4free/releases/tag/' + versions["latest_version"]; @@ -883,7 +822,24 @@ window.onload = async () => { text += versions["version"]; } document.getElementById("version_text").innerHTML = text -})() + + models = await api("models"); + models.forEach((model) => { + let option = document.createElement('option'); + option.value = option.text = model; + modelSelect.appendChild(option); + }); + + providers = await api('providers') + providers.forEach((provider) => { + let option = document.createElement('option'); + option.value = option.text = provider; + providerSelect.appendChild(option); + }) + + load_provider_models(); + load_settings_storage() +} for (const el of [imageInput, cameraInput]) { el.addEventListener('click', async () => { @@ -899,6 +855,7 @@ fileInput.addEventListener('click', async (event) => { fileInput.value = ''; delete fileInput.dataset.text; }); + fileInput.addEventListener('change', async (event) => { if (fileInput.files.length) { type = fileInput.files[0].type; @@ -947,10 +904,73 @@ function get_selected_model() { } } +async function api(ressource, args=null, file=null) { + if (window?.pywebview) { + if (args) { + if (ressource == "models") { + ressource = "provider_models"; + } + return pywebview.api["get_" + ressource](args); + } + return pywebview.api["get_" + ressource](); + } + if (ressource == "models" && args) { + ressource = `${ressource}/${args}`; + } + const url = `/backend-api/v2/${ressource}`; + if (ressource == "conversation") { + const body = JSON.stringify(args); + const headers = { + accept: 'text/event-stream' + } + if (file) { + const formData = new FormData(); + formData.append('file', file); + formData.append('json', body); + body = formData; + } else { + headers['content-type'] = 'application/json'; + } + response = await fetch(url, { + method: 'POST', + signal: window.controller.signal, + headers: headers, + body: body + }); + return read_response(response); + } + response = await fetch(url); + return await response.json(); +} + +async function read_response(response) { + const reader = response.body.pipeThrough(new TextDecoderStream()).getReader(); + let buffer = "" + while (true) { + const { value, done } = await reader.read(); + if (done) { + break; + } + for (const line of value.split("\n")) { + if (!line) { + continue; + } + try { + add_message_chunk(JSON.parse(buffer + line)) + buffer = ""; + } catch { + buffer += line + } + } + } +} + async function load_provider_models() { - provider = providerSelect.options[providerSelect.selectedIndex].value; - response = await fetch('/backend-api/v2/models/' + provider); - models = await response.json(); + const provider = providerSelect.options[providerSelect.selectedIndex].value; + if (!provider) { + return; + } + const models = await api('models', provider); modelProvider.innerHTML = ''; if (models.length > 0) { modelSelect.classList.add("hidden"); diff --git a/g4f/gui/client/js/highlight.min.js b/g4f/gui/client/static/js/highlight.min.js similarity index 100% rename from g4f/gui/client/js/highlight.min.js rename to g4f/gui/client/static/js/highlight.min.js diff --git a/g4f/gui/client/js/highlightjs-copy.min.js b/g4f/gui/client/static/js/highlightjs-copy.min.js similarity index 100% rename from g4f/gui/client/js/highlightjs-copy.min.js rename to g4f/gui/client/static/js/highlightjs-copy.min.js diff --git a/g4f/gui/client/js/icons.js b/g4f/gui/client/static/js/icons.js similarity index 100% rename from g4f/gui/client/js/icons.js rename to g4f/gui/client/static/js/icons.js diff --git a/g4f/gui/server/api.py b/g4f/gui/server/api.py new file mode 100644 index 00000000..df7b487d --- /dev/null +++ b/g4f/gui/server/api.py @@ -0,0 +1,172 @@ +import logging +import json +from typing import Iterator + +try: + import webview +except ImportError: + ... + +from g4f import version, models +from g4f import get_last_provider, ChatCompletion +from g4f.errors import VersionNotFoundError +from g4f.Provider import ProviderType, __providers__, __map__ +from g4f.providers.base_provider import ProviderModelMixin +from g4f.Provider.bing.create_images import patch_provider + +class Api(): + def get_models(self) -> list[str]: + """ + Return a list of all models. + + Fetches and returns a list of all available models in the system. + + Returns: + List[str]: A list of model names. + """ + return models._all_models + + def get_provider_models(self, provider: str) -> list[dict]: + if provider in __map__: + provider: ProviderType = __map__[provider] + if issubclass(provider, ProviderModelMixin): + return [{"model": model, "default": model == provider.default_model} for model in provider.get_models()] + elif provider.supports_gpt_35_turbo or provider.supports_gpt_4: + return [ + *([{"model": "gpt-4", "default": not provider.supports_gpt_4}] if provider.supports_gpt_4 else []), + *([{"model": "gpt-3.5-turbo", "default": not provider.supports_gpt_4}] if provider.supports_gpt_35_turbo else []) + ] + else: + return []; + + def get_providers(self) -> list[str]: + """ + Return a list of all working providers. + """ + return [provider.__name__ for provider in __providers__ if provider.working] + + def get_version(self): + """ + Returns the current and latest version of the application. + + Returns: + dict: A dictionary containing the current and latest version. + """ + try: + current_version = version.utils.current_version + except VersionNotFoundError: + current_version = None + return { + "version": current_version, + "latest_version": version.utils.latest_version, + } + + def generate_title(self): + """ + Generates and returns a title based on the request data. + + Returns: + dict: A dictionary with the generated title. + """ + return {'title': ''} + + def get_conversation(self, options: dict, **kwargs) -> Iterator: + window = webview.active_window() + for message in self._create_response_stream( + self._prepare_conversation_kwargs(options, kwargs) + ): + window.evaluate_js(f"this.add_message_chunk({json.dumps(message)})") + + def _prepare_conversation_kwargs(self, json_data: dict, kwargs: dict): + """ + Prepares arguments for chat completion based on the request data. + + Reads the request and prepares the necessary arguments for handling + a chat completion request. + + Returns: + dict: Arguments prepared for chat completion. + """ + provider = json_data.get('provider', None) + if "image" in kwargs and provider is None: + provider = "Bing" + if provider == 'OpenaiChat': + kwargs['auto_continue'] = True + + messages = json_data['messages'] + if json_data.get('web_search'): + if provider == "Bing": + kwargs['web_search'] = True + else: + from .internet import get_search_message + messages[-1]["content"] = get_search_message(messages[-1]["content"]) + + model = json_data.get('model') + model = model if model else models.default + patch = patch_provider if json_data.get('patch_provider') else None + + return { + "model": model, + "provider": provider, + "messages": messages, + "stream": True, + "ignore_stream": True, + "patch_provider": patch, + **kwargs + } + + def _create_response_stream(self, kwargs) -> Iterator: + """ + Creates and returns a streaming response for the conversation. + + Args: + kwargs (dict): Arguments for creating the chat completion. + + Yields: + str: JSON formatted response chunks for the stream. + + Raises: + Exception: If an error occurs during the streaming process. + """ + try: + first = True + for chunk in ChatCompletion.create(**kwargs): + if first: + first = False + yield self._format_json('provider', get_last_provider(True)) + if isinstance(chunk, Exception): + logging.exception(chunk) + yield self._format_json('message', get_error_message(chunk)) + else: + yield self._format_json('content', chunk) + except Exception as e: + logging.exception(e) + yield self._format_json('error', get_error_message(e)) + + def _format_json(self, response_type: str, content): + """ + Formats and returns a JSON response. + + Args: + response_type (str): The type of the response. + content: The content to be included in the response. + + Returns: + str: A JSON formatted string. + """ + return { + 'type': response_type, + response_type: content + } + +def get_error_message(exception: Exception) -> str: + """ + Generates a formatted error message from an exception. + + Args: + exception (Exception): The exception to format. + + Returns: + str: A formatted error message string. + """ + return f"{get_last_provider().__name__}: {type(exception).__name__}: {exception}" \ No newline at end of file diff --git a/g4f/gui/server/app.py b/g4f/gui/server/app.py index d9e25ca1..869d3880 100644 --- a/g4f/gui/server/app.py +++ b/g4f/gui/server/app.py @@ -2,8 +2,8 @@ import sys, os from flask import Flask if getattr(sys, 'frozen', False): - template_folder = os.path.join(sys._MEIPASS, "client/html") + template_folder = os.path.join(sys._MEIPASS, "client") else: - template_folder = "./../client/html" + template_folder = "../client" -app = Flask(__name__, template_folder=template_folder) \ No newline at end of file +app = Flask(__name__, template_folder=template_folder, static_folder=f"{template_folder}/static") \ No newline at end of file diff --git a/g4f/gui/server/backend.py b/g4f/gui/server/backend.py index e9617c07..8d388708 100644 --- a/g4f/gui/server/backend.py +++ b/g4f/gui/server/backend.py @@ -1,16 +1,9 @@ -import logging import json from flask import request, Flask -from typing import Generator -from g4f import version, models -from g4f import get_last_provider, ChatCompletion from g4f.image import is_allowed_extension, to_image -from g4f.errors import VersionNotFoundError -from g4f.Provider import ProviderType, __providers__, __map__ -from g4f.providers.base_provider import ProviderModelMixin -from g4f.Provider.bing.create_images import patch_provider +from .api import Api -class Backend_Api: +class Backend_Api(Api): """ Handles various endpoints in a Flask application for backend operations. @@ -59,7 +52,7 @@ class Backend_Api: 'methods': ['POST'] } } - + def handle_error(self): """ Initialize the backend API with the given Flask application. @@ -69,63 +62,7 @@ class Backend_Api: """ print(request.json) return 'ok', 200 - - def get_models(self): - """ - Return a list of all models. - Fetches and returns a list of all available models in the system. - - Returns: - List[str]: A list of model names. - """ - return models._all_models - - def get_provider_models(self, provider: str): - if provider in __map__: - provider: ProviderType = __map__[provider] - if issubclass(provider, ProviderModelMixin): - return [{"model": model, "default": model == provider.default_model} for model in provider.get_models()] - elif provider.supports_gpt_35_turbo or provider.supports_gpt_4: - return [ - *([{"model": "gpt-3.5-turbo", "default": not provider.supports_gpt_4}] if provider.supports_gpt_35_turbo else []), - *([{"model": "gpt-4", "default": not provider.supports_gpt_4}] if provider.supports_gpt_4 else []) - ] - else: - return []; - return 404, "Provider not found" - - def get_providers(self): - """ - Return a list of all working providers. - """ - return [provider.__name__ for provider in __providers__ if provider.working] - - def get_version(self): - """ - Returns the current and latest version of the application. - - Returns: - dict: A dictionary containing the current and latest version. - """ - try: - current_version = version.utils.current_version - except VersionNotFoundError: - current_version = None - return { - "version": current_version, - "latest_version": version.utils.latest_version, - } - - def generate_title(self): - """ - Generates and returns a title based on the request data. - - Returns: - dict: A dictionary with the generated title. - """ - return {'title': ''} - def handle_conversation(self): """ Handles conversation requests and streams responses back. @@ -133,26 +70,10 @@ class Backend_Api: Returns: Response: A Flask response object for streaming. """ - kwargs = self._prepare_conversation_kwargs() - - return self.app.response_class( - self._create_response_stream(kwargs), - mimetype='text/event-stream' - ) - - def _prepare_conversation_kwargs(self): - """ - Prepares arguments for chat completion based on the request data. - - Reads the request and prepares the necessary arguments for handling - a chat completion request. - - Returns: - dict: Arguments prepared for chat completion. - """ + kwargs = {} - if "image" in request.files: - file = request.files['image'] + if "file" in request.files: + file = request.files['file'] if file.filename != '' and is_allowed_extension(file.filename): kwargs['image'] = to_image(file.stream, file.filename.endswith('.svg')) kwargs['image_name'] = file.filename @@ -160,66 +81,20 @@ class Backend_Api: json_data = json.loads(request.form['json']) else: json_data = request.json - - provider = json_data.get('provider', '').replace('g4f.Provider.', '') - provider = provider if provider and provider != "Auto" else None - if "image" in kwargs and not provider: - provider = "Bing" - if provider == 'OpenaiChat': - kwargs['auto_continue'] = True + kwargs = self._prepare_conversation_kwargs(json_data, kwargs) - messages = json_data['messages'] - if json_data.get('web_search'): - if provider == "Bing": - kwargs['web_search'] = True - else: - # ResourceWarning: unclosed event loop - from .internet import get_search_message - messages[-1]["content"] = get_search_message(messages[-1]["content"]) + return self.app.response_class( + self._create_response_stream(kwargs), + mimetype='text/event-stream' + ) - model = json_data.get('model') - model = model if model else models.default - patch = patch_provider if json_data.get('patch_provider') else None + def get_provider_models(self, provider: str): + models = super().get_provider_models(provider) + if models is None: + return 404, "Provider not found" + return models - return { - "model": model, - "provider": provider, - "messages": messages, - "stream": True, - "ignore_stream": True, - "patch_provider": patch, - **kwargs - } - - def _create_response_stream(self, kwargs) -> Generator[str, None, None]: - """ - Creates and returns a streaming response for the conversation. - - Args: - kwargs (dict): Arguments for creating the chat completion. - - Yields: - str: JSON formatted response chunks for the stream. - - Raises: - Exception: If an error occurs during the streaming process. - """ - try: - first = True - for chunk in ChatCompletion.create(**kwargs): - if first: - first = False - yield self._format_json('provider', get_last_provider(True)) - if isinstance(chunk, Exception): - logging.exception(chunk) - yield self._format_json('message', get_error_message(chunk)) - else: - yield self._format_json('content', str(chunk)) - except Exception as e: - logging.exception(e) - yield self._format_json('error', get_error_message(e)) - def _format_json(self, response_type: str, content) -> str: """ Formats and returns a JSON response. @@ -231,19 +106,4 @@ class Backend_Api: Returns: str: A JSON formatted string. """ - return json.dumps({ - 'type': response_type, - response_type: content - }) + "\n" - -def get_error_message(exception: Exception) -> str: - """ - Generates a formatted error message from an exception. - - Args: - exception (Exception): The exception to format. - - Returns: - str: A formatted error message string. - """ - return f"{get_last_provider().__name__}: {type(exception).__name__}: {exception}" \ No newline at end of file + return json.dumps(super()._format_json(response_type, content)) + "\n" \ No newline at end of file diff --git a/g4f/gui/server/website.py b/g4f/gui/server/website.py index 9aa98b76..4e611177 100644 --- a/g4f/gui/server/website.py +++ b/g4f/gui/server/website.py @@ -1,12 +1,5 @@ -from flask import render_template, send_file, redirect -from time import time -from os import urandom -import sys, os - -if getattr(sys, 'frozen', False): - assets_folder = os.path.join(sys._MEIPASS, "client") -else: - assets_folder = "./../client" +import uuid +from flask import render_template, redirect class Website: def __init__(self, app) -> None: @@ -24,23 +17,12 @@ class Website: 'function': self._chat, 'methods': ['GET', 'POST'] }, - '/assets//': { - 'function': self._assets, - 'methods': ['GET', 'POST'] - } } def _chat(self, conversation_id): if '-' not in conversation_id: return redirect('/chat') - - return render_template('index.html', chat_id = conversation_id) + return render_template('index.html', chat_id=conversation_id) def _index(self): - return render_template('index.html', chat_id = f'{urandom(4).hex()}-{urandom(2).hex()}-{urandom(2).hex()}-{urandom(2).hex()}-{hex(int(time() * 1000))[2:]}') - - def _assets(self, folder: str, file: str): - try: - return send_file(f"{assets_folder}/{folder}/{file}", as_attachment=False) - except: - return "File not found", 404 \ No newline at end of file + return render_template('index.html', chat_id=str(uuid.uuid4())) \ No newline at end of file diff --git a/g4f/gui/webview.py b/g4f/gui/webview.py index 16f143c4..a8e745f3 100644 --- a/g4f/gui/webview.py +++ b/g4f/gui/webview.py @@ -1,34 +1,32 @@ import webview -from functools import partial try: from platformdirs import user_config_dir has_platformdirs = True except ImportError: has_platformdirs = False -from g4f.gui import run_gui from g4f.gui.run import gui_parser +from g4f.gui.server.api import Api import g4f.version import g4f.debug def run_webview( - host: str = "0.0.0.0", - port: int = 8080, debug: bool = False, storage_path: str = None ): webview.create_window( f"g4f - {g4f.version.utils.current_version}", - f"http://{host}:{port}/", - text_select=True + "client/index.html", + text_select=True, + js_api=Api(), ) if has_platformdirs and storage_path is None: storage_path = user_config_dir("g4f-webview") webview.start( - partial(run_gui, host, port), private_mode=False, storage_path=storage_path, - debug=debug + debug=debug, + ssl=True ) if __name__ == "__main__": @@ -36,4 +34,4 @@ if __name__ == "__main__": args = parser.parse_args() if args.debug: g4f.debug.logging = True - run_webview(args.host, args.port, args.debug) \ No newline at end of file + run_webview(args.debug) \ No newline at end of file