summaryrefslogtreecommitdiffstats
path: root/g4f
diff options
context:
space:
mode:
Diffstat (limited to 'g4f')
-rw-r--r--g4f/Provider/AI365VIP.py2
-rw-r--r--g4f/Provider/AiMathGPT.py4
-rw-r--r--g4f/Provider/Blackbox.py2
-rw-r--r--g4f/Provider/GizAI.py151
-rw-r--r--g4f/Provider/__init__.py1
-rw-r--r--g4f/Provider/nexra/NexraChatGPT.py270
-rw-r--r--g4f/Provider/nexra/NexraChatGPT4o.py86
-rw-r--r--g4f/Provider/nexra/NexraChatGptV2.py92
-rw-r--r--g4f/Provider/nexra/NexraChatGptWeb.py64
-rw-r--r--g4f/Provider/nexra/__init__.py3
-rw-r--r--g4f/api/__init__.py147
-rw-r--r--g4f/client/client.py30
-rw-r--r--g4f/gui/client/index.html488
-rw-r--r--g4f/gui/client/static/css/style.css87
-rw-r--r--g4f/gui/client/static/js/chat.v1.js126
-rw-r--r--g4f/gui/server/api.py63
-rw-r--r--g4f/gui/server/backend.py24
-rw-r--r--g4f/gui/server/website.py6
-rw-r--r--g4f/models.py47
19 files changed, 1017 insertions, 676 deletions
diff --git a/g4f/Provider/AI365VIP.py b/g4f/Provider/AI365VIP.py
index c7ebf6b5..511ad568 100644
--- a/g4f/Provider/AI365VIP.py
+++ b/g4f/Provider/AI365VIP.py
@@ -10,7 +10,7 @@ from .helper import format_prompt
class AI365VIP(AsyncGeneratorProvider, ProviderModelMixin):
url = "https://chat.ai365vip.com"
api_endpoint = "/api/chat"
- working = True
+ working = False
default_model = 'gpt-3.5-turbo'
models = [
'gpt-3.5-turbo',
diff --git a/g4f/Provider/AiMathGPT.py b/g4f/Provider/AiMathGPT.py
index 4399320a..90931691 100644
--- a/g4f/Provider/AiMathGPT.py
+++ b/g4f/Provider/AiMathGPT.py
@@ -60,10 +60,6 @@ class AiMathGPT(AsyncGeneratorProvider, ProviderModelMixin):
data = {
"messages": [
{
- "role": "system",
- "content": ""
- },
- {
"role": "user",
"content": format_prompt(messages)
}
diff --git a/g4f/Provider/Blackbox.py b/g4f/Provider/Blackbox.py
index 5cd43eed..6d8a467d 100644
--- a/g4f/Provider/Blackbox.py
+++ b/g4f/Provider/Blackbox.py
@@ -51,7 +51,6 @@ class Blackbox(AsyncGeneratorProvider, ProviderModelMixin):
'ReactAgent',
'XcodeAgent',
'AngularJSAgent',
- 'RepoMap',
]
agentMode = {
@@ -78,7 +77,6 @@ class Blackbox(AsyncGeneratorProvider, ProviderModelMixin):
'ReactAgent': {'mode': True, 'id': "React Agent"},
'XcodeAgent': {'mode': True, 'id': "Xcode Agent"},
'AngularJSAgent': {'mode': True, 'id': "AngularJS Agent"},
- 'RepoMap': {'mode': True, 'id': "repomap"},
}
userSelectedModel = {
diff --git a/g4f/Provider/GizAI.py b/g4f/Provider/GizAI.py
new file mode 100644
index 00000000..127edc9e
--- /dev/null
+++ b/g4f/Provider/GizAI.py
@@ -0,0 +1,151 @@
+from __future__ import annotations
+
+import json
+from aiohttp import ClientSession
+
+from ..typing import AsyncResult, Messages
+from ..image import ImageResponse
+from .base_provider import AsyncGeneratorProvider, ProviderModelMixin
+from .helper import format_prompt
+
+class GizAI(AsyncGeneratorProvider, ProviderModelMixin):
+ url = "https://app.giz.ai/assistant/"
+ api_endpoint = "https://app.giz.ai/api/data/users/inferenceServer.infer"
+ working = True
+
+ supports_system_message = True
+ supports_message_history = True
+
+ # Chat models
+ default_model = 'chat-gemini-flash'
+ chat_models = [
+ default_model,
+ 'chat-gemini-pro',
+ 'chat-gpt4m',
+ 'chat-gpt4',
+ 'claude-sonnet',
+ 'claude-haiku',
+ 'llama-3-70b',
+ 'llama-3-8b',
+ 'mistral-large',
+ 'chat-o1-mini'
+ ]
+
+ # Image models
+ image_models = [
+ 'flux1',
+ 'sdxl',
+ 'sd',
+ 'sd35',
+ ]
+
+ models = [*chat_models, *image_models]
+
+ model_aliases = {
+ # Chat model aliases
+ "gemini-flash": "chat-gemini-flash",
+ "gemini-pro": "chat-gemini-pro",
+ "gpt-4o-mini": "chat-gpt4m",
+ "gpt-4o": "chat-gpt4",
+ "claude-3.5-sonnet": "claude-sonnet",
+ "claude-3-haiku": "claude-haiku",
+ "llama-3.1-70b": "llama-3-70b",
+ "llama-3.1-8b": "llama-3-8b",
+ "o1-mini": "chat-o1-mini",
+ # Image model aliases
+ "sd-1.5": "sd",
+ "sd-3.5": "sd35",
+ "flux-schnell": "flux1",
+ }
+
+ @classmethod
+ def get_model(cls, model: str) -> str:
+ if model in cls.models:
+ return model
+ elif model in cls.model_aliases:
+ return cls.model_aliases[model]
+ else:
+ return cls.default_model
+
+ @classmethod
+ def is_image_model(cls, model: str) -> bool:
+ return model in cls.image_models
+
+ @classmethod
+ async def create_async_generator(
+ cls,
+ model: str,
+ messages: Messages,
+ proxy: str = None,
+ **kwargs
+ ) -> AsyncResult:
+ model = cls.get_model(model)
+
+ headers = {
+ 'Accept': 'application/json, text/plain, */*',
+ 'Accept-Language': 'en-US,en;q=0.9',
+ 'Cache-Control': 'no-cache',
+ 'Connection': 'keep-alive',
+ 'Content-Type': 'application/json',
+ 'Origin': 'https://app.giz.ai',
+ 'Pragma': 'no-cache',
+ 'Sec-Fetch-Dest': 'empty',
+ 'Sec-Fetch-Mode': 'cors',
+ 'Sec-Fetch-Site': 'same-origin',
+ 'User-Agent': 'Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/130.0.0.0 Safari/537.36',
+ 'sec-ch-ua': '"Not?A_Brand";v="99", "Chromium";v="130"',
+ 'sec-ch-ua-mobile': '?0',
+ 'sec-ch-ua-platform': '"Linux"'
+ }
+
+ async with ClientSession() as session:
+ if cls.is_image_model(model):
+ # Image generation
+ prompt = messages[-1]["content"]
+ data = {
+ "model": model,
+ "input": {
+ "width": "1024",
+ "height": "1024",
+ "steps": 4,
+ "output_format": "webp",
+ "batch_size": 1,
+ "mode": "plan",
+ "prompt": prompt
+ }
+ }
+ async with session.post(
+ cls.api_endpoint,
+ headers=headers,
+ data=json.dumps(data),
+ proxy=proxy
+ ) as response:
+ response.raise_for_status()
+ response_data = await response.json()
+ if response_data.get('status') == 'completed' and response_data.get('output'):
+ for url in response_data['output']:
+ yield ImageResponse(images=url, alt="Generated Image")
+ else:
+ # Chat completion
+ data = {
+ "model": model,
+ "input": {
+ "messages": [
+ {
+ "type": "human",
+ "content": format_prompt(messages)
+ }
+ ],
+ "mode": "plan"
+ },
+ "noStream": True
+ }
+ async with session.post(
+ cls.api_endpoint,
+ headers=headers,
+ data=json.dumps(data),
+ proxy=proxy
+ ) as response:
+ response.raise_for_status()
+ result = await response.json()
+ yield result.get('output', '')
diff --git a/g4f/Provider/__init__.py b/g4f/Provider/__init__.py
index 8f36606b..1caf8aaf 100644
--- a/g4f/Provider/__init__.py
+++ b/g4f/Provider/__init__.py
@@ -47,6 +47,7 @@ from .FreeChatgpt import FreeChatgpt
from .FreeGpt import FreeGpt
from .FreeNetfly import FreeNetfly
from .GeminiPro import GeminiPro
+from .GizAI import GizAI
from .GPROChat import GPROChat
from .HuggingChat import HuggingChat
from .HuggingFace import HuggingFace
diff --git a/g4f/Provider/nexra/NexraChatGPT.py b/g4f/Provider/nexra/NexraChatGPT.py
index fc5051ee..074a0363 100644
--- a/g4f/Provider/nexra/NexraChatGPT.py
+++ b/g4f/Provider/nexra/NexraChatGPT.py
@@ -1,45 +1,52 @@
from __future__ import annotations
+import asyncio
import json
import requests
+from typing import Any, Dict
-from ...typing import CreateResult, Messages
-from ..base_provider import ProviderModelMixin, AbstractProvider
+from ...typing import AsyncResult, Messages
+from ..base_provider import AsyncGeneratorProvider, ProviderModelMixin
from ..helper import format_prompt
-class NexraChatGPT(AbstractProvider, ProviderModelMixin):
+
+class NexraChatGPT(AsyncGeneratorProvider, ProviderModelMixin):
label = "Nexra ChatGPT"
url = "https://nexra.aryahcr.cc/documentation/chatgpt/en"
- api_endpoint = "https://nexra.aryahcr.cc/api/chat/gpt"
+ api_endpoint_nexra_chatgpt = "https://nexra.aryahcr.cc/api/chat/gpt"
+ api_endpoint_nexra_chatgpt4o = "https://nexra.aryahcr.cc/api/chat/complements"
+ api_endpoint_nexra_chatgpt_v2 = "https://nexra.aryahcr.cc/api/chat/complements"
+ api_endpoint_nexra_gptweb = "https://nexra.aryahcr.cc/api/chat/gptweb"
working = True
+ supports_system_message = True
+ supports_message_history = True
+ supports_stream = True
default_model = 'gpt-3.5-turbo'
- models = ['gpt-4', 'gpt-4-0613', 'gpt-4-0314', 'gpt-4-32k-0314', default_model, 'gpt-3.5-turbo-16k', 'gpt-3.5-turbo-0613', 'gpt-3.5-turbo-16k-0613', 'gpt-3.5-turbo-0301', 'text-davinci-003', 'text-davinci-002', 'code-davinci-002', 'gpt-3', 'text-curie-001', 'text-babbage-001', 'text-ada-001', 'davinci', 'curie', 'babbage', 'ada', 'babbage-002', 'davinci-002']
+ nexra_chatgpt = [
+ 'gpt-4', 'gpt-4-0613', 'gpt-4-0314', 'gpt-4-32k-0314',
+ default_model, 'gpt-3.5-turbo-16k', 'gpt-3.5-turbo-0613', 'gpt-3.5-turbo-16k-0613', 'gpt-3.5-turbo-0301',
+ 'text-davinci-003', 'text-davinci-002', 'code-davinci-002', 'gpt-3', 'text-curie-001', 'text-babbage-001', 'text-ada-001', 'davinci', 'curie', 'babbage', 'ada', 'babbage-002', 'davinci-002'
+ ]
+ nexra_chatgpt4o = ['gpt-4o']
+ nexra_chatgptv2 = ['chatgpt']
+ nexra_gptweb = ['gptweb']
+ models = nexra_chatgpt + nexra_chatgpt4o + nexra_chatgptv2 + nexra_gptweb
model_aliases = {
"gpt-4": "gpt-4-0613",
- "gpt-4": "gpt-4-32k",
- "gpt-4": "gpt-4-0314",
- "gpt-4": "gpt-4-32k-0314",
-
+ "gpt-4-32k": "gpt-4-32k-0314",
"gpt-3.5-turbo": "gpt-3.5-turbo-16k",
- "gpt-3.5-turbo": "gpt-3.5-turbo-0613",
- "gpt-3.5-turbo": "gpt-3.5-turbo-16k-0613",
- "gpt-3.5-turbo": "gpt-3.5-turbo-0301",
-
+ "gpt-3.5-turbo-0613": "gpt-3.5-turbo-16k-0613",
"gpt-3": "text-davinci-003",
- "gpt-3": "text-davinci-002",
- "gpt-3": "code-davinci-002",
- "gpt-3": "text-curie-001",
- "gpt-3": "text-babbage-001",
- "gpt-3": "text-ada-001",
- "gpt-3": "text-ada-001",
- "gpt-3": "davinci",
- "gpt-3": "curie",
- "gpt-3": "babbage",
- "gpt-3": "ada",
- "gpt-3": "babbage-002",
- "gpt-3": "davinci-002",
+ "text-davinci-002": "code-davinci-002",
+ "text-curie-001": "text-babbage-001",
+ "text-ada-001": "davinci",
+ "curie": "babbage",
+ "ada": "babbage-002",
+ "davinci-002": "davinci-002",
+ "chatgpt": "chatgpt",
+ "gptweb": "gptweb"
}
@classmethod
@@ -50,40 +57,229 @@ class NexraChatGPT(AbstractProvider, ProviderModelMixin):
return cls.model_aliases[model]
else:
return cls.default_model
-
+
@classmethod
- def create_completion(
+ async def create_async_generator(
cls,
model: str,
messages: Messages,
+ stream: bool = False,
proxy: str = None,
markdown: bool = False,
**kwargs
- ) -> CreateResult:
- model = cls.get_model(model)
+ ) -> AsyncResult:
+ if model in cls.nexra_chatgpt:
+ async for chunk in cls._create_async_generator_nexra_chatgpt(model, messages, proxy, **kwargs):
+ yield chunk
+ elif model in cls.nexra_chatgpt4o:
+ async for chunk in cls._create_async_generator_nexra_chatgpt4o(model, messages, stream, proxy, markdown, **kwargs):
+ yield chunk
+ elif model in cls.nexra_chatgptv2:
+ async for chunk in cls._create_async_generator_nexra_chatgpt_v2(model, messages, stream, proxy, markdown, **kwargs):
+ yield chunk
+ elif model in cls.nexra_gptweb:
+ async for chunk in cls._create_async_generator_nexra_gptweb(model, messages, proxy, **kwargs):
+ yield chunk
+ @classmethod
+ async def _create_async_generator_nexra_chatgpt(
+ cls,
+ model: str,
+ messages: Messages,
+ proxy: str = None,
+ markdown: bool = False,
+ **kwargs
+ ) -> AsyncResult:
+ model = cls.get_model(model)
+
headers = {
- 'Content-Type': 'application/json'
+ "Content-Type": "application/json"
}
+ prompt = format_prompt(messages)
data = {
- "messages": [],
- "prompt": format_prompt(messages),
+ "messages": messages,
+ "prompt": prompt,
"model": model,
"markdown": markdown
}
+
+ loop = asyncio.get_event_loop()
+ try:
+ response = await loop.run_in_executor(None, cls._sync_post_request, cls.api_endpoint_nexra_chatgpt, data, headers, proxy)
+ filtered_response = cls._filter_response(response)
+
+ for chunk in filtered_response:
+ yield chunk
+ except Exception as e:
+ print(f"Error during API request (nexra_chatgpt): {e}")
+
+ @classmethod
+ async def _create_async_generator_nexra_chatgpt4o(
+ cls,
+ model: str,
+ messages: Messages,
+ stream: bool = False,
+ proxy: str = None,
+ markdown: bool = False,
+ **kwargs
+ ) -> AsyncResult:
+ model = cls.get_model(model)
- response = requests.post(cls.api_endpoint, headers=headers, json=data)
+ headers = {
+ "Content-Type": "application/json"
+ }
+
+ prompt = format_prompt(messages)
+ data = {
+ "messages": [
+ {
+ "role": "user",
+ "content": prompt
+ }
+ ],
+ "stream": stream,
+ "markdown": markdown,
+ "model": model
+ }
- return cls.process_response(response)
+ loop = asyncio.get_event_loop()
+ try:
+ response = await loop.run_in_executor(None, cls._sync_post_request, cls.api_endpoint_nexra_chatgpt4o, data, headers, proxy, stream)
+
+ if stream:
+ async for chunk in cls._process_streaming_response(response):
+ yield chunk
+ else:
+ for chunk in cls._process_non_streaming_response(response):
+ yield chunk
+ except Exception as e:
+ print(f"Error during API request (nexra_chatgpt4o): {e}")
@classmethod
- def process_response(cls, response):
+ async def _create_async_generator_nexra_chatgpt_v2(
+ cls,
+ model: str,
+ messages: Messages,
+ stream: bool = False,
+ proxy: str = None,
+ markdown: bool = False,
+ **kwargs
+ ) -> AsyncResult:
+ model = cls.get_model(model)
+
+ headers = {
+ "Content-Type": "application/json"
+ }
+
+ prompt = format_prompt(messages)
+ data = {
+ "messages": [
+ {
+ "role": "user",
+ "content": prompt
+ }
+ ],
+ "stream": stream,
+ "markdown": markdown,
+ "model": model
+ }
+
+ loop = asyncio.get_event_loop()
+ try:
+ response = await loop.run_in_executor(None, cls._sync_post_request, cls.api_endpoint_nexra_chatgpt_v2, data, headers, proxy, stream)
+
+ if stream:
+ async for chunk in cls._process_streaming_response(response):
+ yield chunk
+ else:
+ for chunk in cls._process_non_streaming_response(response):
+ yield chunk
+ except Exception as e:
+ print(f"Error during API request (nexra_chatgpt_v2): {e}")
+
+ @classmethod
+ async def _create_async_generator_nexra_gptweb(
+ cls,
+ model: str,
+ messages: Messages,
+ proxy: str = None,
+ markdown: bool = False,
+ **kwargs
+ ) -> AsyncResult:
+ model = cls.get_model(model)
+
+ headers = {
+ "Content-Type": "application/json"
+ }
+
+ prompt = format_prompt(messages)
+ data = {
+ "prompt": prompt,
+ "markdown": markdown,
+ }
+
+ loop = asyncio.get_event_loop()
+ try:
+ response = await loop.run_in_executor(None, cls._sync_post_request, cls.api_endpoint_nexra_gptweb, data, headers, proxy)
+
+ for chunk in response.iter_content(1024):
+ if chunk:
+ decoded_chunk = chunk.decode().lstrip('_')
+ try:
+ response_json = json.loads(decoded_chunk)
+ if response_json.get("status"):
+ yield response_json.get("gpt", "")
+ except json.JSONDecodeError:
+ continue
+ except Exception as e:
+ print(f"Error during API request (nexra_gptweb): {e}")
+
+ @staticmethod
+ def _sync_post_request(url: str, data: Dict[str, Any], headers: Dict[str, str], proxy: str = None, stream: bool = False) -> requests.Response:
+ proxies = {
+ "http": proxy,
+ "https": proxy,
+ } if proxy else None
+
+ try:
+ response = requests.post(url, json=data, headers=headers, proxies=proxies, stream=stream)
+ response.raise_for_status()
+ return response
+ except requests.RequestException as e:
+ print(f"Request failed: {e}")
+ raise
+
+ @staticmethod
+ def _process_non_streaming_response(response: requests.Response) -> str:
if response.status_code == 200:
try:
- data = response.json()
- return data.get('gpt', '')
+ content = response.text.lstrip('')
+ data = json.loads(content)
+ return data.get('message', '')
except json.JSONDecodeError:
return "Error: Unable to decode JSON response"
else:
return f"Error: {response.status_code}"
+
+ @staticmethod
+ async def _process_streaming_response(response: requests.Response):
+ full_message = ""
+ for line in response.iter_lines(decode_unicode=True):
+ if line:
+ try:
+ line = line.lstrip('')
+ data = json.loads(line)
+ if data.get('finish'):
+ break
+ message = data.get('message', '')
+ if message:
+ yield message[len(full_message):]
+ full_message = message
+ except json.JSONDecodeError:
+ pass
+
+ @staticmethod
+ def _filter_response(response: requests.Response) -> str:
+ response_json = response.json()
+ return response_json.get("gpt", "")
diff --git a/g4f/Provider/nexra/NexraChatGPT4o.py b/g4f/Provider/nexra/NexraChatGPT4o.py
deleted file mode 100644
index 126d32b8..00000000
--- a/g4f/Provider/nexra/NexraChatGPT4o.py
+++ /dev/null
@@ -1,86 +0,0 @@
-from __future__ import annotations
-
-import json
-import requests
-
-from ...typing import CreateResult, Messages
-from ..base_provider import ProviderModelMixin, AbstractProvider
-from ..helper import format_prompt
-
-class NexraChatGPT4o(AbstractProvider, ProviderModelMixin):
- label = "Nexra ChatGPT4o"
- url = "https://nexra.aryahcr.cc/documentation/chatgpt/en"
- api_endpoint = "https://nexra.aryahcr.cc/api/chat/complements"
- working = True
- supports_stream = True
-
- default_model = "gpt-4o"
- models = [default_model]
-
- @classmethod
- def get_model(cls, model: str) -> str:
- return cls.default_model
-
- @classmethod
- def create_completion(
- cls,
- model: str,
- messages: Messages,
- stream: bool,
- proxy: str = None,
- markdown: bool = False,
- **kwargs
- ) -> CreateResult:
- model = cls.get_model(model)
-
- headers = {
- 'Content-Type': 'application/json'
- }
-
- data = {
- "messages": [
- {
- "role": "user",
- "content": format_prompt(messages)
- }
- ],
- "stream": stream,
- "markdown": markdown,
- "model": model
- }
-
- response = requests.post(cls.api_endpoint, headers=headers, json=data, stream=stream)
-
- if stream:
- return cls.process_streaming_response(response)
- else:
- return cls.process_non_streaming_response(response)
-
- @classmethod
- def process_non_streaming_response(cls, response):
- if response.status_code == 200:
- try:
- content = response.text.lstrip('')
- data = json.loads(content)
- return data.get('message', '')
- except json.JSONDecodeError:
- return "Error: Unable to decode JSON response"
- else:
- return f"Error: {response.status_code}"
-
- @classmethod
- def process_streaming_response(cls, response):
- full_message = ""
- for line in response.iter_lines(decode_unicode=True):
- if line:
- try:
- line = line.lstrip('')
- data = json.loads(line)
- if data.get('finish'):
- break
- message = data.get('message', '')
- if message and message != full_message:
- yield message[len(full_message):]
- full_message = message
- except json.JSONDecodeError:
- pass
diff --git a/g4f/Provider/nexra/NexraChatGptV2.py b/g4f/Provider/nexra/NexraChatGptV2.py
deleted file mode 100644
index 1ff42705..00000000
--- a/g4f/Provider/nexra/NexraChatGptV2.py
+++ /dev/null
@@ -1,92 +0,0 @@
-from __future__ import annotations
-
-import json
-import requests
-
-from ...typing import CreateResult, Messages
-from ..base_provider import ProviderModelMixin, AbstractProvider
-from ..helper import format_prompt
-
-class NexraChatGptV2(AbstractProvider, ProviderModelMixin):
- label = "Nexra ChatGPT v2"
- url = "https://nexra.aryahcr.cc/documentation/chatgpt/en"
- api_endpoint = "https://nexra.aryahcr.cc/api/chat/complements"
- working = True
- supports_stream = True
-
- default_model = 'chatgpt'
- models = [default_model]
- model_aliases = {"gpt-4": "chatgpt"}
-
- @classmethod
- def get_model(cls, model: str) -> str:
- if model in cls.models:
- return model
- elif model in cls.model_aliases:
- return cls.model_aliases[model]
- else:
- return cls.default_model
-
- @classmethod
- def create_completion(
- cls,
- model: str,
- messages: Messages,
- stream: bool,
- proxy: str = None,
- markdown: bool = False,
- **kwargs
- ) -> CreateResult:
- model = cls.get_model(model)
-
- headers = {
- 'Content-Type': 'application/json'
- }
-
- data = {
- "messages": [
- {
- "role": "user",
- "content": format_prompt(messages)
- }
- ],
- "stream": stream,
- "markdown": markdown,
- "model": model
- }
-
- response = requests.post(cls.api_endpoint, headers=headers, json=data, stream=stream)
-
- if stream:
- return cls.process_streaming_response(response)
- else:
- return cls.process_non_streaming_response(response)
-
- @classmethod
- def process_non_streaming_response(cls, response):
- if response.status_code == 200:
- try:
- content = response.text.lstrip('')
- data = json.loads(content)
- return data.get('message', '')
- except json.JSONDecodeError:
- return "Error: Unable to decode JSON response"
- else:
- return f"Error: {response.status_code}"
-
- @classmethod
- def process_streaming_response(cls, response):
- full_message = ""
- for line in response.iter_lines(decode_unicode=True):
- if line:
- try:
- line = line.lstrip('')
- data = json.loads(line)
- if data.get('finish'):
- break
- message = data.get('message', '')
- if message:
- yield message[len(full_message):]
- full_message = message
- except json.JSONDecodeError:
- pass
diff --git a/g4f/Provider/nexra/NexraChatGptWeb.py b/g4f/Provider/nexra/NexraChatGptWeb.py
deleted file mode 100644
index f82694d4..00000000
--- a/g4f/Provider/nexra/NexraChatGptWeb.py
+++ /dev/null
@@ -1,64 +0,0 @@
-from __future__ import annotations
-
-import json
-import requests
-
-from ...typing import CreateResult, Messages
-from ..base_provider import ProviderModelMixin, AbstractProvider
-from ..helper import format_prompt
-
-class NexraChatGptWeb(AbstractProvider, ProviderModelMixin):
- label = "Nexra ChatGPT Web"
- url = "https://nexra.aryahcr.cc/documentation/chatgpt/en"
- working = True
-
- default_model = "gptweb"
- models = [default_model]
- model_aliases = {"gpt-4": "gptweb"}
- api_endpoints = {"gptweb": "https://nexra.aryahcr.cc/api/chat/gptweb"}
-
- @classmethod
- def get_model(cls, model: str) -> str:
- if model in cls.models:
- return model
- elif model in cls.model_aliases:
- return cls.model_aliases[model]
- else:
- return cls.default_model
-
- @classmethod
- def create_completion(
- cls,
- model: str,
- messages: Messages,
- proxy: str = None,
- markdown: bool = False,
- **kwargs
- ) -> CreateResult:
- model = cls.get_model(model)
- api_endpoint = cls.api_endpoints.get(model, cls.api_endpoints[cls.default_model])
-
- headers = {
- 'Content-Type': 'application/json'
- }
-
- data = {
- "prompt": format_prompt(messages),
- "markdown": markdown
- }
-
- response = requests.post(api_endpoint, headers=headers, json=data)
-
- return cls.process_response(response)
-
- @classmethod
- def process_response(cls, response):
- if response.status_code == 200:
- try:
- content = response.text.lstrip('_')
- json_response = json.loads(content)
- return json_response.get('gpt', '')
- except json.JSONDecodeError:
- return "Error: Unable to decode JSON response"
- else:
- return f"Error: {response.status_code}"
diff --git a/g4f/Provider/nexra/__init__.py b/g4f/Provider/nexra/__init__.py
index 6121fdc0..bebc1fb6 100644
--- a/g4f/Provider/nexra/__init__.py
+++ b/g4f/Provider/nexra/__init__.py
@@ -1,9 +1,6 @@
from .NexraBing import NexraBing
from .NexraBlackbox import NexraBlackbox
from .NexraChatGPT import NexraChatGPT
-from .NexraChatGPT4o import NexraChatGPT4o
-from .NexraChatGptV2 import NexraChatGptV2
-from .NexraChatGptWeb import NexraChatGptWeb
from .NexraDallE import NexraDallE
from .NexraDallE2 import NexraDallE2
from .NexraEmi import NexraEmi
diff --git a/g4f/api/__init__.py b/g4f/api/__init__.py
index 83df469a..754a48f1 100644
--- a/g4f/api/__init__.py
+++ b/g4f/api/__init__.py
@@ -14,17 +14,18 @@ from starlette.status import HTTP_422_UNPROCESSABLE_ENTITY, HTTP_401_UNAUTHORIZE
from fastapi.encoders import jsonable_encoder
from fastapi.middleware.cors import CORSMiddleware
from pydantic import BaseModel
-from typing import Union, Optional
+from typing import Union, Optional, Iterator
import g4f
import g4f.debug
-from g4f.client import Client
+from g4f.client import Client, ChatCompletion, ChatCompletionChunk, ImagesResponse
from g4f.typing import Messages
from g4f.cookies import read_cookie_files
-def create_app():
+def create_app(g4f_api_key: str = None):
app = FastAPI()
- api = Api(app)
+
+ # Add CORS middleware
app.add_middleware(
CORSMiddleware,
allow_origin_regex=".*",
@@ -32,18 +33,19 @@ def create_app():
allow_methods=["*"],
allow_headers=["*"],
)
+
+ api = Api(app, g4f_api_key=g4f_api_key)
api.register_routes()
api.register_authorization()
api.register_validation_exception_handler()
+
+ # Read cookie files if not ignored
if not AppConfig.ignore_cookie_files:
read_cookie_files()
- return app
-def create_app_debug():
- g4f.debug.logging = True
- return create_app()
+ return app
-class ChatCompletionsForm(BaseModel):
+class ChatCompletionsConfig(BaseModel):
messages: Messages
model: str
provider: Optional[str] = None
@@ -55,15 +57,12 @@ class ChatCompletionsForm(BaseModel):
web_search: Optional[bool] = None
proxy: Optional[str] = None
-class ImagesGenerateForm(BaseModel):
- model: Optional[str] = None
- provider: Optional[str] = None
+class ImageGenerationConfig(BaseModel):
prompt: str
- response_format: Optional[str] = None
- api_key: Optional[str] = None
- proxy: Optional[str] = None
+ model: Optional[str] = None
+ response_format: str = "url"
-class AppConfig():
+class AppConfig:
ignored_providers: Optional[list[str]] = None
g4f_api_key: Optional[str] = None
ignore_cookie_files: bool = False
@@ -74,16 +73,23 @@ class AppConfig():
for key, value in data.items():
setattr(cls, key, value)
+list_ignored_providers: list[str] = None
+
+def set_list_ignored_providers(ignored: list[str]):
+ global list_ignored_providers
+ list_ignored_providers = ignored
+
class Api:
- def __init__(self, app: FastAPI) -> None:
+ def __init__(self, app: FastAPI, g4f_api_key=None) -> None:
self.app = app
self.client = Client()
+ self.g4f_api_key = g4f_api_key
self.get_g4f_api_key = APIKeyHeader(name="g4f-api-key")
def register_authorization(self):
@self.app.middleware("http")
async def authorization(request: Request, call_next):
- if AppConfig.g4f_api_key and request.url.path in ["/v1/chat/completions", "/v1/completions"]:
+ if self.g4f_api_key and request.url.path in ["/v1/chat/completions", "/v1/completions", "/v1/images/generate"]:
try:
user_g4f_api_key = await self.get_g4f_api_key(request)
except HTTPException as e:
@@ -92,22 +98,26 @@ class Api:
status_code=HTTP_401_UNAUTHORIZED,
content=jsonable_encoder({"detail": "G4F API key required"}),
)
- if not secrets.compare_digest(AppConfig.g4f_api_key, user_g4f_api_key):
+ if not secrets.compare_digest(self.g4f_api_key, user_g4f_api_key):
return JSONResponse(
status_code=HTTP_403_FORBIDDEN,
content=jsonable_encoder({"detail": "Invalid G4F API key"}),
)
- return await call_next(request)
+
+ response = await call_next(request)
+ return response
def register_validation_exception_handler(self):
@self.app.exception_handler(RequestValidationError)
async def validation_exception_handler(request: Request, exc: RequestValidationError):
details = exc.errors()
- modified_details = [{
- "loc": error["loc"],
- "message": error["msg"],
- "type": error["type"],
- } for error in details]
+ modified_details = []
+ for error in details:
+ modified_details.append({
+ "loc": error["loc"],
+ "message": error["msg"],
+ "type": error["type"],
+ })
return JSONResponse(
status_code=HTTP_422_UNPROCESSABLE_ENTITY,
content=jsonable_encoder({"detail": modified_details}),
@@ -121,25 +131,23 @@ class Api:
@self.app.get("/v1")
async def read_root_v1():
return HTMLResponse('g4f API: Go to '
- '<a href="/v1/chat/completions">chat/completions</a> '
- 'or <a href="/v1/models">models</a>.')
+ '<a href="/v1/chat/completions">chat/completions</a>, '
+ '<a href="/v1/models">models</a>, or '
+ '<a href="/v1/images/generate">images/generate</a>.')
@self.app.get("/v1/models")
async def models():
- model_list = {
- model: g4f.models.ModelUtils.convert[model]
+ model_list = dict(
+ (model, g4f.models.ModelUtils.convert[model])
for model in g4f.Model.__all__()
- }
+ )
model_list = [{
'id': model_id,
'object': 'model',
'created': 0,
'owned_by': model.base_provider
} for model_id, model in model_list.items()]
- return JSONResponse({
- "object": "list",
- "data": model_list,
- })
+ return JSONResponse(model_list)
@self.app.get("/v1/models/{model_name}")
async def model_info(model_name: str):
@@ -155,7 +163,7 @@ class Api:
return JSONResponse({"error": "The model does not exist."})
@self.app.post("/v1/chat/completions")
- async def chat_completions(config: ChatCompletionsForm, request: Request = None, provider: str = None):
+ async def chat_completions(config: ChatCompletionsConfig, request: Request = None, provider: str = None):
try:
config.provider = provider if config.provider is None else config.provider
if config.api_key is None and request is not None:
@@ -164,17 +172,27 @@ class Api:
auth_header = auth_header.split(None, 1)[-1]
if auth_header and auth_header != "Bearer":
config.api_key = auth_header
- # Use the asynchronous create method and await it
- response = await self.client.chat.completions.async_create(
+
+ # Create the completion response
+ response = self.client.chat.completions.create(
**{
**AppConfig.defaults,
**config.dict(exclude_none=True),
},
ignored=AppConfig.ignored_providers
)
- if not config.stream:
+
+ # Check if the response is synchronous or asynchronous
+ if isinstance(response, ChatCompletion):
+ # Synchronous response
return JSONResponse(response.to_json())
+ if not config.stream:
+ # If the response is an iterator but not streaming, collect the result
+ response_list = list(response) if isinstance(response, Iterator) else [response]
+ return JSONResponse(response_list[0].to_json())
+
+ # Streaming response
async def streaming():
try:
async for chunk in response:
@@ -185,41 +203,38 @@ class Api:
logging.exception(e)
yield f'data: {format_exception(e, config)}\n\n'
yield "data: [DONE]\n\n"
+
return StreamingResponse(streaming(), media_type="text/event-stream")
except Exception as e:
logging.exception(e)
return Response(content=format_exception(e, config), status_code=500, media_type="application/json")
- @self.app.post("/v1/completions")
- async def completions():
- return Response(content=json.dumps({'info': 'Not working yet.'}, indent=4), media_type="application/json")
-
- @self.app.post("/v1/images/generations")
- async def images_generate(config: ImagesGenerateForm, request: Request = None, provider: str = None):
+ @self.app.post("/v1/images/generate")
+ async def generate_image(config: ImageGenerationConfig):
try:
- config.provider = provider if config.provider is None else config.provider
- if config.api_key is None and request is not None:
- auth_header = request.headers.get("Authorization")
- if auth_header is not None:
- auth_header = auth_header.split(None, 1)[-1]
- if auth_header and auth_header != "Bearer":
- config.api_key = auth_header
- # Use the asynchronous generate method and await it
- response = await self.client.images.async_generate(
- **config.dict(exclude_none=True),
+ response: ImagesResponse = await self.client.images.async_generate(
+ prompt=config.prompt,
+ model=config.model,
+ response_format=config.response_format
)
- return JSONResponse(response.to_json())
+ # Convert Image objects to dictionaries
+ response_data = [image.to_dict() for image in response.data]
+ return JSONResponse({"data": response_data})
except Exception as e:
logging.exception(e)
return Response(content=format_exception(e, config), status_code=500, media_type="application/json")
-def format_exception(e: Exception, config: ChatCompletionsForm) -> str:
+ @self.app.post("/v1/completions")
+ async def completions():
+ return Response(content=json.dumps({'info': 'Not working yet.'}, indent=4), media_type="application/json")
+
+def format_exception(e: Exception, config: Union[ChatCompletionsConfig, ImageGenerationConfig]) -> str:
last_provider = g4f.get_last_provider(True)
return json.dumps({
"error": {"message": f"{e.__class__.__name__}: {e}"},
- "model": last_provider.get("model") if last_provider else config.model,
- "provider": last_provider.get("name") if last_provider else config.provider
+ "model": last_provider.get("model") if last_provider else getattr(config, 'model', None),
+ "provider": last_provider.get("name") if last_provider else getattr(config, 'provider', None)
})
def run_api(
@@ -228,18 +243,22 @@ def run_api(
bind: str = None,
debug: bool = False,
workers: int = None,
- use_colors: bool = None
+ use_colors: bool = None,
+ g4f_api_key: str = None
) -> None:
print(f'Starting server... [g4f v-{g4f.version.utils.current_version}]' + (" (debug)" if debug else ""))
if use_colors is None:
use_colors = debug
if bind is not None:
host, port = bind.split(":")
+ if debug:
+ g4f.debug.logging = True
uvicorn.run(
- f"g4f.api:create_app{'_debug' if debug else ''}",
- host=host, port=int(port),
- workers=workers,
- use_colors=use_colors,
- factory=True,
+ "g4f.api:create_app",
+ host=host,
+ port=int(port),
+ workers=workers,
+ use_colors=use_colors,
+ factory=True,
reload=debug
)
diff --git a/g4f/client/client.py b/g4f/client/client.py
index 41238df5..2772f9bb 100644
--- a/g4f/client/client.py
+++ b/g4f/client/client.py
@@ -149,6 +149,7 @@ class Completions:
self,
messages: Messages,
model: str,
+ system: str = None, # Added system parameter
provider: ProviderType = None,
stream: bool = False,
proxy: str = None,
@@ -161,6 +162,12 @@ class Completions:
ignore_stream: bool = False,
**kwargs
) -> Union[ChatCompletion, Iterator[ChatCompletionChunk]]:
+ # If a system prompt is provided, prepend it to the messages
+ if system:
+ system_message = {"role": "system", "content": system}
+ messages = [system_message] + messages
+
+ # Existing implementation continues...
model, provider = get_model_and_provider(
model,
self.provider if provider is None else provider,
@@ -221,6 +228,7 @@ class Completions:
self,
messages: Messages,
model: str,
+ system: str = None, # Added system parameter
provider: ProviderType = None,
stream: bool = False,
proxy: str = None,
@@ -233,6 +241,12 @@ class Completions:
ignore_stream: bool = False,
**kwargs
) -> Union[ChatCompletion, AsyncIterator[ChatCompletionChunk]]:
+ # If a system prompt is provided, prepend it to the messages
+ if system:
+ system_message = {"role": "system", "content": system}
+ messages = [system_message] + messages
+
+ # Existing implementation continues...
model, provider = get_model_and_provider(
model,
self.provider if provider is None else provider,
@@ -271,16 +285,18 @@ class Completions:
**kwargs
)
- # Removed 'await' here since 'async_iter_response' returns an async generator
- response = async_iter_response(response, stream, response_format, max_tokens, stop)
- response = async_iter_append_model_and_provider(response)
-
+ # Handle streaming or non-streaming responses
if stream:
+ response = async_iter_response(response, stream, response_format, max_tokens, stop)
+ response = async_iter_append_model_and_provider(response)
return response
else:
+ response = async_iter_response(response, stream, response_format, max_tokens, stop)
+ response = async_iter_append_model_and_provider(response)
async for result in response:
return result
+
class Chat:
completions: Completions
@@ -401,6 +417,12 @@ class Image:
def __repr__(self):
return f"Image(url={self.url}, b64_json={'<base64 data>' if self.b64_json else None})"
+ def to_dict(self):
+ return {
+ "url": self.url,
+ "b64_json": self.b64_json
+ }
+
class ImagesResponse:
def __init__(self, data: list[Image]):
self.data = data
diff --git a/g4f/gui/client/index.html b/g4f/gui/client/index.html
index 1a660062..f8c11ea2 100644
--- a/g4f/gui/client/index.html
+++ b/g4f/gui/client/index.html
@@ -1,258 +1,272 @@
<!DOCTYPE html>
<html lang="en" data-framework="javascript">
-
-<head>
- <meta charset="UTF-8">
- <meta http-equiv="X-UA-Compatible" content="IE=edge">
- <meta name="viewport" content="width=device-width, initial-scale=1.0 maximum-scale=1.0">
- <meta name="description" content="A conversational AI system that listens, learns, and challenges">
- <meta property="og:title" content="ChatGPT">
- <meta property="og:image" content="https://openai.com/content/images/2022/11/ChatGPT.jpg">
- <meta property="og:description" content="A conversational AI system that listens, learns, and challenges">
- <meta property="og:url" content="https://g4f.ai">
- <link rel="stylesheet" href="/static/css/style.css">
- <link rel="apple-touch-icon" sizes="180x180" href="/static/img/apple-touch-icon.png">
- <link rel="icon" type="image/png" sizes="32x32" href="/static/img/favicon-32x32.png">
- <link rel="icon" type="image/png" sizes="16x16" href="/static/img/favicon-16x16.png">
- <link rel="manifest" href="/static/img/site.webmanifest">
- <script src="/static/js/icons.js"></script>
- <script src="/static/js/highlightjs-copy.min.js"></script>
- <script src="/static/js/chat.v1.js" defer></script>
- <script src="https://cdn.jsdelivr.net/npm/markdown-it@13.0.1/dist/markdown-it.min.js"></script>
- <link rel="stylesheet" href="/static/css/dracula.min.css">
- <script>
- MathJax = {
- chtml: {
- scale: 1,
- displayAlign: 'left'
- }
- };
- </script>
- <script id="MathJax-script" src="https://cdn.jsdelivr.net/npm/mathjax@3/es5/tex-mml-chtml.js" async></script>
- <script type="module" src="https://cdn.jsdelivr.net/npm/mistral-tokenizer-js" async>
- import mistralTokenizer from "mistral-tokenizer-js"
- </script>
- <script type="module" src="https://cdn.jsdelivr.net/gh/belladoreai/llama-tokenizer-js@master/llama-tokenizer.js" async>
- import llamaTokenizer from "llama-tokenizer-js"
- </script>
- <script src="https://cdn.jsdelivr.net/npm/gpt-tokenizer/dist/cl100k_base.js" async></script>
- <script src="/static/js/text_to_speech/index.js" async></script>
- <!--
- <script src="/static/js/whisper-web/index.js" async></script>
- -->
- <script>
- const user_image = '<img src="/static/img/user.png" alt="your avatar">';
- const gpt_image = '<img src="/static/img/gpt.png" alt="your avatar">';
- </script>
- <script src="/static/js/highlight.min.js"></script>
- <script>window.conversation_id = "{{chat_id}}"</script>
- <title>g4f - gui</title>
-</head>
-
-<body>
- <div class="gradient"></div>
- <div class="row">
- <div class="box conversations">
+ <head>
+ <meta charset="UTF-8">
+ <meta http-equiv="X-UA-Compatible" content="IE=edge">
+ <meta name="viewport" content="width=device-width, initial-scale=1.0 maximum-scale=1.0">
+ <meta name="description" content="A conversational AI system that listens, learns, and challenges">
+ <meta property="og:title" content="ChatGPT">
+ <meta property="og:image" content="https://openai.com/content/images/2022/11/ChatGPT.jpg">
+ <meta property="og:description" content="A conversational AI system that listens, learns, and challenges">
+ <meta property="og:url" content="https://g4f.ai">
+ <link rel="stylesheet" href="/static/css/style.css">
+ <link rel="apple-touch-icon" sizes="180x180" href="/static/img/apple-touch-icon.png">
+ <link rel="icon" type="image/png" sizes="32x32" href="/static/img/favicon-32x32.png">
+ <link rel="icon" type="image/png" sizes="16x16" href="/static/img/favicon-16x16.png">
+ <link rel="manifest" href="/static/img/site.webmanifest">
+ <script src="/static/js/icons.js"></script>
+ <script src="/static/js/highlightjs-copy.min.js"></script>
+ <script src="/static/js/chat.v1.js" defer></script>
+ <script src="https://cdn.jsdelivr.net/npm/markdown-it@13.0.1/dist/markdown-it.min.js"></script>
+ <link rel="stylesheet" href="/static/css/dracula.min.css">
+ <script>
+ MathJax = {
+ chtml: {
+ scale: 1,
+ displayAlign: 'left'
+ }
+ };
+ </script>
+ <script id="MathJax-script" src="https://cdn.jsdelivr.net/npm/mathjax@3/es5/tex-mml-chtml.js" async></script>
+ <script type="module" src="https://cdn.jsdelivr.net/npm/mistral-tokenizer-js" async>
+ import mistralTokenizer from "mistral-tokenizer-js"
+ </script>
+ <script type="module" src="https://cdn.jsdelivr.net/gh/belladoreai/llama-tokenizer-js@master/llama-tokenizer.js" async>
+ import llamaTokenizer from "llama-tokenizer-js"
+ </script>
+ <script src="https://cdn.jsdelivr.net/npm/gpt-tokenizer/dist/cl100k_base.js" async></script>
+ <script src="/static/js/text_to_speech/index.js" async></script>
+ <!--
+ <script src="/static/js/whisper-web/index.js" async></script>
+ -->
+ <script>
+ const user_image = '<img src="/static/img/user.png" alt="your avatar">';
+ const gpt_image = '<img src="/static/img/gpt.png" alt="your avatar">';
+ </script>
+ <script src="/static/js/highlight.min.js"></script>
+ <script>window.conversation_id = "{{chat_id}}"</script>
+ <title>g4f - gui</title>
+ </head>
+ <body>
+ <div class="gradient"></div>
+ <div class="row">
+ <div class="box conversations">
<div class="top">
- <button class="new_convo" onclick="new_conversation()">
- <i class="fa-regular fa-plus"></i>
- <span>New Conversation</span>
- </button>
+ <button class="new_convo" onclick="new_conversation()">
+ <i class="fa-regular fa-plus"></i>
+ <span>New Conversation</span>
+ </button>
</div>
<div class="bottom_buttons">
- <!--
- <button onclick="open_album();">
- <i class="fa-solid fa-toolbox"></i>
- <span>Images Album</span>
- </button>
- -->
- <button onclick="open_settings();">
- <i class="fa-solid fa-toolbox"></i>
- <span>Open Settings</span>
- </button>
- <div class="info">
- <i class="fa-brands fa-discord"></i>
- <span class="convo-title">discord ~ <a href="https://discord.gg/XfybzPXPH5">discord.gg/XfybzPXPH5</a>
- </span>
- </div>
- <div class="info">
- <i class="fa-brands fa-github"></i>
- <span class="convo-title">github ~ <a href="https://github.com/xtekky/gpt4free">@xtekky/gpt4free</a>
- </span>
- </div>
- <div class="info">
- <i class="fa-solid fa-star"></i>
- <span id="version_text" class="convo-title"></span>
- </div>
+ <button onclick="open_album();">
+ <i class="fa-solid fa-toolbox"></i>
+ <span>Images Album</span>
+ </button>
+ <button onclick="open_settings();">
+ <i class="fa-solid fa-toolbox"></i>
+ <span>Open Settings</span>
+ </button>
+ <div class="info">
+ <i class="fa-brands fa-discord"></i>
+ <span class="convo-title">discord ~ <a href="https://discord.gg/XfybzPXPH5">discord.gg/XfybzPXPH5</a>
+ </span>
+ </div>
+ <div class="info">
+ <i class="fa-brands fa-github"></i>
+ <span class="convo-title">github ~ <a href="https://github.com/xtekky/gpt4free">@xtekky/gpt4free</a>
+ </span>
+ </div>
+ <div class="info">
+ <i class="fa-solid fa-star"></i>
+ <span id="version_text" class="convo-title"></span>
+ </div>
</div>
- </div>
- <div class="images hidden">
-
- </div>
- <div class="settings hidden">
+ </div>
+ <div class="images hidden">
+ </div>
+ <div class="settings hidden">
<div class="paper">
- <h3>Settings</h3>
- <div class="field">
- <span class="label">Enable Dark Mode</span>
- <input type="checkbox" id="darkMode" checked />
- <label for="darkMode" class="toogle" title=""></label>
- </div>
- <div class="field">
- <span class="label">Web Access with DuckDuckGo</span>
- <input type="checkbox" id="switch" />
- <label for="switch" class="toogle" title="Add the pages of the first 5 search results to the query."></label>
- </div>
- <div class="field">
- <span class="label">Disable Conversation History</span>
- <input type="checkbox" id="history" />
- <label for="history" class="toogle" title="To improve the reaction time or if you have trouble with large conversations."></label>
- </div>
- <div class="field">
- <span class="label">Hide System-prompt</span>
- <input type="checkbox" id="hide-systemPrompt" />
- <label for="hide-systemPrompt" class="toogle" title="For more space on phones"></label>
- </div>
- <div class="field">
- <span class="label">Auto continue in ChatGPT</span>
- <input id="auto_continue" type="checkbox" name="auto_continue" checked/>
- <label for="auto_continue" class="toogle" title="Continue large responses in OpenaiChat"></label>
- </div>
- <div class="field box">
- <label for="message-input-height" class="label" title="">Input max. height</label>
- <input type="number" id="message-input-height" value="200"/>
- </div>
- <div class="field box">
- <label for="recognition-language" class="label" title="">Speech recognition lang</label>
- <input type="text" id="recognition-language" value="" placeholder="navigator.language"/>
- </div>
- <div class="field box">
- <label for="BingCreateImages-api_key" class="label" title="">Microsoft Designer in Bing:</label>
- <textarea id="BingCreateImages-api_key" name="BingCreateImages[api_key]" placeholder="&quot;_U&quot; cookie"></textarea>
- </div>
- <div class="field box">
- <label for="DeepInfra-api_key" class="label" title="">DeepInfra:</label>
- <textarea id="DeepInfra-api_key" name="DeepInfra[api_key]" class="DeepInfraImage-api_key" placeholder="api_key"></textarea>
- </div>
- <div class="field box">
- <label for="GeminiPro-api_key" class="label" title="">Gemini API:</label>
- <textarea id="GeminiPro-api_key" name="GeminiPro[api_key]" placeholder="api_key"></textarea>
- </div>
- <div class="field box">
- <label for="Groq-api_key" class="label" title="">Groq:</label>
- <textarea id="Groq-api_key" name="Groq[api_key]" placeholder="api_key"></textarea>
- </div>
- <div class="field box">
- <label for="HuggingFace-api_key" class="label" title="">HuggingFace:</label>
- <textarea id="HuggingFace-api_key" name="HuggingFace[api_key]" placeholder="api_key"></textarea>
- </div>
- <div class="field box">
- <label for="Openai-api_key" class="label" title="">OpenAI API:</label>
- <textarea id="Openai-api_key" name="Openai[api_key]" placeholder="api_key"></textarea>
- </div>
- <div class="field box">
- <label for="OpenRouter-api_key" class="label" title="">OpenRouter:</label>
- <textarea id="OpenRouter-api_key" name="OpenRouter[api_key]" placeholder="api_key"></textarea>
- </div>
- <div class="field box">
- <label for="PerplexityApi-api_key" class="label" title="">Perplexity API:</label>
- <textarea id="PerplexityApi-api_key" name="PerplexityApi[api_key]" placeholder="api_key"></textarea>
- </div>
- <div class="field box">
- <label for="Replicate-api_key" class="label" title="">Replicate:</label>
- <textarea id="Replicate-api_key" name="Replicate[api_key]" class="ReplicateImage-api_key" placeholder="api_key"></textarea>
- </div>
+ <h3>Settings</h3>
+ <div class="field">
+ <span class="label">Enable Dark Mode</span>
+ <input type="checkbox" id="darkMode" checked />
+ <label for="darkMode" class="toogle" title=""></label>
+ </div>
+ <div class="field">
+ <span class="label">Web Access with DuckDuckGo</span>
+ <input type="checkbox" id="switch" />
+ <label for="switch" class="toogle" title="Add the pages of the first 5 search results to the query."></label>
+ </div>
+ <div class="field">
+ <span class="label">Disable Conversation History</span>
+ <input type="checkbox" id="history" />
+ <label for="history" class="toogle" title="To improve the reaction time or if you have trouble with large conversations."></label>
+ </div>
+ <div class="field">
+ <span class="label">Hide System-prompt</span>
+ <input type="checkbox" id="hide-systemPrompt" />
+ <label for="hide-systemPrompt" class="toogle" title="For more space on phones"></label>
+ </div>
+ <div class="field">
+ <span class="label">Auto continue in ChatGPT</span>
+ <input id="auto_continue" type="checkbox" name="auto_continue" checked/>
+ <label for="auto_continue" class="toogle" title="Continue large responses in OpenaiChat"></label>
+ </div>
+ <div class="field box">
+ <label for="message-input-height" class="label" title="">Input max. height</label>
+ <input type="number" id="message-input-height" value="200"/>
+ </div>
+ <div class="field box">
+ <label for="recognition-language" class="label" title="">Speech recognition lang</label>
+ <input type="text" id="recognition-language" value="" placeholder="navigator.language"/>
+ </div>
+ <div class="field box">
+ <label for="BingCreateImages-api_key" class="label" title="">Microsoft Designer in Bing:</label>
+ <textarea id="BingCreateImages-api_key" name="BingCreateImages[api_key]" placeholder="&quot;_U&quot; cookie"></textarea>
+ </div>
+ <div class="field box">
+ <label for="DeepInfra-api_key" class="label" title="">DeepInfra:</label>
+ <textarea id="DeepInfra-api_key" name="DeepInfra[api_key]" class="DeepInfraImage-api_key" placeholder="api_key"></textarea>
+ </div>
+ <div class="field box">
+ <label for="GeminiPro-api_key" class="label" title="">Gemini API:</label>
+ <textarea id="GeminiPro-api_key" name="GeminiPro[api_key]" placeholder="api_key"></textarea>
+ </div>
+ <div class="field box">
+ <label for="Groq-api_key" class="label" title="">Groq:</label>
+ <textarea id="Groq-api_key" name="Groq[api_key]" placeholder="api_key"></textarea>
+ </div>
+ <div class="field box">
+ <label for="HuggingFace-api_key" class="label" title="">HuggingFace:</label>
+ <textarea id="HuggingFace-api_key" name="HuggingFace[api_key]" placeholder="api_key"></textarea>
+ </div>
+ <div class="field box">
+ <label for="Openai-api_key" class="label" title="">OpenAI API:</label>
+ <textarea id="Openai-api_key" name="Openai[api_key]" placeholder="api_key"></textarea>
+ </div>
+ <div class="field box">
+ <label for="OpenRouter-api_key" class="label" title="">OpenRouter:</label>
+ <textarea id="OpenRouter-api_key" name="OpenRouter[api_key]" placeholder="api_key"></textarea>
+ </div>
+ <div class="field box">
+ <label for="PerplexityApi-api_key" class="label" title="">Perplexity API:</label>
+ <textarea id="PerplexityApi-api_key" name="PerplexityApi[api_key]" placeholder="api_key"></textarea>
+ </div>
+ <div class="field box">
+ <label for="Replicate-api_key" class="label" title="">Replicate:</label>
+ <textarea id="Replicate-api_key" name="Replicate[api_key]" class="ReplicateImage-api_key" placeholder="api_key"></textarea>
+ </div>
</div>
<div class="bottom_buttons">
- <button onclick="delete_conversations()">
- <i class="fa-regular fa-trash"></i>
- <span>Clear Conversations</span>
- </button>
- <button onclick="save_storage()">
- <i class="fa-solid fa-download"></i>
- <a href="" onclick="return false;">Export Conversations</a>
- </button>
+ <button onclick="delete_conversations()">
+ <i class="fa-regular fa-trash"></i>
+ <span>Clear Conversations</span>
+ </button>
+ <button onclick="save_storage()">
+ <i class="fa-solid fa-download"></i>
+ <a href="" onclick="return false;">Export Conversations</a>
+ </button>
</div>
- </div>
- <div class="conversation">
+ </div>
+ <div class="conversation">
<textarea id="systemPrompt" class="box" placeholder="System prompt"></textarea>
<div id="messages" class="box"></div>
<button class="slide-systemPrompt">
- <i class="fa-solid fa-angles-up"></i>
+ <i class="fa-solid fa-angles-up"></i>
</button>
<div class="toolbar">
- <div id="input-count" class="">
- <button class="hide-input">
- <i class="fa-solid fa-angles-down"></i>
- </button>
- <span class="text"></span>
- </div>
- <div class="stop_generating stop_generating-hidden">
- <button id="cancelButton">
- <span>Stop Generating</span>
- <i class="fa-regular fa-stop"></i>
- </button>
- </div>
- <div class="regenerate regenerate-hidden">
- <button id="regenerateButton">
- <span>Regenerate</span>
- <i class="fa-solid fa-rotate"></i>
- </button>
- </div>
+ <div id="input-count" class="">
+ <button class="hide-input">
+ <i class="fa-solid fa-angles-down"></i>
+ </button>
+ <span class="text"></span>
+ </div>
+ <div class="stop_generating stop_generating-hidden">
+ <button id="cancelButton">
+ <span>Stop Generating</span>
+ <i class="fa-regular fa-stop"></i>
+ </button>
+ </div>
+ <div class="regenerate regenerate-hidden">
+ <button id="regenerateButton">
+ <span>Regenerate</span>
+ <i class="fa-solid fa-rotate"></i>
+ </button>
+ </div>
</div>
<div class="user-input">
- <div class="box input-box">
- <textarea id="message-input" placeholder="Ask a question" cols="30" rows="10"
- style="white-space: pre-wrap;resize: none;"></textarea>
- <label class="file-label image-label" for="image" title="Works with Bing, Gemini, OpenaiChat and You">
- <input type="file" id="image" name="image" accept="image/*" required/>
- <i class="fa-regular fa-image"></i>
- </label>
- <label class="file-label image-label" for="camera">
- <input type="file" id="camera" name="camera" accept="image/*" capture="camera" required/>
- <i class="fa-solid fa-camera"></i>
- </label>
- <label class="file-label" for="file">
- <input type="file" id="file" name="file" accept="text/plain, text/html, text/xml, application/json, text/javascript, .sh, .py, .php, .css, .yaml, .sql, .log, .csv, .twig, .md" required/>
- <i class="fa-solid fa-paperclip"></i>
- </label>
- <label class="micro-label" for="micro">
- <i class="fa-solid fa-microphone-slash"></i>
- </label>
- <div id="send-button">
- <i class="fa-solid fa-paper-plane-top"></i>
- </div>
- </div>
+ <div class="box input-box">
+ <textarea id="message-input" placeholder="Ask a question" cols="30" rows="10"
+ style="white-space: pre-wrap;resize: none;"></textarea>
+ <label class="file-label image-label" for="image" title="Works with Bing, Gemini, OpenaiChat and You">
+ <input type="file" id="image" name="image" accept="image/*" required/>
+ <i class="fa-regular fa-image"></i>
+ </label>
+ <label class="file-label image-label" for="camera">
+ <input type="file" id="camera" name="camera" accept="image/*" capture="camera" required/>
+ <i class="fa-solid fa-camera"></i>
+ </label>
+ <label class="file-label" for="file">
+ <input type="file" id="file" name="file" accept="text/plain, text/html, text/xml, application/json, text/javascript, .sh, .py, .php, .css, .yaml, .sql, .log, .csv, .twig, .md" required/>
+ <i class="fa-solid fa-paperclip"></i>
+ </label>
+ <label class="micro-label" for="micro">
+ <i class="fa-solid fa-microphone-slash"></i>
+ </label>
+ <div id="send-button">
+ <i class="fa-solid fa-paper-plane-top"></i>
+ </div>
+ </div>
</div>
<div class="buttons">
- <div class="field">
- <select name="model" id="model">
- <option value="">Model: Default</option>
- <option value="gpt-4">gpt-4</option>
- <option value="gpt-3.5-turbo">gpt-3.5-turbo</option>
- <option value="llama-3-70b-chat">llama-3-70b-chat</option>
- <option value="llama-3.1-70b">llama-3.1-70b</option>
- <option value="gemini-pro">gemini-pro</option>
- <option value="">----</option>
- </select>
- <select name="model2" id="model2" class="hidden"></select>
- </div>
- <div class="field">
- <select name="provider" id="provider">
- <option value="">Provider: Auto</option>
- <option value="Bing">Bing</option>
- <option value="OpenaiChat">OpenAI ChatGPT</option>
- <option value="Gemini">Gemini</option>
- <option value="Liaobots">Liaobots</option>
- <option value="MetaAI">Meta AI</option>
- <option value="You">You</option>
- <option value="">----</option>
- </select>
- </div>
+ <div class="field">
+ <select name="model" id="model">
+ <option value="">Model: Default</option>
+ <option value="gpt-4">gpt-4</option>
+ <option value="gpt-4o">gpt-4o</option>
+ <option value="gpt-4o-mini">gpt-4o-mini</option>
+ <option value="llama-3.1-70b">llama-3.1-70b</option>
+ <option value="llama-3.1-70b">llama-3.1-405b</option>
+ <option value="llama-3.1-70b">mixtral-8x7b</option>
+ <option value="gemini-pro">gemini-pro</option>
+ <option value="gemini-flash">gemini-flash</option>
+ <option value="claude-3-haiku">claude-3-haiku</option>
+ <option value="claude-3.5-sonnet">claude-3.5-sonnet</option>
+ <option value="">----</option>
+ </select>
+ <select name="model2" id="model2" class="hidden"></select>
+ </div>
+ <div class="field">
+ <select name="provider" id="provider">
+ <option value="">Provider: Auto</option>
+ <option value="OpenaiChat">OpenAI ChatGPT</option>
+ <option value="Gemini">Gemini</option>
+ <option value="MetaAI">Meta AI</option>
+ <option value="DeepInfraChat">DeepInfraChat</option>
+ <option value="ReplicateHome">ReplicateHome</option>
+ <option value="Blackbox">Blackbox</option>
+ <option value="HuggingChat">HuggingChat</option>
+ <option value="DDG">DDG</option>
+ <option value="Pizzagpt">Pizzagpt</option>
+ <option value="">----</option>
+ </select>
+ </div>
</div>
- </div>
- </div>
- <div class="mobile-sidebar">
- <i class="fa-solid fa-bars"></i>
- </div>
-</body>
+ </div>
+ </div>
+ <div id="imageModal" class="modal hidden">
+ <span class="close">&times;</span>
+ <img class="modal-content" id="img01">
+ <div id="caption"></div>
+ <div class="navigation-controls">
+ <button id="prevImage" class="nav-button left">&#10094;</button>
+ <span id="imageCounter" class="image-counter"></span>
+ <button id="nextImage" class="nav-button right">&#10095;</button>
+ </div>
+ </div>
+ <div class="mobile-sidebar">
+ <i class="fa-solid fa-bars"></i>
+ </div>
+ </body>
</html>
+
diff --git a/g4f/gui/client/static/css/style.css b/g4f/gui/client/static/css/style.css
index 441e2042..72f3ec4f 100644
--- a/g4f/gui/client/static/css/style.css
+++ b/g4f/gui/client/static/css/style.css
@@ -474,7 +474,6 @@ body {
.stop_generating, .toolbar .regenerate {
position: absolute;
- z-index: 1000000;
top: 0;
right: 0;
}
@@ -1118,6 +1117,92 @@ a:-webkit-any-link {
display: none;
}
+.album-image {
+ width: 100px;
+ height: auto;
+ margin: 5px;
+ display: inline-block;
+}
+
+.modal {
+ display: none;
+ position: fixed;
+ z-index: 1;
+ left: 0;
+ top: 0;
+ width: 100%;
+ height: 100%;
+ overflow: hidden;
+ background-color: rgba(0,0,0,0.9);
+}
+
+.modal-content {
+ margin: auto;
+ display: block;
+ max-width: 80%;
+ max-height: 80%;
+ transition: transform 0.2s;
+}
+
+.close {
+ position: absolute;
+ top: 15px;
+ right: 35px;
+ color: #f1f1f1;
+ font-size: 40px;
+ font-weight: bold;
+ transition: 0.3s;
+}
+
+.close:hover,
+.close:focus {
+ color: #bbb;
+ text-decoration: none;
+ cursor: pointer;
+}
+
+
+.image-counter {
+ color: #fff;
+ font-size: 18px;
+ margin: auto 10px;
+ user-select: none;
+}
+
+.nav-button {
+ background-color: #555;
+ color: #fff;
+ border: none;
+ padding: 10px;
+ font-size: 20px;
+ cursor: pointer;
+}
+
+.nav-button:hover {
+ background-color: #777;
+}
+
+.nav-button {
+ position: relative;
+}
+
+.nav-button.left {
+ left: 0;
+}
+
+.nav-button.right {
+ right: 0;
+}
+
+.navigation-controls {
+ position: absolute;
+ bottom: 20px;
+ left: 50%;
+ transform: translateX(-50%);
+ display: flex;
+ gap: 10px;
+}
+
.blink {
animation: blinker 1s step-start infinite;
}
diff --git a/g4f/gui/client/static/js/chat.v1.js b/g4f/gui/client/static/js/chat.v1.js
index d5a1ffaa..9bf07046 100644
--- a/g4f/gui/client/static/js/chat.v1.js
+++ b/g4f/gui/client/static/js/chat.v1.js
@@ -338,6 +338,14 @@ const prepare_messages = (messages, message_index = -1) => {
messages = messages.filter((_, index) => message_index >= index);
}
+ let new_messages = [];
+ if (systemPrompt?.value) {
+ new_messages.push({
+ "role": "system",
+ "content": systemPrompt.value
+ });
+ }
+
// Remove history, if it's selected
if (document.getElementById('history')?.checked) {
if (message_index == null) {
@@ -347,13 +355,6 @@ const prepare_messages = (messages, message_index = -1) => {
}
}
- let new_messages = [];
- if (systemPrompt?.value) {
- new_messages.push({
- "role": "system",
- "content": systemPrompt.value
- });
- }
messages.forEach((new_message) => {
// Include only not regenerated messages
if (new_message && !new_message.regenerate) {
@@ -366,6 +367,7 @@ const prepare_messages = (messages, message_index = -1) => {
return new_messages;
}
+
async function add_message_chunk(message) {
if (message.type == "conversation") {
console.info("Conversation used:", message.conversation)
@@ -934,17 +936,127 @@ function open_settings() {
}
}
+async function loadImages() {
+ try {
+ const response = await fetch('/images');
+ const images = await response.json();
+ console.log(images);
+ displayImages(images);
+ } catch (error) {
+ console.error('Error fetching images:', error);
+ }
+}
+
+function displayImages(images) {
+ const album = document.querySelector('.images');
+ album.innerHTML = '';
+ images.forEach(image => {
+ const imgElement = document.createElement('img');
+ imgElement.src = image;
+ imgElement.alt = 'Generated Image';
+ imgElement.classList.add('album-image');
+ album.appendChild(imgElement);
+ });
+}
+
+document.addEventListener('DOMContentLoaded', () => {
+ loadImages();
+});
+
function open_album() {
+ const album = document.querySelector('.images');
if (album.classList.contains("hidden")) {
sidebar.classList.remove("shown");
settings.classList.add("hidden");
album.classList.remove("hidden");
history.pushState({}, null, "/images/");
+ loadImages();
} else {
album.classList.add("hidden");
}
}
+let currentScale = 1;
+let currentImageIndex = 0;
+let imagesList = [];
+
+function displayImages(images) {
+ imagesList = images;
+ const album = document.querySelector('.images');
+ album.innerHTML = '';
+ images.forEach((image, index) => {
+ const imgElement = document.createElement('img');
+ imgElement.src = image;
+ imgElement.alt = 'Generated Image';
+ imgElement.classList.add('album-image');
+ imgElement.style.cursor = 'pointer';
+ imgElement.addEventListener('click', () => openImageModal(index));
+ album.appendChild(imgElement);
+ });
+}
+
+function openImageModal(index) {
+ currentImageIndex = index;
+ const modal = document.getElementById('imageModal');
+ const modalImg = document.getElementById('img01');
+ const imageCounter = document.getElementById('imageCounter');
+ modal.style.display = 'block';
+ modalImg.src = imagesList[index];
+ currentScale = 1;
+ modalImg.style.transform = `scale(${currentScale})`;
+ imageCounter.textContent = `${index + 1} / ${imagesList.length}`;
+}
+
+const modal = document.getElementById('imageModal');
+const span = document.getElementsByClassName('close')[0];
+const prevImageButton = document.getElementById('prevImage');
+const nextImageButton = document.getElementById('nextImage');
+
+span.onclick = function() {
+ modal.style.display = 'none';
+}
+
+window.onclick = function(event) {
+ if (event.target == modal) {
+ modal.style.display = 'none';
+ }
+}
+
+document.getElementById('img01').addEventListener('wheel', function(event) {
+ event.preventDefault();
+ if (event.deltaY < 0) {
+ currentScale += 0.1;
+ } else if (currentScale > 0.1) {
+ currentScale -= 0.1;
+ }
+ document.getElementById('img01').style.transform = `scale(${currentScale})`;
+});
+
+prevImageButton.onclick = function() {
+ if (currentImageIndex > 0) {
+ currentImageIndex--;
+ openImageModal(currentImageIndex);
+ }
+}
+
+nextImageButton.onclick = function() {
+ if (currentImageIndex < imagesList.length - 1) {
+ currentImageIndex++;
+ openImageModal(currentImageIndex);
+ }
+}
+
+document.addEventListener('keydown', function(event) {
+ if (modal.style.display === 'block') {
+ if (event.key === 'ArrowLeft') {
+ prevImageButton.click();
+ } else if (event.key === 'ArrowRight') {
+ nextImageButton.click();
+ }
+ }
+});
+
+
const register_settings_storage = async () => {
optionElements.forEach((element) => {
if (element.type == "textarea") {
diff --git a/g4f/gui/server/api.py b/g4f/gui/server/api.py
index 57f3eaa1..51cf3d32 100644
--- a/g4f/gui/server/api.py
+++ b/g4f/gui/server/api.py
@@ -2,12 +2,11 @@ from __future__ import annotations
import logging
import os
-import os.path
import uuid
import asyncio
import time
from aiohttp import ClientSession
-from typing import Iterator, Optional
+from typing import Iterator, Optional, AsyncIterator, Union
from flask import send_from_directory
from g4f import version, models
@@ -20,21 +19,20 @@ from g4f.Provider import ProviderType, __providers__, __map__
from g4f.providers.base_provider import ProviderModelMixin, FinishReason
from g4f.providers.conversation import BaseConversation
-conversations: dict[dict[str, BaseConversation]] = {}
+# Define the directory for generated images
images_dir = "./generated_images"
+# Function to ensure the images directory exists
+def ensure_images_dir():
+ if not os.path.exists(images_dir):
+ os.makedirs(images_dir)
+
+conversations: dict[dict[str, BaseConversation]] = {}
+
class Api:
@staticmethod
def get_models() -> list[str]:
- """
- Return a list of all models.
-
- Fetches and returns a list of all available models in the system.
-
- Returns:
- List[str]: A list of model names.
- """
return models._all_models
@staticmethod
@@ -82,9 +80,6 @@ class Api:
@staticmethod
def get_providers() -> list[str]:
- """
- Return a list of all working providers.
- """
return {
provider.__name__: (
provider.label if hasattr(provider, "label") else provider.__name__
@@ -99,12 +94,6 @@ class Api:
@staticmethod
def get_version():
- """
- Returns the current and latest version of the application.
-
- Returns:
- dict: A dictionary containing the current and latest version.
- """
try:
current_version = version.utils.current_version
except VersionNotFoundError:
@@ -115,18 +104,10 @@ class Api:
}
def serve_images(self, name):
+ ensure_images_dir()
return send_from_directory(os.path.abspath(images_dir), name)
def _prepare_conversation_kwargs(self, json_data: dict, kwargs: dict):
- """
- Prepares arguments for chat completion based on the request data.
-
- Reads the request and prepares the necessary arguments for handling
- a chat completion request.
-
- Returns:
- dict: Arguments prepared for chat completion.
- """
model = json_data.get('model') or models.default
provider = json_data.get('provider')
messages = json_data['messages']
@@ -159,13 +140,11 @@ class Api:
result = ChatCompletion.create(**kwargs)
first = True
if isinstance(result, ImageResponse):
- # Якщо результат є ImageResponse, обробляємо його як одиночний елемент
if first:
first = False
yield self._format_json("provider", get_last_provider(True))
yield self._format_json("content", str(result))
else:
- # Якщо результат є ітерабельним, обробляємо його як раніше
for chunk in result:
if first:
first = False
@@ -181,7 +160,6 @@ class Api:
elif isinstance(chunk, ImagePreview):
yield self._format_json("preview", chunk.to_string())
elif isinstance(chunk, ImageResponse):
- # Обробка ImageResponse
images = asyncio.run(self._copy_images(chunk.get_list(), chunk.options.get("cookies")))
yield self._format_json("content", str(ImageResponse(images, chunk.alt)))
elif not isinstance(chunk, FinishReason):
@@ -190,8 +168,8 @@ class Api:
logging.exception(e)
yield self._format_json('error', get_error_message(e))
- # Додайте цей метод до класу Api
async def _copy_images(self, images: list[str], cookies: Optional[Cookies] = None):
+ ensure_images_dir()
async with ClientSession(
connector=get_connector(None, os.environ.get("G4F_PROXY")),
cookies=cookies
@@ -212,16 +190,6 @@ class Api:
return await asyncio.gather(*[copy_image(image) for image in images])
def _format_json(self, response_type: str, content):
- """
- Formats and returns a JSON response.
-
- Args:
- response_type (str): The type of the response.
- content: The content to be included in the response.
-
- Returns:
- str: A JSON formatted string.
- """
return {
'type': response_type,
response_type: content
@@ -229,15 +197,6 @@ class Api:
def get_error_message(exception: Exception) -> str:
- """
- Generates a formatted error message from an exception.
-
- Args:
- exception (Exception): The exception to format.
-
- Returns:
- str: A formatted error message string.
- """
message = f"{type(exception).__name__}: {exception}"
provider = get_last_provider()
if provider is None:
diff --git a/g4f/gui/server/backend.py b/g4f/gui/server/backend.py
index dc1b1080..e24d4da2 100644
--- a/g4f/gui/server/backend.py
+++ b/g4f/gui/server/backend.py
@@ -1,5 +1,6 @@
import json
-from flask import request, Flask
+import os
+from flask import request, Flask, jsonify, send_from_directory
from g4f.image import is_allowed_extension, to_image
from .api import Api
@@ -54,6 +55,10 @@ class Backend_Api(Api):
'/images/<path:name>': {
'function': self.serve_images,
'methods': ['GET']
+ },
+ '/images': {
+ 'function': self.get_images,
+ 'methods': ['GET']
}
}
@@ -110,4 +115,19 @@ class Backend_Api(Api):
Returns:
str: A JSON formatted string.
"""
- return json.dumps(super()._format_json(response_type, content)) + "\n" \ No newline at end of file
+ return json.dumps(super()._format_json(response_type, content)) + "\n"
+
+ @staticmethod
+ def get_images():
+ images_dir = "./generated_images"
+ try:
+ images = [f for f in os.listdir(images_dir) if os.path.isfile(os.path.join(images_dir, f))]
+ images = [f"/images/{image}" for image in images if image.lower().endswith(('.png', '.jpg', '.jpeg', '.gif', '.webp'))]
+ return jsonify(images)
+ except Exception as e:
+ return str(e), 500
+
+ @staticmethod
+ def serve_images(name):
+ images_dir = "./generated_images"
+ return send_from_directory(os.path.abspath(images_dir), name)
diff --git a/g4f/gui/server/website.py b/g4f/gui/server/website.py
index 5e633674..3cabcdf3 100644
--- a/g4f/gui/server/website.py
+++ b/g4f/gui/server/website.py
@@ -27,6 +27,10 @@ class Website:
'function': redirect_home,
'methods': ['GET', 'POST']
},
+ '/images/': {
+ 'function': redirect_home,
+ 'methods': ['GET', 'POST']
+ },
}
def _chat(self, conversation_id):
@@ -35,4 +39,4 @@ class Website:
return render_template('index.html', chat_id=conversation_id)
def _index(self):
- return render_template('index.html', chat_id=str(uuid.uuid4())) \ No newline at end of file
+ return render_template('index.html', chat_id=str(uuid.uuid4()))
diff --git a/g4f/models.py b/g4f/models.py
index 1cea6447..32a12d10 100644
--- a/g4f/models.py
+++ b/g4f/models.py
@@ -23,7 +23,6 @@ from .Provider import (
DDG,
DeepInfra,
DeepInfraChat,
- DeepInfraImage,
Editee,
Free2GPT,
FreeChatgpt,
@@ -31,6 +30,7 @@ from .Provider import (
FreeNetfly,
Gemini,
GeminiPro,
+ GizAI,
GigaChat,
GPROChat,
HuggingChat,
@@ -42,9 +42,6 @@ from .Provider import (
NexraBing,
NexraBlackbox,
NexraChatGPT,
- NexraChatGPT4o,
- NexraChatGptV2,
- NexraChatGptWeb,
NexraDallE,
NexraDallE2,
NexraEmi,
@@ -87,6 +84,8 @@ class Model:
"""Returns a list of all model names."""
return _all_models
+
+### Default ###
default = Model(
name = "",
base_provider = "",
@@ -113,6 +112,8 @@ default = Model(
])
)
+
+
############
### Text ###
############
@@ -136,13 +137,13 @@ gpt_35_turbo = Model(
gpt_4o = Model(
name = 'gpt-4o',
base_provider = 'OpenAI',
- best_provider = IterListProvider([NexraChatGPT4o, Blackbox, ChatGptEs, AmigoChat, DarkAI, Editee, Liaobots, Airforce, OpenaiChat])
+ best_provider = IterListProvider([NexraChatGPT, Blackbox, ChatGptEs, AmigoChat, DarkAI, Editee, GizAI, Airforce, Liaobots, OpenaiChat])
)
gpt_4o_mini = Model(
name = 'gpt-4o-mini',
base_provider = 'OpenAI',
- best_provider = IterListProvider([DDG, ChatGptEs, FreeNetfly, Pizzagpt, MagickPen, AmigoChat, RubiksAI, Liaobots, Airforce, ChatgptFree, Koala, OpenaiChat, ChatGpt])
+ best_provider = IterListProvider([DDG, ChatGptEs, FreeNetfly, Pizzagpt, MagickPen, AmigoChat, RubiksAI, Liaobots, Airforce, GizAI, ChatgptFree, Koala, OpenaiChat, ChatGpt])
)
gpt_4_turbo = Model(
@@ -154,7 +155,7 @@ gpt_4_turbo = Model(
gpt_4 = Model(
name = 'gpt-4',
base_provider = 'OpenAI',
- best_provider = IterListProvider([Chatgpt4Online, Ai4Chat, NexraBing, NexraChatGPT, NexraChatGptV2, NexraChatGptWeb, Airforce, Bing, OpenaiChat, gpt_4_turbo.best_provider, gpt_4o.best_provider, gpt_4o_mini.best_provider])
+ best_provider = IterListProvider([Chatgpt4Online, Ai4Chat, NexraBing, NexraChatGPT, Airforce, Bing, OpenaiChat, gpt_4_turbo.best_provider, gpt_4o.best_provider, gpt_4o_mini.best_provider])
)
# o1
@@ -167,7 +168,7 @@ o1 = Model(
o1_mini = Model(
name = 'o1-mini',
base_provider = 'OpenAI',
- best_provider = AmigoChat
+ best_provider = IterListProvider([AmigoChat, GizAI])
)
@@ -216,13 +217,13 @@ llama_3_70b = Model(
llama_3_1_8b = Model(
name = "llama-3.1-8b",
base_provider = "Meta Llama",
- best_provider = IterListProvider([Blackbox, DeepInfraChat, ChatHub, Cloudflare, Airforce, PerplexityLabs])
+ best_provider = IterListProvider([Blackbox, DeepInfraChat, ChatHub, Cloudflare, Airforce, GizAI, PerplexityLabs])
)
llama_3_1_70b = Model(
name = "llama-3.1-70b",
base_provider = "Meta Llama",
- best_provider = IterListProvider([DDG, HuggingChat, Blackbox, FreeGpt, TeachAnything, Free2GPT, DeepInfraChat, DarkAI, Airforce, AiMathGPT, RubiksAI, HuggingFace, PerplexityLabs])
+ best_provider = IterListProvider([DDG, HuggingChat, Blackbox, FreeGpt, TeachAnything, Free2GPT, DeepInfraChat, DarkAI, Airforce, AiMathGPT, RubiksAI, GizAI, HuggingFace, PerplexityLabs])
)
llama_3_1_405b = Model(
@@ -299,7 +300,7 @@ mistral_nemo = Model(
mistral_large = Model(
name = "mistral-large",
base_provider = "Mistral",
- best_provider = Editee
+ best_provider = IterListProvider([Editee, GizAI])
)
@@ -347,13 +348,13 @@ phi_3_5_mini = Model(
gemini_pro = Model(
name = 'gemini-pro',
base_provider = 'Google DeepMind',
- best_provider = IterListProvider([GeminiPro, Blackbox, AIChatFree, GPROChat, NexraGeminiPro, AmigoChat, Editee, Liaobots, Airforce])
+ best_provider = IterListProvider([GeminiPro, Blackbox, AIChatFree, GPROChat, NexraGeminiPro, AmigoChat, Editee, GizAI, Airforce, Liaobots])
)
gemini_flash = Model(
name = 'gemini-flash',
base_provider = 'Google DeepMind',
- best_provider = IterListProvider([Blackbox, Liaobots, Airforce])
+ best_provider = IterListProvider([Blackbox, GizAI, Airforce, Liaobots])
)
gemini = Model(
@@ -424,14 +425,14 @@ claude_3_sonnet = Model(
claude_3_haiku = Model(
name = 'claude-3-haiku',
base_provider = 'Anthropic',
- best_provider = IterListProvider([DDG, Airforce, Liaobots])
+ best_provider = IterListProvider([DDG, Airforce, GizAI, Liaobots])
)
# claude 3.5
claude_3_5_sonnet = Model(
name = 'claude-3.5-sonnet',
base_provider = 'Anthropic',
- best_provider = IterListProvider([Blackbox, Editee, AmigoChat, Airforce, Liaobots])
+ best_provider = IterListProvider([Blackbox, Editee, AmigoChat, Airforce, GizAI, Liaobots])
)
@@ -753,14 +754,14 @@ sdxl_lora = Model(
sdxl = Model(
name = 'sdxl',
base_provider = 'Stability AI',
- best_provider = IterListProvider([ReplicateHome, DeepInfraImage])
+ best_provider = IterListProvider([ReplicateHome])
)
sd_1_5 = Model(
name = 'sd-1.5',
base_provider = 'Stability AI',
- best_provider = NexraSD15
+ best_provider = IterListProvider([NexraSD15, GizAI])
)
@@ -771,6 +772,13 @@ sd_3 = Model(
)
+sd_3_5 = Model(
+ name = 'sd-3.5',
+ base_provider = 'Stability AI',
+ best_provider = GizAI
+
+)
+
### Playground ###
playground_v2_5 = Model(
name = 'playground-v2.5',
@@ -791,7 +799,7 @@ flux = Model(
flux_pro = Model(
name = 'flux-pro',
base_provider = 'Flux AI',
- best_provider = IterListProvider([AmigoChat, NexraFluxPro])
+ best_provider = IterListProvider([NexraFluxPro, AmigoChat])
)
@@ -840,7 +848,7 @@ flux_4o = Model(
flux_schnell = Model(
name = 'flux-schnell',
base_provider = 'Flux AI',
- best_provider = ReplicateHome
+ best_provider = IterListProvider([ReplicateHome, GizAI])
)
@@ -1123,6 +1131,7 @@ class ModelUtils:
'sdxl-turbo': sdxl_turbo,
'sd-1.5': sd_1_5,
'sd-3': sd_3,
+'sd-3.5': sd_3_5,
### Playground ###