diff options
Diffstat (limited to 'g4f/Provider')
-rw-r--r-- | g4f/Provider/AItianhuSpace.py | 73 | ||||
-rw-r--r-- | g4f/Provider/Aivvm.py | 5 | ||||
-rw-r--r-- | g4f/Provider/Vercel.py | 30 | ||||
-rw-r--r-- | g4f/Provider/__init__.py | 2 | ||||
-rw-r--r-- | g4f/Provider/base_provider.py | 31 |
5 files changed, 124 insertions, 17 deletions
diff --git a/g4f/Provider/AItianhuSpace.py b/g4f/Provider/AItianhuSpace.py new file mode 100644 index 00000000..8beb3355 --- /dev/null +++ b/g4f/Provider/AItianhuSpace.py @@ -0,0 +1,73 @@ +from __future__ import annotations + +import random, json + +from g4f.requests import AsyncSession, StreamRequest +from .base_provider import AsyncGeneratorProvider, format_prompt + +domains = { + "gpt-3.5-turbo": ".aitianhu.space", + "gpt-4": ".aitianhu.website", +} + +class AItianhuSpace(AsyncGeneratorProvider): + url = "https://chat3.aiyunos.top/" + working = True + supports_gpt_35_turbo = True + + @classmethod + async def create_async_generator( + cls, + model: str, + messages: list[dict[str, str]], + stream: bool = True, + **kwargs + ) -> str: + if not model: + model = "gpt-3.5-turbo" + elif not model in domains: + raise ValueError(f"Model are not supported: {model}") + + chars = 'abcdefghijklmnopqrstuvwxyz0123456789' + rand = ''.join(random.choice(chars) for _ in range(6)) + domain = domains[model] + url = f'https://{rand}{domain}/api/chat-process' + + headers = { + "User-Agent": "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/113.0.0.0 Safari/537.36", + } + async with AsyncSession(headers=headers, impersonate="chrome107", verify=False) as session: + data = { + "prompt": format_prompt(messages), + "options": {}, + "systemMessage": "You are ChatGPT, a large language model trained by OpenAI. Follow the user's instructions carefully.", + "temperature": 0.8, + "top_p": 1, + **kwargs + } + async with StreamRequest(session, "POST", url, json=data) as response: + response.raise_for_status() + async for line in response.content: + line = json.loads(line.rstrip()) + if "detail" in line: + content = line["detail"]["choices"][0]["delta"].get("content") + if content: + yield content + elif "message" in line and "AI-4接口非常昂贵" in line["message"]: + raise RuntimeError("Rate limit for GPT 4 reached") + else: + raise RuntimeError("Response: {line}") + + + @classmethod + @property + def params(cls): + params = [ + ("model", "str"), + ("messages", "list[dict[str, str]]"), + ("stream", "bool"), + ("temperature", "float"), + ("top_p", "int"), + ] + param = ", ".join([": ".join(p) for p in params]) + return f"g4f.provider.{cls.__name__} supports: ({param})" diff --git a/g4f/Provider/Aivvm.py b/g4f/Provider/Aivvm.py index c38c4a74..7a3d57bd 100644 --- a/g4f/Provider/Aivvm.py +++ b/g4f/Provider/Aivvm.py @@ -60,9 +60,10 @@ class Aivvm(BaseProvider): response = requests.post( "https://chat.aivvm.com/api/chat", headers=headers, json=json_data, stream=True) + response.raise_for_status() - for line in response.iter_content(chunk_size=1048): - yield line.decode('utf-8') + for chunk in response.iter_content(chunk_size=None): + yield chunk.decode('utf-8') @classmethod @property diff --git a/g4f/Provider/Vercel.py b/g4f/Provider/Vercel.py index ca124fec..4102c07b 100644 --- a/g4f/Provider/Vercel.py +++ b/g4f/Provider/Vercel.py @@ -18,7 +18,13 @@ class Vercel(BaseProvider): def create_completion( model: str, messages: list[dict[str, str]], - stream: bool, **kwargs ) -> CreateResult: + stream: bool, + **kwargs + ) -> CreateResult: + if not model: + model = "gpt-3.5-turbo" + elif model not in model_info: + raise ValueError(f"Model are not supported: {model}") headers = { 'authority' : 'sdk.vercel.ai', @@ -26,7 +32,7 @@ class Vercel(BaseProvider): 'accept-language' : 'en,fr-FR;q=0.9,fr;q=0.8,es-ES;q=0.7,es;q=0.6,en-US;q=0.5,am;q=0.4,de;q=0.3', 'cache-control' : 'no-cache', 'content-type' : 'application/json', - 'custom-encoding' : AntiBotToken(), + 'custom-encoding' : get_anti_bot_token(), 'origin' : 'https://sdk.vercel.ai', 'pragma' : 'no-cache', 'referer' : 'https://sdk.vercel.ai/', @@ -48,22 +54,20 @@ class Vercel(BaseProvider): 'playgroundId': str(uuid.uuid4()), 'chatIndex' : 0} | model_info[model]['default_params'] - server_error = True - retries = 0 max_retries = kwargs.get('max_retries', 20) - - while server_error and not retries > max_retries: + for i in range(max_retries): response = requests.post('https://sdk.vercel.ai/api/generate', headers=headers, json=json_data, stream=True) + try: + response.raise_for_status() + except: + continue + for token in response.iter_content(chunk_size=8): + yield token.decode() + break - for token in response.iter_content(chunk_size=2046): - if token != b'Internal Server Error': - server_error = False - yield (token.decode()) - - retries += 1 -def AntiBotToken() -> str: +def get_anti_bot_token() -> str: headers = { 'authority' : 'sdk.vercel.ai', 'accept' : '*/*', diff --git a/g4f/Provider/__init__.py b/g4f/Provider/__init__.py index aa19ade3..ebe01603 100644 --- a/g4f/Provider/__init__.py +++ b/g4f/Provider/__init__.py @@ -4,6 +4,7 @@ from .Aichat import Aichat from .Ails import Ails from .AiService import AiService from .AItianhu import AItianhu +from .AItianhuSpace import AItianhuSpace from .Aivvm import Aivvm from .Bard import Bard from .Bing import Bing @@ -52,6 +53,7 @@ __all__ = [ 'Ails', 'AiService', 'AItianhu', + 'AItianhuSpace', 'Aivvm', 'Bard', 'Bing', diff --git a/g4f/Provider/base_provider.py b/g4f/Provider/base_provider.py index ea81502f..e8a54f78 100644 --- a/g4f/Provider/base_provider.py +++ b/g4f/Provider/base_provider.py @@ -1,7 +1,9 @@ from __future__ import annotations import asyncio -from asyncio import SelectorEventLoop +import functools +from asyncio import SelectorEventLoop, AbstractEventLoop +from concurrent.futures import ThreadPoolExecutor from abc import ABC, abstractmethod import browser_cookie3 @@ -27,6 +29,31 @@ class BaseProvider(ABC): ) -> CreateResult: raise NotImplementedError() + @classmethod + async def create_async( + cls, + model: str, + messages: list[dict[str, str]], + *, + loop: AbstractEventLoop = None, + executor: ThreadPoolExecutor = None, + **kwargs + ) -> str: + if not loop: + loop = asyncio.get_event_loop() + + partial_func = functools.partial( + cls.create_completion, + model, + messages, + False, + **kwargs + ) + response = await loop.run_in_executor( + executor, + partial_func + ) + return "".join(response) @classmethod @property @@ -127,7 +154,7 @@ def create_event_loop() -> SelectorEventLoop: except RuntimeError: return SelectorEventLoop() raise RuntimeError( - 'Use "create_async" instead of "create" function in a async loop.') + 'Use "create_async" instead of "create" function in a running event loop.') _cookies = {} |