From aa265acf30bec0ea7a3f411e46da49dd05b383d0 Mon Sep 17 00:00:00 2001 From: kqlio67 Date: Sun, 1 Sep 2024 19:39:55 +0300 Subject: Update models and providers per community feedback --- g4f/Provider/Cohere.py | 106 ------------------------------ g4f/Provider/GeminiProChat.py | 75 --------------------- g4f/Provider/Marsyoo.py | 64 ------------------ g4f/Provider/TeachAnything.py | 62 ----------------- g4f/Provider/not_working/AItianhu.py | 79 ---------------------- g4f/Provider/not_working/Aichatos.py | 56 ---------------- g4f/Provider/not_working/Bestim.py | 56 ---------------- g4f/Provider/not_working/ChatBase.py | 61 ----------------- g4f/Provider/not_working/ChatForAi.py | 66 ------------------- g4f/Provider/not_working/ChatgptAi.py | 88 ------------------------- g4f/Provider/not_working/ChatgptDemo.py | 70 -------------------- g4f/Provider/not_working/ChatgptDemoAi.py | 56 ---------------- g4f/Provider/not_working/ChatgptLogin.py | 78 ---------------------- g4f/Provider/not_working/ChatgptNext.py | 66 ------------------- g4f/Provider/not_working/ChatgptX.py | 106 ------------------------------ g4f/Provider/not_working/Chatxyz.py | 60 ----------------- g4f/Provider/not_working/Cnote.py | 58 ---------------- g4f/Provider/not_working/Feedough.py | 78 ---------------------- g4f/Provider/not_working/Gpt6.py | 54 --------------- g4f/Provider/not_working/GptChatly.py | 35 ---------- g4f/Provider/not_working/GptForLove.py | 91 ------------------------- g4f/Provider/not_working/GptGo.py | 66 ------------------- g4f/Provider/not_working/GptGod.py | 61 ----------------- g4f/Provider/not_working/OnlineGpt.py | 57 ---------------- g4f/Provider/not_working/__init__.py | 21 ------ g4f/models.py | 1 - 26 files changed, 1671 deletions(-) delete mode 100644 g4f/Provider/Cohere.py delete mode 100644 g4f/Provider/GeminiProChat.py delete mode 100644 g4f/Provider/Marsyoo.py delete mode 100644 g4f/Provider/TeachAnything.py delete mode 100644 g4f/Provider/not_working/AItianhu.py delete mode 100644 g4f/Provider/not_working/Aichatos.py delete mode 100644 g4f/Provider/not_working/Bestim.py delete mode 100644 g4f/Provider/not_working/ChatBase.py delete mode 100644 g4f/Provider/not_working/ChatForAi.py delete mode 100644 g4f/Provider/not_working/ChatgptAi.py delete mode 100644 g4f/Provider/not_working/ChatgptDemo.py delete mode 100644 g4f/Provider/not_working/ChatgptDemoAi.py delete mode 100644 g4f/Provider/not_working/ChatgptLogin.py delete mode 100644 g4f/Provider/not_working/ChatgptNext.py delete mode 100644 g4f/Provider/not_working/ChatgptX.py delete mode 100644 g4f/Provider/not_working/Chatxyz.py delete mode 100644 g4f/Provider/not_working/Cnote.py delete mode 100644 g4f/Provider/not_working/Feedough.py delete mode 100644 g4f/Provider/not_working/Gpt6.py delete mode 100644 g4f/Provider/not_working/GptChatly.py delete mode 100644 g4f/Provider/not_working/GptForLove.py delete mode 100644 g4f/Provider/not_working/GptGo.py delete mode 100644 g4f/Provider/not_working/GptGod.py delete mode 100644 g4f/Provider/not_working/OnlineGpt.py delete mode 100644 g4f/Provider/not_working/__init__.py diff --git a/g4f/Provider/Cohere.py b/g4f/Provider/Cohere.py deleted file mode 100644 index eac04ab4..00000000 --- a/g4f/Provider/Cohere.py +++ /dev/null @@ -1,106 +0,0 @@ -from __future__ import annotations - -import json, random, requests, threading -from aiohttp import ClientSession - -from ..typing import CreateResult, Messages -from .base_provider import AbstractProvider -from .helper import format_prompt - -class Cohere(AbstractProvider): - url = "https://cohereforai-c4ai-command-r-plus.hf.space" - working = False - supports_gpt_35_turbo = False - supports_gpt_4 = False - supports_stream = True - - @staticmethod - def create_completion( - model: str, - messages: Messages, - stream: bool, - proxy: str = None, - max_retries: int = 6, - **kwargs - ) -> CreateResult: - - prompt = format_prompt(messages) - - headers = { - 'accept': 'text/event-stream', - 'accept-language': 'en,fr-FR;q=0.9,fr;q=0.8,es-ES;q=0.7,es;q=0.6,en-US;q=0.5,am;q=0.4,de;q=0.3', - 'cache-control': 'no-cache', - 'pragma': 'no-cache', - 'referer': 'https://cohereforai-c4ai-command-r-plus.hf.space/?__theme=light', - 'sec-ch-ua': '"Google Chrome";v="123", "Not:A-Brand";v="8", "Chromium";v="123"', - 'sec-ch-ua-mobile': '?0', - 'sec-ch-ua-platform': '"macOS"', - 'sec-fetch-dest': 'empty', - 'sec-fetch-mode': 'cors', - 'sec-fetch-site': 'same-origin', - 'user-agent': 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/123.0.0.0 Safari/537.36', - } - - session_hash = ''.join(random.choices("abcdefghijklmnopqrstuvwxyz0123456789", k=11)) - - params = { - 'fn_index': '1', - 'session_hash': session_hash, - } - - response = requests.get( - 'https://cohereforai-c4ai-command-r-plus.hf.space/queue/join', - params=params, - headers=headers, - stream=True - ) - - completion = '' - - for line in response.iter_lines(): - if line: - json_data = json.loads(line[6:]) - - if b"send_data" in (line): - event_id = json_data["event_id"] - - threading.Thread(target=send_data, args=[session_hash, event_id, prompt]).start() - - if b"process_generating" in line or b"process_completed" in line: - token = (json_data['output']['data'][0][0][1]) - - yield (token.replace(completion, "")) - completion = token - -def send_data(session_hash, event_id, prompt): - headers = { - 'accept': '*/*', - 'accept-language': 'en,fr-FR;q=0.9,fr;q=0.8,es-ES;q=0.7,es;q=0.6,en-US;q=0.5,am;q=0.4,de;q=0.3', - 'cache-control': 'no-cache', - 'content-type': 'application/json', - 'origin': 'https://cohereforai-c4ai-command-r-plus.hf.space', - 'pragma': 'no-cache', - 'referer': 'https://cohereforai-c4ai-command-r-plus.hf.space/?__theme=light', - 'sec-ch-ua': '"Google Chrome";v="123", "Not:A-Brand";v="8", "Chromium";v="123"', - 'sec-ch-ua-mobile': '?0', - 'sec-ch-ua-platform': '"macOS"', - 'sec-fetch-dest': 'empty', - 'sec-fetch-mode': 'cors', - 'sec-fetch-site': 'same-origin', - 'user-agent': 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/123.0.0.0 Safari/537.36', - } - - json_data = { - 'data': [ - prompt, - '', - [], - ], - 'event_data': None, - 'fn_index': 1, - 'session_hash': session_hash, - 'event_id': event_id - } - - requests.post('https://cohereforai-c4ai-command-r-plus.hf.space/queue/data', - json = json_data, headers=headers) \ No newline at end of file diff --git a/g4f/Provider/GeminiProChat.py b/g4f/Provider/GeminiProChat.py deleted file mode 100644 index 208ca773..00000000 --- a/g4f/Provider/GeminiProChat.py +++ /dev/null @@ -1,75 +0,0 @@ -from __future__ import annotations - -import time -from hashlib import sha256 - -from aiohttp import BaseConnector, ClientSession - -from ..errors import RateLimitError -from ..requests import raise_for_status -from ..requests.aiohttp import get_connector -from ..typing import AsyncResult, Messages -from .base_provider import AsyncGeneratorProvider, ProviderModelMixin - - -class GeminiProChat(AsyncGeneratorProvider, ProviderModelMixin): - url = "https://gemini-pro.chat/" - working = True - supports_message_history = True - default_model = 'gemini-pro' - - @classmethod - async def create_async_generator( - cls, - model: str, - messages: Messages, - proxy: str = None, - connector: BaseConnector = None, - **kwargs, - ) -> AsyncResult: - headers = { - "User-Agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:122.0) Gecko/20100101 Firefox/122.0", - "Accept": "*/*", - "Accept-Language": "en-US,en;q=0.5", - "Accept-Encoding": "gzip, deflate, br", - "Content-Type": "text/plain;charset=UTF-8", - "Referer": f"{cls.url}/", - "Origin": cls.url, - "Sec-Fetch-Dest": "empty", - "Sec-Fetch-Mode": "cors", - "Sec-Fetch-Site": "same-origin", - "Connection": "keep-alive", - "TE": "trailers", - } - async with ClientSession( - connector=get_connector(connector, proxy), headers=headers - ) as session: - timestamp = int(time.time() * 1e3) - data = { - "messages": [ - { - "role": "model" if message["role"] == "assistant" else "user", - "parts": [{"text": message["content"]}], - } - for message in messages - ], - "time": timestamp, - "pass": None, - "sign": generate_signature(timestamp, messages[-1]["content"]), - } - async with session.post( - f"{cls.url}/api/generate", json=data, proxy=proxy - ) as response: - if response.status == 500: - if "Quota exceeded" in await response.text(): - raise RateLimitError( - f"Response {response.status}: Rate limit reached" - ) - await raise_for_status(response) - async for chunk in response.content.iter_any(): - yield chunk.decode(errors="ignore") - - -def generate_signature(time: int, text: str, secret: str = ""): - message = f"{time}:{text}:{secret}" - return sha256(message.encode()).hexdigest() diff --git a/g4f/Provider/Marsyoo.py b/g4f/Provider/Marsyoo.py deleted file mode 100644 index 1c5fa9fd..00000000 --- a/g4f/Provider/Marsyoo.py +++ /dev/null @@ -1,64 +0,0 @@ -from __future__ import annotations - -import json -from aiohttp import ClientSession, ClientResponseError - -from ..typing import AsyncResult, Messages -from .base_provider import AsyncGeneratorProvider, ProviderModelMixin -from .helper import format_prompt - - -class Marsyoo(AsyncGeneratorProvider, ProviderModelMixin): - url = "https://aiagent.marsyoo.com" - api_endpoint = "/api/chat-messages" - working = True - supports_gpt_4 = True - default_model = 'gpt-4o' - - @classmethod - async def create_async_generator( - cls, - model: str, - messages: Messages, - proxy: str = None, - **kwargs - ) -> AsyncResult: - headers = { - "Accept": "*/*", - "Accept-Language": "en-US,en;q=0.9", - "Connection": "keep-alive", - "DNT": "1", - "Origin": cls.url, - "Referer": f"{cls.url}/chat", - "Sec-Fetch-Dest": "empty", - "Sec-Fetch-Mode": "cors", - "Sec-Fetch-Site": "same-origin", - "User-Agent": "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/126.0.0.0 Safari/537.36", - "authorization": "Bearer eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpc3MiOiI0MWNkOTE3MS1mNTg1LTRjMTktOTY0Ni01NzgxMTBjYWViNTciLCJzdWIiOiJXZWIgQVBJIFBhc3Nwb3J0IiwiYXBwX2lkIjoiNDFjZDkxNzEtZjU4NS00YzE5LTk2NDYtNTc4MTEwY2FlYjU3IiwiYXBwX2NvZGUiOiJMakhzdWJqNjhMTXZCT0JyIiwiZW5kX3VzZXJfaWQiOiI4YjE5YjY2Mi05M2E1LTRhYTktOGNjNS03MDhmNWE0YmQxNjEifQ.pOzdQ4wTrQjjRlEv1XY9TZitkW5KW1K-wbcUJAoBJ5I", - "content-type": "application/json", - "sec-ch-ua": '"Not/A)Brand";v="8", "Chromium";v="126"', - "sec-ch-ua-mobile": "?0", - "sec-ch-ua-platform": "Linux", - } - async with ClientSession(headers=headers) as session: - prompt = format_prompt(messages) - data = { - "response_mode": "streaming", - "query": prompt, - "inputs": {}, - } - try: - async with session.post(f"{cls.url}{cls.api_endpoint}", json=data, proxy=proxy) as response: - response.raise_for_status() - async for line in response.content: - if line: - try: - json_data = json.loads(line.decode('utf-8').strip().lstrip('data: ')) - if json_data['event'] == 'message': - yield json_data['answer'] - elif json_data['event'] == 'message_end': - return - except json.JSONDecodeError: - continue - except ClientResponseError as e: - yield f"Error: HTTP {e.status}: {e.message}" diff --git a/g4f/Provider/TeachAnything.py b/g4f/Provider/TeachAnything.py deleted file mode 100644 index 908dd56e..00000000 --- a/g4f/Provider/TeachAnything.py +++ /dev/null @@ -1,62 +0,0 @@ -from __future__ import annotations - -from typing import Any, Dict - -from aiohttp import ClientSession, ClientTimeout - -from ..typing import AsyncResult, Messages -from .base_provider import AsyncGeneratorProvider, ProviderModelMixin -from .helper import format_prompt - - -class TeachAnything(AsyncGeneratorProvider, ProviderModelMixin): - url = "https://www.teach-anything.com" - api_endpoint = "/api/generate" - working = True - default_model = "llama-3-70b-instruct" - - @classmethod - async def create_async_generator( - cls, - model: str, - messages: Messages, - proxy: str | None = None, - **kwargs: Any - ) -> AsyncResult: - headers = cls._get_headers() - - async with ClientSession(headers=headers) as session: - prompt = format_prompt(messages) - data = {"prompt": prompt} - - timeout = ClientTimeout(total=60) - - async with session.post( - f"{cls.url}{cls.api_endpoint}", - json=data, - proxy=proxy, - timeout=timeout - ) as response: - response.raise_for_status() - async for chunk in response.content.iter_any(): - if chunk: - yield chunk.decode() - - @staticmethod - def _get_headers() -> Dict[str, str]: - return { - "accept": "*/*", - "accept-language": "en-US,en;q=0.9", - "content-type": "application/json", - "dnt": "1", - "origin": "https://www.teach-anything.com", - "priority": "u=1, i", - "referer": "https://www.teach-anything.com/", - "sec-ch-ua": '"Not/A)Brand";v="8", "Chromium";v="126"', - "sec-ch-ua-mobile": "?0", - "sec-ch-ua-platform": '"Linux"', - "sec-fetch-dest": "empty", - "sec-fetch-mode": "cors", - "sec-fetch-site": "same-origin", - "user-agent": "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/126.0.0.0 Safari/537.36" - } diff --git a/g4f/Provider/not_working/AItianhu.py b/g4f/Provider/not_working/AItianhu.py deleted file mode 100644 index 501b334e..00000000 --- a/g4f/Provider/not_working/AItianhu.py +++ /dev/null @@ -1,79 +0,0 @@ -from __future__ import annotations - -import json - -from ...typing import AsyncResult, Messages -from ...requests import StreamSession -from ..base_provider import AsyncGeneratorProvider, format_prompt, get_cookies - - -class AItianhu(AsyncGeneratorProvider): - url = "https://www.aitianhu.com" - working = False - supports_gpt_35_turbo = True - - @classmethod - async def create_async_generator( - cls, - model: str, - messages: Messages, - proxy: str = None, - cookies: dict = None, - timeout: int = 120, **kwargs) -> AsyncResult: - - if not cookies: - cookies = get_cookies(domain_name='www.aitianhu.com') - if not cookies: - raise RuntimeError(f"g4f.provider.{cls.__name__} requires cookies [refresh https://www.aitianhu.com on chrome]") - - data = { - "prompt": format_prompt(messages), - "options": {}, - "systemMessage": "You are ChatGPT, a large language model trained by OpenAI. Follow the user's instructions carefully.", - "temperature": 0.8, - "top_p": 1, - **kwargs - } - - headers = { - 'authority': 'www.aitianhu.com', - 'accept': 'application/json, text/plain, */*', - 'accept-language': 'en,fr-FR;q=0.9,fr;q=0.8,es-ES;q=0.7,es;q=0.6,en-US;q=0.5,am;q=0.4,de;q=0.3', - 'content-type': 'application/json', - 'origin': 'https://www.aitianhu.com', - 'referer': 'https://www.aitianhu.com/', - 'sec-ch-ua': '"Chromium";v="118", "Google Chrome";v="118", "Not=A?Brand";v="99"', - 'sec-ch-ua-mobile': '?0', - 'sec-ch-ua-platform': '"macOS"', - 'sec-fetch-dest': 'empty', - 'sec-fetch-mode': 'cors', - 'sec-fetch-site': 'same-origin', - 'user-agent': 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/118.0.0.0 Safari/537.36', - } - - async with StreamSession(headers=headers, - cookies=cookies, - timeout=timeout, - proxies={"https": proxy}, - impersonate="chrome107", verify=False) as session: - - async with session.post(f"{cls.url}/api/chat-process", json=data) as response: - response.raise_for_status() - - async for line in response.iter_lines(): - if line == b"