summaryrefslogblamecommitdiffstats
path: root/g4f/Provider/FreeChatgpt.py
blob: 82fc3ef41c5aef168c2566956cbb385cb1bafeff (plain) (tree)
1
2
3
4
5
6
7
8
9








                                                 








                                         














                                          
                                                                          
                   











                                                                                                                           
                                                             












                                                                                                                                           
from __future__ import annotations

import json
from aiohttp import ClientSession

from ..typing import AsyncResult, Messages
from .base_provider import AsyncGeneratorProvider



models = {
     "claude-v1":"claude-2.1",
     "claude-v2":"claude-2.0",
     "gpt_35_turbo":"gpt-3.5-turbo-1106",
     "gpt-4":"gpt-4",
     "gemini-pro":"google-gemini-pro"
}

class FreeChatgpt(AsyncGeneratorProvider):
    url = "https://free.chatgpt.org.uk"
    working = True
    supports_gpt_35_turbo = True
    supports_gpt_4 = True
    supports_message_history = True

    @classmethod
    async def create_async_generator(
        cls,
        model: str,
        messages: Messages,
        proxy: str = None,
        **kwargs
    ) -> AsyncResult:
        model = models[model] if model in models else "gpt-3.5-turbo-1106"
        headers = {
    "Accept": "application/json, text/event-stream",
    "Content-Type":"application/json",
    "Accept-Encoding": "gzip, deflate, br",
    "Accept-Language": "en-US,en;q=0.5",
    "Host":"free.chatgpt.org.uk",
    "Referer":f"{cls.url}/",
    "Origin":f"{cls.url}",
    "Sec-Fetch-Dest": "empty",
    "Sec-Fetch-Mode": "cors",
    "Sec-Fetch-Site": "same-origin",
    "User-Agent": "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/120.0.0.0 Safari/537.36", 
}
        async with ClientSession(headers=headers) as session:
            data = {"messages":messages,"stream":True,"model":model,"temperature":0.5,"presence_penalty":0,"frequency_penalty":0,"top_p":1}
            async with session.post(f'{cls.url}/api/openai/v1/chat/completions',json=data) as result:
                async for chunk in result.content:
                    
                    line = chunk.decode()
                    if line.startswith("data: [DONE]"):
                            break
                    elif line.startswith("data: "):
                         line = json.loads(line[6:])
                         if(line["choices"]==[]):
                              continue
                         if(line["choices"][0]["delta"].get("content") and line["choices"][0]["delta"]["content"]!=None):
                              yield line["choices"][0]["delta"]["content"]