diff options
author | Tekky <98614666+xtekky@users.noreply.github.com> | 2024-10-22 23:32:27 +0200 |
---|---|---|
committer | GitHub <noreply@github.com> | 2024-10-22 23:32:27 +0200 |
commit | a63c18de796bd4f3e818ff170b6ff595304f95e0 (patch) | |
tree | 844dbb9a8d3526a8b60564b78f7a19a4e0f605d9 /g4f/Provider/nexra/NexraLLaMA31.py | |
parent | Merge pull request #2282 from Karasiq/patch-1 (diff) | |
parent | Updated docs/providers-and-models.md g4f/models.py g4f/Provider/Upstage.py (diff) | |
download | gpt4free-a63c18de796bd4f3e818ff170b6ff595304f95e0.tar gpt4free-a63c18de796bd4f3e818ff170b6ff595304f95e0.tar.gz gpt4free-a63c18de796bd4f3e818ff170b6ff595304f95e0.tar.bz2 gpt4free-a63c18de796bd4f3e818ff170b6ff595304f95e0.tar.lz gpt4free-a63c18de796bd4f3e818ff170b6ff595304f95e0.tar.xz gpt4free-a63c18de796bd4f3e818ff170b6ff595304f95e0.tar.zst gpt4free-a63c18de796bd4f3e818ff170b6ff595304f95e0.zip |
Diffstat (limited to '')
-rw-r--r-- | g4f/Provider/nexra/NexraLLaMA31.py | 91 |
1 files changed, 0 insertions, 91 deletions
diff --git a/g4f/Provider/nexra/NexraLLaMA31.py b/g4f/Provider/nexra/NexraLLaMA31.py deleted file mode 100644 index d461f2b2..00000000 --- a/g4f/Provider/nexra/NexraLLaMA31.py +++ /dev/null @@ -1,91 +0,0 @@ -from __future__ import annotations - -from aiohttp import ClientSession -import json - -from ...typing import AsyncResult, Messages -from ..base_provider import AsyncGeneratorProvider, ProviderModelMixin -from ..helper import format_prompt - - -class NexraLLaMA31(AsyncGeneratorProvider, ProviderModelMixin): - label = "Nexra LLaMA 3.1" - url = "https://nexra.aryahcr.cc/documentation/llama-3.1/en" - api_endpoint = "https://nexra.aryahcr.cc/api/chat/complements" - working = True - supports_stream = True - - default_model = 'llama-3.1' - models = [default_model] - model_aliases = { - "llama-3.1-8b": "llama-3.1", - } - - @classmethod - def get_model(cls, model: str) -> str: - if model in cls.models: - return model - elif model in cls.model_aliases: - return cls.model_aliases.get(model, cls.default_model) - else: - return cls.default_model - - @classmethod - async def create_async_generator( - cls, - model: str, - messages: Messages, - proxy: str = None, - stream: bool = False, - markdown: bool = False, - **kwargs - ) -> AsyncResult: - model = cls.get_model(model) - - headers = { - "Content-Type": "application/json" - } - - async with ClientSession(headers=headers) as session: - prompt = format_prompt(messages) - data = { - "messages": [ - { - "role": "user", - "content": prompt - } - ], - "stream": stream, - "markdown": markdown, - "model": model - } - - async with session.post(f"{cls.api_endpoint}", json=data, proxy=proxy) as response: - response.raise_for_status() - - if stream: - # Streamed response handling - collected_message = "" - async for chunk in response.content.iter_any(): - if chunk: - decoded_chunk = chunk.decode().strip().split("\x1e") - for part in decoded_chunk: - if part: - message_data = json.loads(part) - - # Collect messages until 'finish': true - if 'message' in message_data and message_data['message']: - collected_message = message_data['message'] - - # When finish is true, yield the final collected message - if message_data.get('finish', False): - yield collected_message - return - else: - # Non-streamed response handling - response_data = await response.json(content_type=None) - - # Yield the message directly from the response - if 'message' in response_data and response_data['message']: - yield response_data['message'] - return |