summaryrefslogtreecommitdiffstats
path: root/g4f/Provider/PerplexityLabs.py
diff options
context:
space:
mode:
authorHeiner Lohaus <hlohaus@users.noreply.github.com>2024-01-23 05:02:14 +0100
committerHeiner Lohaus <hlohaus@users.noreply.github.com>2024-01-23 05:02:14 +0100
commitf5518bb94d9ec78016b6e98da414647473f761e9 (patch)
tree3e65e24c47f6a2600442d7935baf2503ad4f3e4f /g4f/Provider/PerplexityLabs.py
parentRevert "Cleanup unused dependencies" (diff)
downloadgpt4free-f5518bb94d9ec78016b6e98da414647473f761e9.tar
gpt4free-f5518bb94d9ec78016b6e98da414647473f761e9.tar.gz
gpt4free-f5518bb94d9ec78016b6e98da414647473f761e9.tar.bz2
gpt4free-f5518bb94d9ec78016b6e98da414647473f761e9.tar.lz
gpt4free-f5518bb94d9ec78016b6e98da414647473f761e9.tar.xz
gpt4free-f5518bb94d9ec78016b6e98da414647473f761e9.tar.zst
gpt4free-f5518bb94d9ec78016b6e98da414647473f761e9.zip
Diffstat (limited to '')
-rw-r--r--g4f/Provider/PerplexityLabs.py29
1 files changed, 15 insertions, 14 deletions
diff --git a/g4f/Provider/PerplexityLabs.py b/g4f/Provider/PerplexityLabs.py
index 7e76aeef..c989b3da 100644
--- a/g4f/Provider/PerplexityLabs.py
+++ b/g4f/Provider/PerplexityLabs.py
@@ -2,27 +2,28 @@ from __future__ import annotations
import random
import json
-from aiohttp import ClientSession, WSMsgType
+from aiohttp import ClientSession
from ..typing import AsyncResult, Messages
from .base_provider import AsyncGeneratorProvider
API_URL = "https://labs-api.perplexity.ai/socket.io/"
WS_URL = "wss://labs-api.perplexity.ai/socket.io/"
-MODELS = ['pplx-7b-online', 'pplx-70b-online', 'pplx-7b-chat', 'pplx-70b-chat', 'mistral-7b-instruct',
- 'codellama-34b-instruct', 'llama-2-70b-chat', 'llava-7b-chat', 'mixtral-8x7b-instruct',
- 'mistral-medium', 'related']
-DEFAULT_MODEL = MODELS[1]
-MODEL_MAP = {
- "mistralai/Mistral-7B-Instruct-v0.1": "mistral-7b-instruct",
- "meta-llama/Llama-2-70b-chat-hf": "llama-2-70b-chat",
- "mistralai/Mixtral-8x7B-Instruct-v0.1": "mixtral-8x7b-instruct",
-}
class PerplexityLabs(AsyncGeneratorProvider):
url = "https://labs.perplexity.ai"
working = True
supports_gpt_35_turbo = True
+ models = ['pplx-7b-online', 'pplx-70b-online', 'pplx-7b-chat', 'pplx-70b-chat', 'mistral-7b-instruct',
+ 'codellama-34b-instruct', 'llama-2-70b-chat', 'llava-7b-chat', 'mixtral-8x7b-instruct',
+ 'mistral-medium', 'related']
+ default_model = 'pplx-70b-online'
+ model_map = {
+ "mistralai/Mistral-7B-Instruct-v0.1": "mistral-7b-instruct",
+ "meta-llama/Llama-2-70b-chat-hf": "llama-2-70b-chat",
+ "mistralai/Mixtral-8x7B-Instruct-v0.1": "mixtral-8x7b-instruct",
+ "codellama/CodeLlama-34b-Instruct-hf": "codellama-34b-instruct"
+ }
@classmethod
async def create_async_generator(
@@ -33,10 +34,10 @@ class PerplexityLabs(AsyncGeneratorProvider):
**kwargs
) -> AsyncResult:
if not model:
- model = DEFAULT_MODEL
- elif model in MODEL_MAP:
- model = MODEL_MAP[model]
- elif model not in MODELS:
+ model = cls.default_model
+ elif model in cls.model_map:
+ model = cls.model_map[model]
+ elif model not in cls.models:
raise ValueError(f"Model is not supported: {model}")
headers = {
"User-Agent": "Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:121.0) Gecko/20100101 Firefox/121.0",