summaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorTekky <98614666+xtekky@users.noreply.github.com>2024-05-10 14:46:38 +0200
committerGitHub <noreply@github.com>2024-05-10 14:46:38 +0200
commit23f3c5b809921eb40bc7a4eb7a46caa3c907c93d (patch)
tree7d23209e4f5e18ff43e70c84157a2888423760e5
parentMerge pull request #1933 from HyiKi/main (diff)
parentUpdated PerplexityLabs models (diff)
downloadgpt4free-main.tar
gpt4free-main.tar.gz
gpt4free-main.tar.bz2
gpt4free-main.tar.lz
gpt4free-main.tar.xz
gpt4free-main.tar.zst
gpt4free-main.zip
-rw-r--r--g4f/Provider/PerplexityLabs.py5
1 files changed, 4 insertions, 1 deletions
diff --git a/g4f/Provider/PerplexityLabs.py b/g4f/Provider/PerplexityLabs.py
index b6fec53c..91ba63f2 100644
--- a/g4f/Provider/PerplexityLabs.py
+++ b/g4f/Provider/PerplexityLabs.py
@@ -15,7 +15,10 @@ class PerplexityLabs(AsyncGeneratorProvider, ProviderModelMixin):
working = True
default_model = "mixtral-8x7b-instruct"
models = [
- "sonar-small-online", "sonar-medium-online", "sonar-small-chat", "sonar-medium-chat", "dbrx-instruct", "claude-3-haiku-20240307", "llama-3-8b-instruct", "llama-3-70b-instruct", "codellama-70b-instruct", "mistral-7b-instruct", "llava-v1.5-7b-wrapper", "llava-v1.6-34b", "mixtral-8x7b-instruct", "mixtral-8x22b-instruct", "mistral-medium", "gemma-2b-it", "gemma-7b-it", "related"
+ "llama-3-sonar-large-32k-online", "llama-3-sonar-small-32k-online", "llama-3-sonar-large-32k-chat", "llama-3-sonar-small-32k-chat",
+ "dbrx-instruct", "claude-3-haiku-20240307", "llama-3-8b-instruct", "llama-3-70b-instruct", "codellama-70b-instruct", "mistral-7b-instruct",
+ "llava-v1.5-7b-wrapper", "llava-v1.6-34b", "mixtral-8x7b-instruct", "mixtral-8x22b-instruct", "mistral-medium", "gemma-2b-it", "gemma-7b-it",
+ "related"
]
model_aliases = {
"mistralai/Mistral-7B-Instruct-v0.1": "mistral-7b-instruct",