summaryrefslogtreecommitdiffstats
path: root/g4f/models.py
diff options
context:
space:
mode:
authorkqlio67 <kqlio67@users.noreply.github.com>2024-09-27 11:13:11 +0200
committerkqlio67 <kqlio67@users.noreply.github.com>2024-09-27 11:13:11 +0200
commit105b4aac75f26c9c216c45112ee5a8b065b5568d (patch)
treef3181df4790874df2bb884a3dacbd7579e1182a8 /g4f/models.py
parentUpdated g4f/models.py g4f/Provider/Liaobots.py g4f/Provider/__init__.py g4f/Provider/DeepInfraChat.py (diff)
downloadgpt4free-105b4aac75f26c9c216c45112ee5a8b065b5568d.tar
gpt4free-105b4aac75f26c9c216c45112ee5a8b065b5568d.tar.gz
gpt4free-105b4aac75f26c9c216c45112ee5a8b065b5568d.tar.bz2
gpt4free-105b4aac75f26c9c216c45112ee5a8b065b5568d.tar.lz
gpt4free-105b4aac75f26c9c216c45112ee5a8b065b5568d.tar.xz
gpt4free-105b4aac75f26c9c216c45112ee5a8b065b5568d.tar.zst
gpt4free-105b4aac75f26c9c216c45112ee5a8b065b5568d.zip
Diffstat (limited to '')
-rw-r--r--g4f/models.py77
1 files changed, 50 insertions, 27 deletions
diff --git a/g4f/models.py b/g4f/models.py
index bac83b5b..dcd3a525 100644
--- a/g4f/models.py
+++ b/g4f/models.py
@@ -16,6 +16,7 @@ from .Provider import (
Chatgpt4Online,
ChatGptEs,
ChatgptFree,
+ ChatHub,
DDG,
DeepInfra,
DeepInfraChat,
@@ -96,9 +97,7 @@ default = Model(
gpt_3 = Model(
name = 'gpt-3',
base_provider = 'OpenAI',
- best_provider = IterListProvider([
- Nexra,
- ])
+ best_provider = Nexra
)
# gpt-3.5
@@ -167,7 +166,7 @@ meta = Model(
llama_2_13b = Model(
name = "llama-2-13b",
base_provider = "Meta Llama",
- best_provider = IterListProvider([Airforce])
+ best_provider = Airforce
)
# llama 3
@@ -193,7 +192,7 @@ llama_3 = Model(
llama_3_1_8b = Model(
name = "llama-3.1-8b",
base_provider = "Meta Llama",
- best_provider = IterListProvider([Blackbox, DeepInfraChat, Airforce, PerplexityLabs])
+ best_provider = IterListProvider([Blackbox, DeepInfraChat, ChatHub, Airforce, PerplexityLabs])
)
llama_3_1_70b = Model(
@@ -225,7 +224,7 @@ mistral_7b = Model(
mixtral_8x7b = Model(
name = "mixtral-8x7b",
base_provider = "Mistral",
- best_provider = IterListProvider([DDG, ReplicateHome, DeepInfraChat, Airforce, DeepInfra])
+ best_provider = IterListProvider([DDG, ReplicateHome, DeepInfraChat, ChatHub, Airforce, DeepInfra])
)
mixtral_8x22b = Model(
@@ -245,7 +244,7 @@ mistral_nemo = Model(
mixtral_8x7b_dpo = Model(
name = "mixtral-8x7b-dpo",
base_provider = "NousResearch",
- best_provider = IterListProvider([Airforce])
+ best_provider = Airforce
)
hermes_3 = Model(
@@ -259,7 +258,7 @@ hermes_3 = Model(
phi_3_medium_4k = Model(
name = "phi-3-medium-4k",
base_provider = "Microsoft",
- best_provider = IterListProvider([DeepInfraChat])
+ best_provider = DeepInfraChat
)
phi_3_5_mini = Model(
@@ -295,7 +294,7 @@ gemini = Model(
gemma_2b_9b = Model(
name = 'gemma-2b-9b',
base_provider = 'Google',
- best_provider = IterListProvider([Airforce])
+ best_provider = Airforce
)
gemma_2b_27b = Model(
@@ -313,37 +312,43 @@ gemma_2b = Model(
])
)
+gemma_2 = Model(
+ name = 'gemma-2',
+ base_provider = 'Google',
+ best_provider = ChatHub
+)
+
### Anthropic ###
claude_2 = Model(
name = 'claude-2',
base_provider = 'Anthropic',
- best_provider = IterListProvider([You])
+ best_provider = You
)
claude_2_0 = Model(
name = 'claude-2.0',
base_provider = 'Anthropic',
- best_provider = IterListProvider([Liaobots])
+ best_provider = Liaobots
)
claude_2_1 = Model(
name = 'claude-2.1',
base_provider = 'Anthropic',
- best_provider = IterListProvider([Liaobots])
+ best_provider = Liaobots
)
# claude 3
claude_3_opus = Model(
name = 'claude-3-opus',
base_provider = 'Anthropic',
- best_provider = IterListProvider([Liaobots])
+ best_provider = Liaobots
)
claude_3_sonnet = Model(
name = 'claude-3-sonnet',
base_provider = 'Anthropic',
- best_provider = IterListProvider([Liaobots])
+ best_provider = Liaobots
)
claude_3_haiku = Model(
@@ -390,7 +395,7 @@ reka_core = Model(
blackbox = Model(
name = 'blackbox',
base_provider = 'Blackbox AI',
- best_provider = IterListProvider([Blackbox])
+ best_provider = Blackbox
)
@@ -406,7 +411,7 @@ dbrx_instruct = Model(
command_r_plus = Model(
name = 'command-r-plus',
base_provider = 'CohereForAI',
- best_provider = IterListProvider([HuggingChat])
+ best_provider = HuggingChat
)
@@ -422,19 +427,19 @@ sparkdesk_v1_1 = Model(
qwen_1_5_14b = Model(
name = 'qwen-1.5-14b',
base_provider = 'Qwen',
- best_provider = IterListProvider([FreeChatgpt])
+ best_provider = FreeChatgpt
)
qwen_1_5_72b = Model(
name = 'qwen-1.5-72b',
base_provider = 'Qwen',
- best_provider = IterListProvider([Airforce])
+ best_provider = Airforce
)
qwen_1_5_110b = Model(
name = 'qwen-1.5-110b',
base_provider = 'Qwen',
- best_provider = IterListProvider([Airforce])
+ best_provider = Airforce
)
qwen_2_72b = Model(
@@ -446,7 +451,7 @@ qwen_2_72b = Model(
qwen_turbo = Model(
name = 'qwen-turbo',
base_provider = 'Qwen',
- best_provider = IterListProvider([Bixin123])
+ best_provider = Bixin123
)
qwen = Model(
@@ -462,13 +467,13 @@ qwen = Model(
glm_3_6b = Model(
name = 'glm-3-6b',
base_provider = 'Zhipu AI',
- best_provider = IterListProvider([FreeChatgpt])
+ best_provider = FreeChatgpt
)
glm_4_9b = Model(
name = 'glm-4-9B',
base_provider = 'Zhipu AI',
- best_provider = IterListProvider([FreeChatgpt])
+ best_provider = FreeChatgpt
)
glm_4 = Model(
@@ -484,13 +489,13 @@ glm_4 = Model(
yi_1_5_9b = Model(
name = 'yi-1.5-9b',
base_provider = '01-ai',
- best_provider = IterListProvider([FreeChatgpt])
+ best_provider = FreeChatgpt
)
yi_34b = Model(
name = 'yi-34b',
base_provider = '01-ai',
- best_provider = IterListProvider([Airforce])
+ best_provider = Airforce
)
@@ -498,7 +503,7 @@ yi_34b = Model(
solar_1_mini = Model(
name = 'solar-1-mini',
base_provider = 'Upstage',
- best_provider = IterListProvider([Upstage])
+ best_provider = Upstage
)
solar_10_7b = Model(
@@ -525,14 +530,14 @@ pi = Model(
deepseek = Model(
name = 'deepseek',
base_provider = 'DeepSeek',
- best_provider = IterListProvider([Airforce])
+ best_provider = Airforce
)
### WizardLM ###
wizardlm_2_7b = Model(
name = 'wizardlm-2-7b',
base_provider = 'WizardLM',
- best_provider = IterListProvider([DeepInfraChat])
+ best_provider = DeepInfraChat
)
wizardlm_2_8x22b = Model(
@@ -610,6 +615,19 @@ grok_2_mini = Model(
best_provider = Liaobots
)
+# Perplexity
+sonar_online = Model(
+ name = 'sonar-online',
+ base_provider = 'Perplexity',
+ best_provider = IterListProvider([ChatHub, PerplexityLabs])
+)
+
+sonar_chat = Model(
+ name = 'sonar-chat',
+ base_provider = 'Perplexity',
+ best_provider = PerplexityLabs
+)
+
#############
@@ -817,6 +835,7 @@ class ModelUtils:
'gemma-2b': gemma_2b,
'gemma-2b-9b': gemma_2b_9b,
'gemma-2b-27b': gemma_2b_27b,
+'gemma-2': gemma_2,
### Anthropic ###
@@ -928,6 +947,10 @@ class ModelUtils:
'grok-2': grok_2,
'grok-2-mini': grok_2_mini,
+### Perplexity ###
+'sonar-online': sonar_online,
+'sonar-chat': sonar_chat,
+
#############