summaryrefslogtreecommitdiffstats
path: root/g4f/models.py
diff options
context:
space:
mode:
authorH Lohaus <hlohaus@users.noreply.github.com>2024-03-11 07:47:35 +0100
committerGitHub <noreply@github.com>2024-03-11 07:47:35 +0100
commit0b850ac9fcabc52fecdeea8d56e620e44b35d180 (patch)
treef92578686f34476be49422889d4587d8f3cf2fa6 /g4f/models.py
parentMerge pull request #1670 from xtekky/xtekky-patch-1 (diff)
parentFix HuggingChat and PerplexityLabs and add HuggingFace provider (diff)
downloadgpt4free-0b850ac9fcabc52fecdeea8d56e620e44b35d180.tar
gpt4free-0b850ac9fcabc52fecdeea8d56e620e44b35d180.tar.gz
gpt4free-0b850ac9fcabc52fecdeea8d56e620e44b35d180.tar.bz2
gpt4free-0b850ac9fcabc52fecdeea8d56e620e44b35d180.tar.lz
gpt4free-0b850ac9fcabc52fecdeea8d56e620e44b35d180.tar.xz
gpt4free-0b850ac9fcabc52fecdeea8d56e620e44b35d180.tar.zst
gpt4free-0b850ac9fcabc52fecdeea8d56e620e44b35d180.zip
Diffstat (limited to 'g4f/models.py')
-rw-r--r--g4f/models.py27
1 files changed, 21 insertions, 6 deletions
diff --git a/g4f/models.py b/g4f/models.py
index bf58ff25..b3e0d0ee 100644
--- a/g4f/models.py
+++ b/g4f/models.py
@@ -10,6 +10,7 @@ from .Provider import (
GeminiProChat,
ChatgptNext,
HuggingChat,
+ HuggingFace,
ChatgptDemo,
FreeChatgpt,
GptForLove,
@@ -112,32 +113,32 @@ llama2_13b = Model(
llama2_70b = Model(
name = "meta-llama/Llama-2-70b-chat-hf",
base_provider = "meta",
- best_provider = RetryProvider([Llama2, DeepInfra, HuggingChat, PerplexityLabs])
+ best_provider = RetryProvider([Llama2, DeepInfra, HuggingChat])
)
codellama_34b_instruct = Model(
name = "codellama/CodeLlama-34b-Instruct-hf",
base_provider = "meta",
- best_provider = RetryProvider([HuggingChat, PerplexityLabs, DeepInfra])
+ best_provider = RetryProvider([HuggingChat, DeepInfra])
)
codellama_70b_instruct = Model(
name = "codellama/CodeLlama-70b-Instruct-hf",
base_provider = "meta",
- best_provider = DeepInfra
+ best_provider = RetryProvider([DeepInfra, PerplexityLabs])
)
# Mistral
mixtral_8x7b = Model(
name = "mistralai/Mixtral-8x7B-Instruct-v0.1",
base_provider = "huggingface",
- best_provider = RetryProvider([DeepInfra, HuggingChat, PerplexityLabs])
+ best_provider = RetryProvider([DeepInfra, HuggingChat, HuggingFace, PerplexityLabs])
)
mistral_7b = Model(
name = "mistralai/Mistral-7B-Instruct-v0.1",
base_provider = "huggingface",
- best_provider = RetryProvider([DeepInfra, HuggingChat, PerplexityLabs])
+ best_provider = RetryProvider([DeepInfra, HuggingChat, HuggingFace, PerplexityLabs])
)
# Misc models
@@ -184,6 +185,18 @@ claude_v2 = Model(
best_provider = RetryProvider([FreeChatgpt, Vercel])
)
+claude_3_opus = Model(
+ name = 'claude-3-opus',
+ base_provider = 'anthropic',
+ best_provider = You
+)
+
+claude_3_sonnet = Model(
+ name = 'claude-3-sonnet',
+ base_provider = 'anthropic',
+ best_provider = You
+)
+
gpt_35_turbo_16k = Model(
name = 'gpt-3.5-turbo-16k',
base_provider = 'openai',
@@ -223,7 +236,7 @@ gpt_4_32k_0613 = Model(
gemini_pro = Model(
name = 'gemini-pro',
base_provider = 'google',
- best_provider = RetryProvider([FreeChatgpt, GeminiProChat])
+ best_provider = RetryProvider([FreeChatgpt, GeminiProChat, You])
)
pi = Model(
@@ -272,6 +285,8 @@ class ModelUtils:
'gemini': gemini,
'gemini-pro': gemini_pro,
'claude-v2': claude_v2,
+ 'claude-3-opus': claude_3_opus,
+ 'claude-3-sonnet': claude_3_sonnet,
'pi': pi
}