From 11bec81dc49412bf4ff10b5e09d452766ee043d4 Mon Sep 17 00:00:00 2001
From: kqlio67 <kqlio67@users.noreply.github.com>
Date: Fri, 1 Nov 2024 18:01:29 +0200
Subject: Update (g4f/Provider/Airforce.py)

---
 g4f/models.py | 239 ++++++++++++++++++++++++++++++++--------------------------
 1 file changed, 132 insertions(+), 107 deletions(-)

(limited to 'g4f/models.py')

diff --git a/g4f/models.py b/g4f/models.py
index bea09f28..ae3d430d 100644
--- a/g4f/models.py
+++ b/g4f/models.py
@@ -130,7 +130,7 @@ gpt_3 = Model(
 gpt_35_turbo = Model(
     name          = 'gpt-3.5-turbo',
     base_provider = 'OpenAI',
-    best_provider = IterListProvider([Allyfy, NexraChatGPT, Airforce, DarkAI, Liaobots])
+    best_provider = IterListProvider([Allyfy, NexraChatGPT, DarkAI, Airforce, Liaobots])
 )
 
 # gpt-4
@@ -191,7 +191,7 @@ meta = Model(
 llama_2_7b = Model(
     name          = "llama-2-7b",
     base_provider = "Meta Llama",
-    best_provider = Cloudflare
+    best_provider = IterListProvider([Cloudflare, Airforce])
 )
 
 llama_2_13b = Model(
@@ -217,13 +217,13 @@ llama_3_70b = Model(
 llama_3_1_8b = Model(
     name          = "llama-3.1-8b",
     base_provider = "Meta Llama",
-    best_provider = IterListProvider([Blackbox, DeepInfraChat, ChatHub, Cloudflare, Airforce, GizAI, PerplexityLabs])
+    best_provider = IterListProvider([Blackbox, DeepInfraChat, ChatHub, Cloudflare, GizAI, Airforce, PerplexityLabs])
 )
 
 llama_3_1_70b = Model(
     name          = "llama-3.1-70b",
     base_provider = "Meta Llama",
-    best_provider = IterListProvider([DDG, HuggingChat, Blackbox, FreeGpt, TeachAnything, Free2GPT, DeepInfraChat, DarkAI, Airforce, AiMathGPT, RubiksAI, GizAI, HuggingFace, PerplexityLabs])
+    best_provider = IterListProvider([DDG, HuggingChat, Blackbox, FreeGpt, TeachAnything, Free2GPT, DeepInfraChat, DarkAI, AiMathGPT, RubiksAI, GizAI, Airforce, HuggingFace, PerplexityLabs])
 )
 
 llama_3_1_405b = Model(
@@ -236,19 +236,19 @@ llama_3_1_405b = Model(
 llama_3_2_1b = Model(
     name          = "llama-3.2-1b",
     base_provider = "Meta Llama",
-    best_provider = Cloudflare
+    best_provider = IterListProvider([Cloudflare, Airforce])
 )
 
 llama_3_2_3b = Model(
     name          = "llama-3.2-3b",
     base_provider = "Meta Llama",
-    best_provider = Cloudflare
+    best_provider = IterListProvider([Cloudflare, Airforce])
 )
 
 llama_3_2_11b = Model(
     name          = "llama-3.2-11b",
     base_provider = "Meta Llama",
-    best_provider = IterListProvider([Cloudflare, HuggingChat, HuggingFace])
+    best_provider = IterListProvider([Cloudflare, HuggingChat, Airforce, HuggingFace])
 )
 
 llama_3_2_90b = Model(
@@ -271,6 +271,18 @@ llamaguard_2_8b = Model(
     best_provider = Airforce
 )
 
+llamaguard_3_8b = Model(
+    name          = "llamaguard-3-8b",
+    base_provider = "Meta Llama",
+    best_provider = Airforce
+)
+
+llamaguard_3_11b = Model(
+    name          = "llamaguard-3-11b",
+    base_provider = "Meta Llama",
+    best_provider = Airforce
+)
+
 
 ### Mistral ###
 mistral_7b = Model(
@@ -305,14 +317,14 @@ mistral_large = Model(
 
 
 ### NousResearch ###
-mixtral_8x7b_dpo = Model(
-    name          = "mixtral-8x7b-dpo",
+hermes_2 = Model(
+    name          = "hermes-2",
     base_provider = "NousResearch",
     best_provider = Airforce
 )
 
-yi_34b = Model(
-    name          = "yi-34b",
+hermes_2_dpo = Model(
+    name          = "hermes-2-dpo",
     base_provider = "NousResearch",
     best_provider = Airforce
 )
@@ -328,7 +340,7 @@ hermes_3 = Model(
 phi_2 = Model(
     name          = "phi-2",
     base_provider = "Microsoft",
-    best_provider = Cloudflare
+    best_provider = IterListProvider([Cloudflare, Airforce])
 )
 
 phi_3_medium_4k = Model(
@@ -364,10 +376,10 @@ gemini = Model(
 )
 
 # gemma
-gemma_2b_9b = Model(
-    name          = 'gemma-2b-9b',
+gemma_2b = Model(
+    name          = 'gemma-2b',
     base_provider = 'Google',
-    best_provider = Airforce
+    best_provider = IterListProvider([ReplicateHome, Airforce])
 )
 
 gemma_2b_27b = Model(
@@ -376,12 +388,6 @@ gemma_2b_27b = Model(
     best_provider = IterListProvider([DeepInfraChat, Airforce])
 )
 
-gemma_2b = Model(
-    name          = 'gemma-2b',
-    base_provider = 'Google',
-    best_provider = IterListProvider([ReplicateHome, Airforce])
-)
-
 gemma_7b = Model(
     name          = 'gemma-7b',
     base_provider = 'Google',
@@ -389,18 +395,18 @@ gemma_7b = Model(
 )
 
 # gemma 2
-gemma_2_27b = Model(
-    name          = 'gemma-2-27b',
-    base_provider = 'Google',
-    best_provider = Airforce
-)
-
 gemma_2 = Model(
     name          = 'gemma-2',
     base_provider = 'Google',
     best_provider = ChatHub
 )
 
+gemma_2_9b = Model(
+    name          = 'gemma-2-9b',
+    base_provider = 'Google',
+    best_provider = Airforce
+)
+
 
 ### Anthropic ###
 claude_2_1 = Model(
@@ -413,26 +419,26 @@ claude_2_1 = Model(
 claude_3_opus = Model(
     name          = 'claude-3-opus',
     base_provider = 'Anthropic',
-    best_provider = IterListProvider([Airforce, Liaobots])
+    best_provider = IterListProvider([Liaobots])
 )
 
 claude_3_sonnet = Model(
     name          = 'claude-3-sonnet',
     base_provider = 'Anthropic',
-    best_provider = IterListProvider([Airforce, Liaobots])
+    best_provider = IterListProvider([Liaobots])
 )
 
 claude_3_haiku = Model(
     name          = 'claude-3-haiku',
     base_provider = 'Anthropic',
-    best_provider = IterListProvider([DDG, Airforce, GizAI, Liaobots])
+    best_provider = IterListProvider([DDG, GizAI, Liaobots])
 )
 
 # claude 3.5
 claude_3_5_sonnet = Model(
     name          = 'claude-3.5-sonnet',
     base_provider = 'Anthropic',
-    best_provider = IterListProvider([Blackbox, Editee, AmigoChat, Airforce, GizAI, Liaobots])
+    best_provider = IterListProvider([Blackbox, Editee, AmigoChat, GizAI, Liaobots])
 )
 
 
@@ -493,40 +499,34 @@ qwen_1_5_0_5b = Model(
 qwen_1_5_7b = Model(
     name = 'qwen-1.5-7b',
     base_provider = 'Qwen',
-    best_provider = IterListProvider([Cloudflare, Airforce])
+    best_provider = IterListProvider([Cloudflare])
 )
 
 qwen_1_5_14b = Model(
     name = 'qwen-1.5-14b',
     base_provider = 'Qwen',
-    best_provider = IterListProvider([FreeChatgpt, Cloudflare, Airforce])
+    best_provider = IterListProvider([FreeChatgpt, Cloudflare])
 )
 
-qwen_1_5_72b = Model(
-    name = 'qwen-1.5-72b',
+# qwen 2
+qwen_2_72b = Model(
+    name = 'qwen-2-72b',
     base_provider = 'Qwen',
-    best_provider = Airforce
+    best_provider = IterListProvider([DeepInfraChat, HuggingChat, Airforce, HuggingFace])
 )
 
-qwen_1_5_110b = Model(
-    name = 'qwen-1.5-110b',
+qwen_2_5_7b = Model(
+    name = 'qwen-2-5-7b',
     base_provider = 'Qwen',
     best_provider = Airforce
 )
 
-qwen_1_5_1_8b = Model(
-    name = 'qwen-1.5-1.8b',
+qwen_2_5_72b = Model(
+    name = 'qwen-2-5-72b',
     base_provider = 'Qwen',
     best_provider = Airforce
 )
 
-# qwen 2
-qwen_2_72b = Model(
-    name = 'qwen-2-72b',
-    base_provider = 'Qwen',
-    best_provider = IterListProvider([DeepInfraChat, HuggingChat, Airforce, HuggingFace])
-)
-
 qwen = Model(
     name = 'qwen',
     base_provider = 'Qwen',
@@ -556,18 +556,18 @@ yi_1_5_9b = Model(
 )
 
 ### Upstage ###
-solar_1_mini = Model(
-    name = 'solar-1-mini',
-    base_provider = 'Upstage',
-    best_provider = Upstage
-)
-
 solar_10_7b = Model(
     name = 'solar-10-7b',
     base_provider = 'Upstage',
     best_provider = Airforce
 )
 
+solar_mini = Model(
+    name = 'solar-mini',
+    base_provider = 'Upstage',
+    best_provider = Upstage
+)
+
 solar_pro = Model(
     name = 'solar-pro',
     base_provider = 'Upstage',
@@ -583,8 +583,8 @@ pi = Model(
 )
 
 ### DeepSeek ###
-deepseek = Model(
-    name = 'deepseek',
+deepseek_coder = Model(
+    name = 'deepseek-coder',
     base_provider = 'DeepSeek',
     best_provider = Airforce
 )
@@ -630,7 +630,7 @@ lzlv_70b = Model(
 openchat_3_5 = Model(
     name = 'openchat-3.5',
     base_provider = 'OpenChat',
-    best_provider = Cloudflare
+    best_provider = IterListProvider([Cloudflare])
 )
 
 openchat_3_6_8b = Model(
@@ -683,23 +683,6 @@ sonar_chat = Model(
     best_provider = PerplexityLabs
 )
 
-
-### Gryphe ### 
-mythomax_l2_13b = Model(
-    name = 'mythomax-l2-13b',
-    base_provider = 'Gryphe',
-    best_provider = Airforce
-)
-
-
-### Pawan ### 
-cosmosrp = Model(
-    name = 'cosmosrp',
-    base_provider = 'Pawan',
-    best_provider = Airforce
-)
-
-
 ### TheBloke ### 
 german_7b = Model(
     name = 'german-7b',
@@ -708,14 +691,6 @@ german_7b = Model(
 )
 
 
-### Tinyllama ### 
-tinyllama_1_1b = Model(
-    name = 'tinyllama-1.1b',
-    base_provider = 'Tinyllama',
-    best_provider = Cloudflare
-)
-
-
 ### Fblgit ### 
 cybertron_7b = Model(
     name = 'cybertron-7b',
@@ -723,6 +698,7 @@ cybertron_7b = Model(
     best_provider = Cloudflare
 )
 
+
 ### Nvidia ### 
 nemotron_70b = Model(
     name = 'nemotron-70b',
@@ -731,6 +707,46 @@ nemotron_70b = Model(
 )
 
 
+### Teknium ### 
+openhermes_2_5 = Model(
+    name = 'openhermes-2.5',
+    base_provider = 'Teknium',
+    best_provider = Airforce
+)
+
+
+### Pawan ### 
+cosmosrp = Model(
+    name = 'cosmosrp',
+    base_provider = 'Pawan',
+    best_provider = Airforce
+)
+
+
+### Liquid ### 
+lfm_40b = Model(
+    name = 'lfm-40b',
+    base_provider = 'Liquid',
+    best_provider = Airforce
+)
+
+
+### DiscoResearch ### 
+german_7b = Model(
+    name = 'german-7b',
+    base_provider = 'DiscoResearch',
+    best_provider = Airforce
+)
+
+
+### HuggingFaceH4 ### 
+zephyr_7b = Model(
+    name = 'zephyr-7b',
+    base_provider = 'HuggingFaceH4',
+    best_provider = Airforce
+)
+
+
 
 #############
 ### Image ###
@@ -754,7 +770,7 @@ sdxl_lora = Model(
 sdxl = Model(
     name = 'sdxl',
     base_provider = 'Stability AI',
-    best_provider = IterListProvider([ReplicateHome])
+    best_provider = IterListProvider([ReplicateHome, Airforce])
     
 )
 
@@ -947,6 +963,8 @@ class ModelUtils:
 # llamaguard
 'llamaguard-7b': llamaguard_7b,
 'llamaguard-2-8b': llamaguard_2_8b,
+'llamaguard-3-8b': llamaguard_3_8b,
+'llamaguard-3-11b': llamaguard_3_11b,
       
         
 ### Mistral ###
@@ -958,17 +976,17 @@ class ModelUtils:
      
      
 ### NousResearch ###
-'mixtral-8x7b-dpo': mixtral_8x7b_dpo,
+'hermes-2': hermes_2,
+'hermes-2-dpo': hermes_2_dpo,
 'hermes-3': hermes_3,
- 
-'yi-34b': yi_34b,   
-        
-        
+
+                
 ### Microsoft ###
 'phi-2': phi_2,
 'phi_3_medium-4k': phi_3_medium_4k,
 'phi-3.5-mini': phi_3_5_mini,
 
+
 ### Google ###
 # gemini
 'gemini': gemini,
@@ -977,13 +995,12 @@ class ModelUtils:
         
 # gemma
 'gemma-2b': gemma_2b,
-'gemma-2b-9b': gemma_2b_9b,
 'gemma-2b-27b': gemma_2b_27b,
 'gemma-7b': gemma_7b,
 
 # gemma-2
 'gemma-2': gemma_2,
-'gemma-2-27b': gemma_2_27b,
+'gemma-2-9b': gemma_2_9b,
 
 
 ### Anthropic ###
@@ -1028,10 +1045,9 @@ class ModelUtils:
 'qwen-1.5-0.5b': qwen_1_5_0_5b,
 'qwen-1.5-7b': qwen_1_5_7b,
 'qwen-1.5-14b': qwen_1_5_14b,
-'qwen-1.5-72b': qwen_1_5_72b,
-'qwen-1.5-110b': qwen_1_5_110b,
-'qwen-1.5-1.8b': qwen_1_5_1_8b,
 'qwen-2-72b': qwen_2_72b,
+'qwen-2-5-7b': qwen_2_5_7b,
+'qwen-2-5-72b': qwen_2_5_72b,
         
         
 ### Zhipu AI ###
@@ -1044,16 +1060,17 @@ class ModelUtils:
         
         
 ### Upstage ###
-'solar-mini': solar_1_mini,
 'solar-10-7b': solar_10_7b,
+'solar-mini': solar_mini,
 'solar-pro': solar_pro,
 
 
 ### Inflection ###
 'pi': pi,
 
+
 ### DeepSeek ###
-'deepseek': deepseek,
+'deepseek-coder': deepseek_coder,
      
         
 ### Yorickvp ###
@@ -1094,30 +1111,38 @@ class ModelUtils:
 ### Perplexity AI ###
 'sonar-online': sonar_online,
 'sonar-chat': sonar_chat,
-
-
-### Gryphe ###   
-'mythomax-l2-13b': sonar_chat,
-
-   
-### Pawan ###   
-'cosmosrp': cosmosrp,
-        
+     
         
 ### TheBloke ###   
 'german-7b': german_7b,
 
 
-### Tinyllama ###   
-'tinyllama-1.1b': tinyllama_1_1b,
-
-
 ### Fblgit ###   
 'cybertron-7b': cybertron_7b,
         
         
 ### Nvidia ###   
 'nemotron-70b': nemotron_70b,
+
+
+### Teknium ###   
+'openhermes-2.5': openhermes_2_5,
+        
+        
+### Pawan ###   
+'cosmosrp': cosmosrp,
+
+
+### Liquid ###   
+'lfm-40b': lfm_40b,
+      
+        
+### DiscoResearch ###   
+'german-7b': german_7b,
+
+
+### HuggingFaceH4 ###   
+'zephyr-7b': zephyr_7b,
         
         
         
-- 
cgit v1.2.3


From 42aba60eceb0a022e5dc79ddf6f391163dbb2cf5 Mon Sep 17 00:00:00 2001
From: kqlio67 <kqlio67@users.noreply.github.com>
Date: Fri, 1 Nov 2024 22:53:15 +0200
Subject: Update (g4f/Provider/AIUncensored.py)

---
 g4f/models.py | 4 +++-
 1 file changed, 3 insertions(+), 1 deletion(-)

(limited to 'g4f/models.py')

diff --git a/g4f/models.py b/g4f/models.py
index ae3d430d..dd2b827d 100644
--- a/g4f/models.py
+++ b/g4f/models.py
@@ -8,6 +8,7 @@ from .Provider import (
     AIChatFree,
     AiMathGPT,
     Airforce,
+    AIUncensored,
     Allyfy,
     AmigoChat,
     Bing,
@@ -109,6 +110,7 @@ default = Model(
         Cloudflare,
         Editee,
         AiMathGPT,
+        AIUncensored,
     ])
 )
 
@@ -808,7 +810,7 @@ playground_v2_5 = Model(
 flux = Model(
     name = 'flux',
     base_provider = 'Flux AI',
-    best_provider = IterListProvider([Blackbox, Airforce])
+    best_provider = IterListProvider([Blackbox, AIUncensored, Airforce])
     
 )
 
-- 
cgit v1.2.3


From e1b131290217308026e3bb321d90e9e4a54b0693 Mon Sep 17 00:00:00 2001
From: kqlio67 <kqlio67@users.noreply.github.com>
Date: Sat, 2 Nov 2024 13:50:28 +0200
Subject: Disconnected provider (g4f/Provider/AmigoChat.py)

---
 g4f/models.py | 22 ++++++++++------------
 1 file changed, 10 insertions(+), 12 deletions(-)

(limited to 'g4f/models.py')

diff --git a/g4f/models.py b/g4f/models.py
index dd2b827d..944c4e9c 100644
--- a/g4f/models.py
+++ b/g4f/models.py
@@ -10,7 +10,6 @@ from .Provider import (
     Airforce,
     AIUncensored,
     Allyfy,
-    AmigoChat,
     Bing,
     Blackbox,
     ChatGpt,
@@ -105,7 +104,6 @@ default = Model(
         ChatHub,
         ChatGptEs,
         ChatHub,
-        AmigoChat,
         ChatifyAI,
         Cloudflare,
         Editee,
@@ -139,13 +137,13 @@ gpt_35_turbo = Model(
 gpt_4o = Model(
     name          = 'gpt-4o',
     base_provider = 'OpenAI',
-    best_provider = IterListProvider([NexraChatGPT, Blackbox, ChatGptEs, AmigoChat, DarkAI, Editee, GizAI, Airforce, Liaobots, OpenaiChat])
+    best_provider = IterListProvider([NexraChatGPT, Blackbox, ChatGptEs, DarkAI, Editee, GizAI, Airforce, Liaobots, OpenaiChat])
 )
 
 gpt_4o_mini = Model(
     name          = 'gpt-4o-mini',
     base_provider = 'OpenAI',
-    best_provider = IterListProvider([DDG, ChatGptEs, FreeNetfly, Pizzagpt, MagickPen, AmigoChat, RubiksAI, Liaobots, Airforce, GizAI, ChatgptFree, Koala, OpenaiChat, ChatGpt])
+    best_provider = IterListProvider([DDG, ChatGptEs, FreeNetfly, Pizzagpt, MagickPen, RubiksAI, Liaobots, Airforce, GizAI, ChatgptFree, Koala, OpenaiChat, ChatGpt])
 )
 
 gpt_4_turbo = Model(
@@ -164,13 +162,13 @@ gpt_4 = Model(
 o1 = Model(
     name          = 'o1',
     base_provider = 'OpenAI',
-    best_provider = AmigoChat
+    best_provider = None
 )
 
 o1_mini = Model(
     name          = 'o1-mini',
     base_provider = 'OpenAI',
-    best_provider = IterListProvider([AmigoChat, GizAI])
+    best_provider = IterListProvider([GizAI])
 )
 
 
@@ -231,7 +229,7 @@ llama_3_1_70b = Model(
 llama_3_1_405b = Model(
     name          = "llama-3.1-405b",
     base_provider = "Meta Llama",
-    best_provider = IterListProvider([DeepInfraChat, Blackbox, AmigoChat, DarkAI, Airforce])
+    best_provider = IterListProvider([DeepInfraChat, Blackbox, DarkAI, Airforce])
 )
 
 # llama 3.2
@@ -256,7 +254,7 @@ llama_3_2_11b = Model(
 llama_3_2_90b = Model(
     name          = "llama-3.2-90b",
     base_provider = "Meta Llama",
-    best_provider = IterListProvider([AmigoChat, Airforce])
+    best_provider = IterListProvider([Airforce])
 )
 
 
@@ -362,7 +360,7 @@ phi_3_5_mini = Model(
 gemini_pro = Model(
     name          = 'gemini-pro',
     base_provider = 'Google DeepMind',
-    best_provider = IterListProvider([GeminiPro, Blackbox, AIChatFree, GPROChat, NexraGeminiPro, AmigoChat, Editee, GizAI, Airforce, Liaobots])
+    best_provider = IterListProvider([GeminiPro, Blackbox, AIChatFree, GPROChat, NexraGeminiPro, Editee, GizAI, Airforce, Liaobots])
 )
 
 gemini_flash = Model(
@@ -440,7 +438,7 @@ claude_3_haiku = Model(
 claude_3_5_sonnet = Model(
     name          = 'claude-3.5-sonnet',
     base_provider = 'Anthropic',
-    best_provider = IterListProvider([Blackbox, Editee, AmigoChat, GizAI, Liaobots])
+    best_provider = IterListProvider([Blackbox, Editee, GizAI, Liaobots])
 )
 
 
@@ -817,14 +815,14 @@ flux = Model(
 flux_pro = Model(
     name = 'flux-pro',
     base_provider = 'Flux AI',
-    best_provider = IterListProvider([NexraFluxPro, AmigoChat])
+    best_provider = IterListProvider([NexraFluxPro])
     
 )
 
 flux_realism = Model(
     name = 'flux-realism',
     base_provider = 'Flux AI',
-    best_provider = IterListProvider([Airforce, AmigoChat])
+    best_provider = IterListProvider([Airforce])
     
 )
 
-- 
cgit v1.2.3