summaryrefslogblamecommitdiffstats
path: root/tool/readme_table.py
blob: dc7c85f3ec467d8e74d7999abae1343d017592b6 (plain) (tree)
1
2
3
4
5
6
7
8
9
10






                                                  


                                                      
 
























                                                                             






                                                                             























                                                                                                              

                           





















                                                         

                                                             




















                                                                                       




                               
import re
import sys
from pathlib import Path
from urllib.parse import urlparse

sys.path.append(str(Path(__file__).parent.parent))

from g4f import models
from g4f.Provider.base_provider import AsyncProvider
from testing.test_providers import test, get_providers

def print_imports():
    print("##### Providers:")
    print("```py")
    print("from g4f.Provider import (")
    for _provider in get_providers():
        if _provider.working:
            print(f"    {_provider.__name__},")
    print(")")
    print("# Usage:")
    print("response = g4f.ChatCompletion.create(..., provider=ProviderName)")
    print("```")
    print()
    print()

def print_async():
    print("##### Async support:")
    print("```py")
    print("from g4f.Provider import (")
    for _provider in get_providers():
        if issubclass(_provider, AsyncProvider):
            print(f"    {_provider.__name__},")
    print(")")
    print("```")
    print()
    print()

def print_providers():
    lines = [
        "| Website| Provider| gpt-3.5 | gpt-4 | Streaming | Status | Auth |",
        "| ------ | ------- | ------- | ----- | --------- | ------ | ---- |",
    ]
    providers = get_providers()
    for is_working in (True, False):
        for _provider in providers:
            if is_working != _provider.working:
                continue
            netloc = urlparse(_provider.url).netloc
            website = f"[{netloc}]({_provider.url})"

            provider_name = f"g4f.provider.{_provider.__name__}"

            has_gpt_35 = "✔️" if _provider.supports_gpt_35_turbo else "❌"
            has_gpt_4 = "✔️" if _provider.supports_gpt_4 else "❌"
            stream = "✔️" if _provider.supports_stream else "❌"
            if _provider.working:
                if test(_provider):
                    status = '![Active](https://img.shields.io/badge/Active-brightgreen)'
                else:
                    status = '![Unknown](https://img.shields.io/badge/Unknown-grey)'
            else:
                status = '![Inactive](https://img.shields.io/badge/Inactive-red)'
            auth = "✔️" if _provider.needs_auth else "❌"

            lines.append(
                f"| {website} | {provider_name} | {has_gpt_35} | {has_gpt_4} | {stream} | {status} | {auth} |"
            )
    print("\n".join(lines))

def print_models():
    base_provider_names = {
        "cohere": "Cohere",
        "google": "Google",
        "openai": "OpenAI",
        "anthropic": "Anthropic",
        "replicate": "Replicate",
        "huggingface": "Huggingface",
    }
    provider_urls = {
        "Bard": "https://bard.google.com/",
        "H2o": "https://www.h2o.ai/",
        "Vercel": "https://sdk.vercel.ai/",
    }

    lines = [
        "| Model | Base Provider | Provider | Website |",
        "| ----- | ------------- | -------- | ------- |",
    ]

    _models = get_models()
    for model in _models:
        if model.best_provider.__name__ not in provider_urls:
            continue
        split_name = re.split(r":|/", model.name)
        name = split_name[-1]

        base_provider = base_provider_names[model.base_provider]
        provider_name = f"g4f.provider.{model.best_provider.__name__}"

        provider_url = provider_urls[model.best_provider.__name__]
        netloc = urlparse(provider_url).netloc
        website = f"[{netloc}]({provider_url})"
        lines.append(f"| {name} | {base_provider} | {provider_name} | {website} |")

    print("\n".join(lines))


def get_models():
    _models = [item[1] for item in models.__dict__.items()]
    _models = [model for model in _models if type(model) is models.Model]
    return [model for model in _models if model.name not in ["gpt-3.5-turbo", "gpt-4"]]


if __name__ == "__main__":
    print_imports()
    print_async()
    print_providers()
    print("\n", "-" * 50, "\n")
    print_models()