123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152 |
- import re
- from urllib.parse import urlparse
- import asyncio
- from g4f import models, ChatCompletion
- from g4f.providers.types import BaseRetryProvider, ProviderType
- from etc.testing._providers import get_providers
- from g4f import debug
- debug.logging = True
- async def test_async(provider: ProviderType):
- if not provider.working:
- return False
- messages = [{"role": "user", "content": "Hello Assistant!"}]
- try:
- if "webdriver" in provider.get_parameters():
- return False
- response = await asyncio.wait_for(ChatCompletion.create_async(
- model=models.default,
- messages=messages,
- provider=provider
- ), 30)
- return bool(response)
- except Exception as e:
- if debug.logging:
- print(f"{provider.__name__}: {e.__class__.__name__}: {e}")
- return False
- def test_async_list(providers: list[ProviderType]):
- responses: list = [
- asyncio.run(test_async(_provider))
- for _provider in providers
- ]
- return responses
- def print_providers():
- providers = get_providers()
- responses = test_async_list(providers)
- for type in ("GPT-4", "GPT-3.5", "Other"):
- lines = [
- "",
- f"### {type}",
- "",
- "| Website | Provider | GPT-3.5 | GPT-4 | Stream | Status | Auth |",
- "| ------ | ------- | ------- | ----- | ------ | ------ | ---- |",
- ]
- for is_working in (True, False):
- for idx, _provider in enumerate(providers):
- if is_working != _provider.working:
- continue
- do_continue = False
- if type == "GPT-4" and _provider.supports_gpt_4:
- do_continue = True
- elif type == "GPT-3.5" and not _provider.supports_gpt_4 and _provider.supports_gpt_35_turbo:
- do_continue = True
- elif type == "Other" and not _provider.supports_gpt_4 and not _provider.supports_gpt_35_turbo:
- do_continue = True
- if not do_continue:
- continue
- netloc = urlparse(_provider.url).netloc.replace("www.", "")
- website = f"[{netloc}]({_provider.url})"
- provider_name = f"`g4f.Provider.{_provider.__name__}`"
- has_gpt_35 = "✔️" if _provider.supports_gpt_35_turbo else "❌"
- has_gpt_4 = "✔️" if _provider.supports_gpt_4 else "❌"
- stream = "✔️" if _provider.supports_stream else "❌"
- if _provider.working:
- status = '![Active](https://img.shields.io/badge/Active-brightgreen)'
- if responses[idx]:
- status = '![Active](https://img.shields.io/badge/Active-brightgreen)'
- else:
- status = '![Unknown](https://img.shields.io/badge/Unknown-grey)'
- else:
- status = '![Inactive](https://img.shields.io/badge/Inactive-red)'
- auth = "✔️" if _provider.needs_auth else "❌"
- lines.append(
- f"| {website} | {provider_name} | {has_gpt_35} | {has_gpt_4} | {stream} | {status} | {auth} |"
- )
- print("\n".join(lines))
- def print_models():
- base_provider_names = {
- "google": "Google",
- "openai": "OpenAI",
- "huggingface": "Huggingface",
- "anthropic": "Anthropic",
- "inflection": "Inflection",
- "meta": "Meta",
- }
- provider_urls = {
- "google": "https://gemini.google.com/",
- "openai": "https://openai.com/",
- "huggingface": "https://huggingface.co/",
- "anthropic": "https://www.anthropic.com/",
- "inflection": "https://inflection.ai/",
- "meta": "https://llama.meta.com/",
- }
- lines = [
- "| Model | Base Provider | Provider | Website |",
- "| ----- | ------------- | -------- | ------- |",
- ]
- for name, model in models.ModelUtils.convert.items():
- if name.startswith("gpt-3.5") or name.startswith("gpt-4"):
- if name not in ("gpt-3.5-turbo", "gpt-4", "gpt-4-turbo"):
- continue
- name = re.split(r":|/", model.name)[-1]
- if model.base_provider not in base_provider_names:
- continue
- base_provider = base_provider_names[model.base_provider]
- if not isinstance(model.best_provider, BaseRetryProvider):
- provider_name = f"g4f.Provider.{model.best_provider.__name__}"
- else:
- provider_name = f"{len(model.best_provider.providers)}+ Providers"
- provider_url = provider_urls[model.base_provider]
- netloc = urlparse(provider_url).netloc.replace("www.", "")
- website = f"[{netloc}]({provider_url})"
- lines.append(f"| {name} | {base_provider} | {provider_name} | {website} |")
- print("\n".join(lines))
- def print_image_models():
- lines = [
- "| Label | Provider | Image Model | Vision Model | Website |",
- "| ----- | -------- | ----------- | ------------ | ------- |",
- ]
- from g4f.gui.server.api import Api
- for image_model in Api.get_image_models():
- provider_url = image_model["url"]
- netloc = urlparse(provider_url).netloc.replace("www.", "")
- website = f"[{netloc}]({provider_url})"
- label = image_model["provider"] if image_model["label"] is None else image_model["label"]
- if image_model["image_model"] is None:
- image_model["image_model"] = "❌"
- if image_model["vision_model"] is None:
- image_model["vision_model"] = "❌"
- lines.append(f'| {label} | `g4f.Provider.{image_model["provider"]}` | {image_model["image_model"]}| {image_model["vision_model"]} | {website} |')
- print("\n".join(lines))
- if __name__ == "__main__":
- #print_providers()
- #print("\n", "-" * 50, "\n")
- #print_models()
- print("\n", "-" * 50, "\n")
- print_image_models()
|