models.py 1.9 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748
  1. default_model = "Qwen/Qwen2.5-72B-Instruct"
  2. default_image_model = "black-forest-labs/FLUX.1-dev"
  3. image_models = [
  4. default_image_model,
  5. "black-forest-labs/FLUX.1-schnell",
  6. ]
  7. fallback_models = [
  8. default_model,
  9. 'meta-llama/Llama-3.3-70B-Instruct',
  10. 'CohereForAI/c4ai-command-r-plus-08-2024',
  11. 'deepseek-ai/DeepSeek-R1-Distill-Qwen-32B',
  12. 'Qwen/QwQ-32B-Preview',
  13. 'nvidia/Llama-3.1-Nemotron-70B-Instruct-HF',
  14. 'Qwen/Qwen2.5-Coder-32B-Instruct',
  15. 'meta-llama/Llama-3.2-11B-Vision-Instruct',
  16. 'mistralai/Mistral-Nemo-Instruct-2407',
  17. 'microsoft/Phi-3.5-mini-instruct',
  18. ] + image_models
  19. model_aliases = {
  20. ### Chat ###
  21. "qwen-2.5-72b": "Qwen/Qwen2.5-Coder-32B-Instruct",
  22. "llama-3.3-70b": "meta-llama/Llama-3.3-70B-Instruct",
  23. "command-r-plus": "CohereForAI/c4ai-command-r-plus-08-2024",
  24. "deepseek-r1": "deepseek-ai/DeepSeek-R1-Distill-Qwen-32B",
  25. "qwq-32b": "Qwen/QwQ-32B-Preview",
  26. "nemotron-70b": "nvidia/Llama-3.1-Nemotron-70B-Instruct-HF",
  27. "qwen-2.5-coder-32b": "Qwen/Qwen2.5-Coder-32B-Instruct",
  28. "llama-3.2-11b": "meta-llama/Llama-3.2-11B-Vision-Instruct",
  29. "mistral-nemo": "mistralai/Mistral-Nemo-Instruct-2407",
  30. "phi-3.5-mini": "microsoft/Phi-3.5-mini-instruct",
  31. ### Image ###
  32. "flux": "black-forest-labs/FLUX.1-dev",
  33. "flux-dev": "black-forest-labs/FLUX.1-dev",
  34. "flux-schnell": "black-forest-labs/FLUX.1-schnell",
  35. ### Used in other providers ###
  36. "qwen-2-vl-7b": "Qwen/Qwen2-VL-7B-Instruct",
  37. "gemma-2-27b": "google/gemma-2-27b-it",
  38. "qwen-2-72b": "Qwen/Qwen2-72B-Instruct",
  39. "qvq-72b": "Qwen/QVQ-72B-Preview",
  40. "sd-3.5": "stabilityai/stable-diffusion-3.5-large",
  41. }
  42. extra_models = [
  43. "meta-llama/Llama-3.2-11B-Vision-Instruct",
  44. "nvidia/Llama-3.1-Nemotron-70B-Instruct-HF",
  45. "NousResearch/Hermes-3-Llama-3.1-8B",
  46. ]
  47. default_vision_model = "meta-llama/Llama-3.2-11B-Vision-Instruct"
  48. vision_models = [default_vision_model, "Qwen/Qwen2-VL-7B-Instruct"]