Spaces:
Running
Running
| import os | |
| import together_gradio | |
| from utils import get_app | |
| demo = get_app( | |
| models=[ | |
| "meta-llama/Llama-Vision-Free", | |
| "meta-llama/Llama-3.2-11B-Vision-Instruct-Turbo", | |
| "meta-llama/Llama-3.2-90B-Vision-Instruct-Turbo", | |
| "meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo", | |
| "meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo", | |
| "meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo", | |
| "meta-llama/Meta-Llama-3-8B-Instruct-Turbo", | |
| "meta-llama/Meta-Llama-3-70B-Instruct-Turbo", | |
| "meta-llama/Llama-3.2-3B-Instruct-Turbo", | |
| "meta-llama/Meta-Llama-3-8B-Instruct-Lite", | |
| "meta-llama/Meta-Llama-3-70B-Instruct-Lite", | |
| "meta-llama/Llama-3-8b-chat-hf", | |
| "meta-llama/Llama-3-70b-chat-hf", | |
| "nvidia/Llama-3.1-Nemotron-70B-Instruct-HF", | |
| "Qwen/Qwen2.5-Coder-32B-Instruct", | |
| "microsoft/WizardLM-2-8x22B", | |
| "google/gemma-2-27b-it", | |
| "google/gemma-2-9b-it", | |
| "databricks/dbrx-instruct", | |
| "mistralai/Mixtral-8x7B-Instruct-v0.1", | |
| "mistralai/Mixtral-8x22B-Instruct-v0.1", | |
| "Qwen/Qwen2.5-7B-Instruct-Turbo", | |
| "Qwen/Qwen2.5-72B-Instruct-Turbo", | |
| "Qwen/Qwen2-72B-Instruct", | |
| "deepseek-ai/deepseek-llm-67b-chat", | |
| "google/gemma-2b-it", | |
| "Gryphe/MythoMax-L2-13b", | |
| "meta-llama/Llama-2-13b-chat-hf", | |
| "mistralai/Mistral-7B-Instruct-v0.1", | |
| "mistralai/Mistral-7B-Instruct-v0.2", | |
| "mistralai/Mistral-7B-Instruct-v0.3", | |
| "NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO", | |
| "togethercomputer/StripedHyena-Nous-7B", | |
| "upstage/SOLAR-10.7B-Instruct-v1.0", | |
| ], | |
| default_model="meta-llama/Llama-3.2-11B-Vision-Instruct-Turbo", | |
| src=together_gradio.registry, | |
| accept_token=not os.getenv("TOGETHER_API_KEY"), | |
| multimodal=True, | |
| ) | |
| if __name__ == "__main__": | |
| demo.launch() | |