|
import os |
|
|
|
import together_gradio |
|
|
|
from utils import get_app |
|
|
|
demo = get_app( |
|
models=[ |
|
"meta-llama/Llama-Vision-Free", |
|
"meta-llama/Llama-3.2-11B-Vision-Instruct-Turbo", |
|
"meta-llama/Llama-3.2-90B-Vision-Instruct-Turbo", |
|
"meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo", |
|
"meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo", |
|
"meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo", |
|
"meta-llama/Meta-Llama-3-8B-Instruct-Turbo", |
|
"meta-llama/Meta-Llama-3-70B-Instruct-Turbo", |
|
"meta-llama/Llama-3.2-3B-Instruct-Turbo", |
|
"meta-llama/Meta-Llama-3-8B-Instruct-Lite", |
|
"meta-llama/Meta-Llama-3-70B-Instruct-Lite", |
|
"meta-llama/Llama-3-8b-chat-hf", |
|
"meta-llama/Llama-3-70b-chat-hf", |
|
"nvidia/Llama-3.1-Nemotron-70B-Instruct-HF", |
|
"Qwen/Qwen2.5-Coder-32B-Instruct", |
|
"microsoft/WizardLM-2-8x22B", |
|
"google/gemma-2-27b-it", |
|
"google/gemma-2-9b-it", |
|
"databricks/dbrx-instruct", |
|
"mistralai/Mixtral-8x7B-Instruct-v0.1", |
|
"mistralai/Mixtral-8x22B-Instruct-v0.1", |
|
"Qwen/Qwen2.5-7B-Instruct-Turbo", |
|
"Qwen/Qwen2.5-72B-Instruct-Turbo", |
|
"Qwen/Qwen2-72B-Instruct", |
|
"deepseek-ai/deepseek-llm-67b-chat", |
|
"google/gemma-2b-it", |
|
"Gryphe/MythoMax-L2-13b", |
|
"meta-llama/Llama-2-13b-chat-hf", |
|
"mistralai/Mistral-7B-Instruct-v0.1", |
|
"mistralai/Mistral-7B-Instruct-v0.2", |
|
"mistralai/Mistral-7B-Instruct-v0.3", |
|
"NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO", |
|
"togethercomputer/StripedHyena-Nous-7B", |
|
"upstage/SOLAR-10.7B-Instruct-v1.0", |
|
], |
|
default_model="meta-llama/Llama-3.2-11B-Vision-Instruct-Turbo", |
|
src=together_gradio.registry, |
|
accept_token=not os.getenv("TOGETHER_API_KEY"), |
|
multimodal=True, |
|
) |
|
|
|
if __name__ == "__main__": |
|
demo.launch() |
|
|