Spaces:
Paused
Paused
Update endpoints.py
Browse files- endpoints.py +2 -1
endpoints.py
CHANGED
@@ -32,11 +32,12 @@ def LLM(llm_name, length):
|
|
32 |
# base_model = AutoModelForCausalLM.from_pretrained("WizardLM/WizardCoder-1B-V1.0")
|
33 |
# Mistral 7B
|
34 |
# mistral_llm = LLM("mistralai/Mistral-7B-v0.1",30000)
|
35 |
-
mistral_llm = LLM("microsoft/phi-2",2000)
|
36 |
|
37 |
# WizardCoder 13B
|
38 |
# wizard_llm = LLM("WizardLM/WizardCoder-Python-13B-V1.0",8000)
|
39 |
wizard_llm = LLM("WizardLM/WizardCoder-3B-V1.0",4000)
|
|
|
40 |
# hf_llm = HuggingFacePipeline(pipeline=pipe)
|
41 |
|
42 |
def ask_model(model, prompt):
|
|
|
32 |
# base_model = AutoModelForCausalLM.from_pretrained("WizardLM/WizardCoder-1B-V1.0")
|
33 |
# Mistral 7B
|
34 |
# mistral_llm = LLM("mistralai/Mistral-7B-v0.1",30000)
|
35 |
+
# mistral_llm = LLM("microsoft/phi-2",2000)
|
36 |
|
37 |
# WizardCoder 13B
|
38 |
# wizard_llm = LLM("WizardLM/WizardCoder-Python-13B-V1.0",8000)
|
39 |
wizard_llm = LLM("WizardLM/WizardCoder-3B-V1.0",4000)
|
40 |
+
mistral_llm = wizard_llm
|
41 |
# hf_llm = HuggingFacePipeline(pipeline=pipe)
|
42 |
|
43 |
def ask_model(model, prompt):
|