Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -24,7 +24,7 @@ model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=torch.bfloa
|
|
24 |
@torch.inference_mode()
|
25 |
@spaces.GPU
|
26 |
def predict_math_bot(user_message, system_message="", max_new_tokens=125, temperature=0.1, top_p=0.9, repetition_penalty=1.9, do_sample=False):
|
27 |
-
prompt = f"<s><INST>{user_message}{system_message}
|
28 |
inputs = tokenizer(prompt, return_tensors='pt', add_special_tokens=False)
|
29 |
input_ids = inputs["input_ids"].to(model.device)
|
30 |
|
|
|
24 |
@torch.inference_mode()
|
25 |
@spaces.GPU
|
26 |
def predict_math_bot(user_message, system_message="", max_new_tokens=125, temperature=0.1, top_p=0.9, repetition_penalty=1.9, do_sample=False):
|
27 |
+
prompt = f"<s><INST>{user_message}{system_message}<\INST>" if system_message else user_message
|
28 |
inputs = tokenizer(prompt, return_tensors='pt', add_special_tokens=False)
|
29 |
input_ids = inputs["input_ids"].to(model.device)
|
30 |
|