Spaces:
Sleeping
Sleeping
Added comment
Browse files
app.py
CHANGED
@@ -19,7 +19,8 @@ DEFAULT_MAX_NEW_TOKENS = 128
|
|
19 |
MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "4096"))
|
20 |
|
21 |
if torch.cuda.is_available():
|
22 |
-
model_id = "
|
|
|
23 |
model = AutoModelForCausalLM.from_pretrained(model_id)
|
24 |
tokenizer = AutoTokenizer.from_pretrained(model_id)
|
25 |
tokenizer.pad_token_id = tokenizer.eos_token_id
|
|
|
19 |
MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "4096"))
|
20 |
|
21 |
if torch.cuda.is_available():
|
22 |
+
model_id = "scb10x/typhoon-7b" # original model
|
23 |
+
model_id = "bandhit/typhoon-7b-q4-bnb_cuda-ts-1703352224" # quantized model
|
24 |
model = AutoModelForCausalLM.from_pretrained(model_id)
|
25 |
tokenizer = AutoTokenizer.from_pretrained(model_id)
|
26 |
tokenizer.pad_token_id = tokenizer.eos_token_id
|