lruizap commited on
Commit
0ce09fc
·
1 Parent(s): 63944e7

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +7 -4
app.py CHANGED
@@ -2,6 +2,9 @@ from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
2
  from transformers import pipeline
3
  import torch
4
  import gradio as gr
 
 
 
5
 
6
  # chatgpt-gpt4-prompts-bart-large-cnn-samsum
7
  tokenizer = AutoTokenizer.from_pretrained(
@@ -10,7 +13,7 @@ model = AutoModelForSeq2SeqLM.from_pretrained(
10
  "Kaludi/chatgpt-gpt4-prompts-bart-large-cnn-samsum", from_tf=True)
11
 
12
  # zephyr
13
- pipe = pipeline("text-generation", model="HuggingFaceH4/zephyr-7b-alpha",torch_dtype=torch.bfloat16, device_map="auto")
14
 
15
  def generate(inputuno, inputdos, max_new_tokens=3556, top_p=0.95, repetition_penalty=1.0):
16
  top_p = float(top_p)
@@ -35,15 +38,15 @@ def generate(inputuno, inputdos, max_new_tokens=3556, top_p=0.95, repetition_pen
35
  "role": "user", "content": str(promptdos)
36
  },
37
  ]
38
-
39
- stream = pipe.tokenizer.apply_chat_template(messages, **generate_kwargs, stream=True, details=True, return_full_text=False)
40
  output = ""
41
 
42
  for response in stream:
43
  output += response.token.text
44
  yield output
45
  return output
46
-
47
 
48
  def generatePrompt(inputuno, inputdos):
49
 
 
2
  from transformers import pipeline
3
  import torch
4
  import gradio as gr
5
+ from huggingface_hub import InferenceClient
6
+
7
+ client = InferenceClient("HuggingFaceH4/zephyr-7b-beta")
8
 
9
  # chatgpt-gpt4-prompts-bart-large-cnn-samsum
10
  tokenizer = AutoTokenizer.from_pretrained(
 
13
  "Kaludi/chatgpt-gpt4-prompts-bart-large-cnn-samsum", from_tf=True)
14
 
15
  # zephyr
16
+ # pipe = pipeline("text-generation", model="HuggingFaceH4/zephyr-7b-alpha",torch_dtype=torch.bfloat16, device_map="auto")
17
 
18
  def generate(inputuno, inputdos, max_new_tokens=3556, top_p=0.95, repetition_penalty=1.0):
19
  top_p = float(top_p)
 
38
  "role": "user", "content": str(promptdos)
39
  },
40
  ]
41
+
42
+ stream = client.text_generation(formatted_prompt, **generate_kwargs, stream=True, details=True, return_full_text=False)
43
  output = ""
44
 
45
  for response in stream:
46
  output += response.token.text
47
  yield output
48
  return output
49
+
50
 
51
  def generatePrompt(inputuno, inputdos):
52