Spaces:
Sleeping
Sleeping
# Load model directly | |
import streamlit as st | |
import transformers | |
st.title("A Simple Interface for a Language Model") | |
st.subheader("Input Text") | |
input_text = st.text_area("Enter your text here", "Type something here...") | |
if st.button("Generate Response"): | |
# Initialize tokenizer and model | |
tokenizer = transformers.AutoTokenizer.from_pretrained("microsoft/phi-2") | |
model = transformers.AutoModelForCausalLM.from_pretrained("microsoft/phi-2") | |
# Encode input text | |
inputs = tokenizer(input_text, return_tensors="pt") | |
# Generate response | |
response = model.generate(**inputs, max_length=100, do_sample=True) | |
# Decode response | |
st.subheader("Generated Response") | |
st.write(tokenizer.decode(response[0])) |