Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -1381,11 +1381,11 @@ if filename is not None:
|
|
1381 |
session_state = {}
|
1382 |
if "search_queries" not in session_state:
|
1383 |
session_state["search_queries"] = []
|
|
|
1384 |
example_input = st.text_input("AI Search ArXiV Scholarly Articles", value=session_state["search_queries"][-1] if session_state["search_queries"] else "")
|
|
|
1385 |
if example_input:
|
1386 |
session_state["search_queries"].append(example_input)
|
1387 |
-
|
1388 |
-
# Search AI
|
1389 |
query=example_input
|
1390 |
if query:
|
1391 |
result = search_arxiv(query)
|
@@ -1462,27 +1462,14 @@ client = OpenAI(api_key= os.getenv('OPENAI_API_KEY'), organization=os.getenv('OP
|
|
1462 |
#MODEL = "gpt-4o"
|
1463 |
MODEL = "gpt-4o-2024-05-13"
|
1464 |
|
1465 |
-
def
|
1466 |
-
text_input = st.text_input("Enter your text:")
|
1467 |
-
if text_input:
|
1468 |
-
completion = client.chat.completions.create(
|
1469 |
-
model=MODEL,
|
1470 |
-
messages=[
|
1471 |
-
{"role": "system", "content": "You are a helpful assistant. Help me with my math homework!"},
|
1472 |
-
{"role": "user", "content": f"Hello! Could you solve {text_input}?"}
|
1473 |
-
]
|
1474 |
-
)
|
1475 |
-
st.write("Assistant: " + completion.choices[0].message.content)
|
1476 |
-
|
1477 |
-
|
1478 |
-
|
1479 |
-
def process_text():
|
1480 |
-
text_input = st.text_input("Enter your text:")
|
1481 |
if text_input:
|
1482 |
|
1483 |
st.session_state.messages.append({"role": "user", "content": text_input})
|
|
|
1484 |
with st.chat_message("user"):
|
1485 |
st.markdown(text_input)
|
|
|
1486 |
with st.chat_message("assistant"):
|
1487 |
completion = client.chat.completions.create(
|
1488 |
model=MODEL,
|
@@ -1490,14 +1477,13 @@ def process_text():
|
|
1490 |
{"role": m["role"], "content": m["content"]}
|
1491 |
for m in st.session_state.messages
|
1492 |
],
|
1493 |
-
stream=
|
1494 |
)
|
1495 |
return_text = completion.choices[0].message.content
|
1496 |
st.write("Assistant: " + return_text)
|
1497 |
filename = generate_filename(text_input, "md")
|
1498 |
create_file(filename, text_input, return_text, should_save)
|
1499 |
-
|
1500 |
-
st.session_state.messages.append({"role": "assistant", "content": response})
|
1501 |
|
1502 |
#st.write("Assistant: " + completion.choices[0].message.content)
|
1503 |
|
@@ -1673,8 +1659,12 @@ def process_audio_and_video(video_input):
|
|
1673 |
],
|
1674 |
temperature=0,
|
1675 |
)
|
1676 |
-
|
1677 |
-
st.markdown(
|
|
|
|
|
|
|
|
|
1678 |
|
1679 |
|
1680 |
def main():
|
@@ -1682,7 +1672,10 @@ def main():
|
|
1682 |
st.markdown("##### GPT-4o Omni Model: Text, Audio, Image, & Video")
|
1683 |
option = st.selectbox("Select an option", ("Text", "Image", "Audio", "Video"))
|
1684 |
if option == "Text":
|
1685 |
-
|
|
|
|
|
|
|
1686 |
elif option == "Image":
|
1687 |
image_input = st.file_uploader("Upload an image", type=["jpg", "jpeg", "png"])
|
1688 |
image_response = process_image(image_input)
|
|
|
1381 |
session_state = {}
|
1382 |
if "search_queries" not in session_state:
|
1383 |
session_state["search_queries"] = []
|
1384 |
+
|
1385 |
example_input = st.text_input("AI Search ArXiV Scholarly Articles", value=session_state["search_queries"][-1] if session_state["search_queries"] else "")
|
1386 |
+
|
1387 |
if example_input:
|
1388 |
session_state["search_queries"].append(example_input)
|
|
|
|
|
1389 |
query=example_input
|
1390 |
if query:
|
1391 |
result = search_arxiv(query)
|
|
|
1462 |
#MODEL = "gpt-4o"
|
1463 |
MODEL = "gpt-4o-2024-05-13"
|
1464 |
|
1465 |
+
def process_text(text_input):
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1466 |
if text_input:
|
1467 |
|
1468 |
st.session_state.messages.append({"role": "user", "content": text_input})
|
1469 |
+
|
1470 |
with st.chat_message("user"):
|
1471 |
st.markdown(text_input)
|
1472 |
+
|
1473 |
with st.chat_message("assistant"):
|
1474 |
completion = client.chat.completions.create(
|
1475 |
model=MODEL,
|
|
|
1477 |
{"role": m["role"], "content": m["content"]}
|
1478 |
for m in st.session_state.messages
|
1479 |
],
|
1480 |
+
stream=False
|
1481 |
)
|
1482 |
return_text = completion.choices[0].message.content
|
1483 |
st.write("Assistant: " + return_text)
|
1484 |
filename = generate_filename(text_input, "md")
|
1485 |
create_file(filename, text_input, return_text, should_save)
|
1486 |
+
st.session_state.messages.append({"role": "assistant", "content": return_text})
|
|
|
1487 |
|
1488 |
#st.write("Assistant: " + completion.choices[0].message.content)
|
1489 |
|
|
|
1659 |
],
|
1660 |
temperature=0,
|
1661 |
)
|
1662 |
+
results = response.choices[0].message.content
|
1663 |
+
st.markdown(results)
|
1664 |
+
|
1665 |
+
filename = generate_filename(transcript, "md")
|
1666 |
+
create_file(filename, transcript, results, should_save)
|
1667 |
+
|
1668 |
|
1669 |
|
1670 |
def main():
|
|
|
1672 |
st.markdown("##### GPT-4o Omni Model: Text, Audio, Image, & Video")
|
1673 |
option = st.selectbox("Select an option", ("Text", "Image", "Audio", "Video"))
|
1674 |
if option == "Text":
|
1675 |
+
text_input = st.text_input("Enter your text:")
|
1676 |
+
if (text_input > ''):
|
1677 |
+
textResponse = process_text(text_input)
|
1678 |
+
|
1679 |
elif option == "Image":
|
1680 |
image_input = st.file_uploader("Upload an image", type=["jpg", "jpeg", "png"])
|
1681 |
image_response = process_image(image_input)
|