Update app.py
Browse files
app.py
CHANGED
@@ -4,6 +4,7 @@ import os
|
|
4 |
import logging
|
5 |
|
6 |
from langchain_ollama import ChatOllama
|
|
|
7 |
|
8 |
from langchain_community.document_loaders import PyPDFLoader
|
9 |
|
@@ -142,6 +143,9 @@ def main() -> None:
|
|
142 |
## from llama3.2:latest to unsloth/Llama-3.2-3B
|
143 |
model = ChatOllama(model="unsloth/Llama-3.2-3B")
|
144 |
|
|
|
|
|
|
|
145 |
rag_chain = (
|
146 |
{"context": retriever|format_docs, "question": RunnablePassthrough()}
|
147 |
| prompt
|
@@ -160,10 +164,7 @@ def main() -> None:
|
|
160 |
if submit and text:
|
161 |
with st.spinner("Generating response..."):
|
162 |
# Ken (12/11/2024): modify start
|
163 |
-
|
164 |
-
##retriever = configure_retriever(pdf_loader)
|
165 |
-
##chain = create_chain(retriever, prompt, model)
|
166 |
-
response = inference(rag_chain, text)
|
167 |
# Ken (12/11/2024): modify end
|
168 |
st.session_state['chat_history'].append({"user": text, "ollama": response})
|
169 |
st.write(response)
|
|
|
4 |
import logging
|
5 |
|
6 |
from langchain_ollama import ChatOllama
|
7 |
+
from langchain_community.llms import Ollama
|
8 |
|
9 |
from langchain_community.document_loaders import PyPDFLoader
|
10 |
|
|
|
143 |
## from llama3.2:latest to unsloth/Llama-3.2-3B
|
144 |
model = ChatOllama(model="unsloth/Llama-3.2-3B")
|
145 |
|
146 |
+
test_llm = model
|
147 |
+
test_llm.invoke("Why is the sky blue?")
|
148 |
+
|
149 |
rag_chain = (
|
150 |
{"context": retriever|format_docs, "question": RunnablePassthrough()}
|
151 |
| prompt
|
|
|
164 |
if submit and text:
|
165 |
with st.spinner("Generating response..."):
|
166 |
# Ken (12/11/2024): modify start
|
167 |
+
response = generate_response(rag_chain, text)
|
|
|
|
|
|
|
168 |
# Ken (12/11/2024): modify end
|
169 |
st.session_state['chat_history'].append({"user": text, "ollama": response})
|
170 |
st.write(response)
|