m96tkmok commited on
Commit
1d5ae18
·
verified ·
1 Parent(s): 146ce62

Update app.py

Browse files

Try inference instead of generate_response

Files changed (1) hide show
  1. app.py +11 -2
app.py CHANGED
@@ -9,7 +9,6 @@ from langchain_community.document_loaders import PyPDFLoader
9
 
10
  from langchain_text_splitters import RecursiveCharacterTextSplitter
11
 
12
- #from langchain_ollama import OllamaEmbeddings
13
  from langchain.embeddings import HuggingFaceEmbeddings
14
 
15
  import faiss
@@ -83,6 +82,11 @@ def get_pdf(uploaded_file):
83
  docs = loader.load()
84
  return docs
85
 
 
 
 
 
 
86
 
87
  def main() -> None:
88
 
@@ -155,7 +159,12 @@ def main() -> None:
155
 
156
  if submit and text:
157
  with st.spinner("Generating response..."):
158
- response = generate_response(rag_chain, text)
 
 
 
 
 
159
  st.session_state['chat_history'].append({"user": text, "ollama": response})
160
  st.write(response)
161
 
 
9
 
10
  from langchain_text_splitters import RecursiveCharacterTextSplitter
11
 
 
12
  from langchain.embeddings import HuggingFaceEmbeddings
13
 
14
  import faiss
 
82
  docs = loader.load()
83
  return docs
84
 
85
+ def inference(chain, input_query):
86
+ """Invoke the processing chain with the input query."""
87
+ result = chain.invoke(input_query)
88
+ return result
89
+
90
 
91
  def main() -> None:
92
 
 
159
 
160
  if submit and text:
161
  with st.spinner("Generating response..."):
162
+ # Ken (12/11/2024): modify start
163
+ #response = generate_response(rag_chain, text)
164
+ ##retriever = configure_retriever(pdf_loader)
165
+ ##chain = create_chain(retriever, prompt, model)
166
+ response = inference(rag_chain, input_query)
167
+ # Ken (12/11/2024): modify end
168
  st.session_state['chat_history'].append({"user": text, "ollama": response})
169
  st.write(response)
170