Update functions.py
Browse files- functions.py +2 -2
functions.py
CHANGED
@@ -512,7 +512,7 @@ def generate_eval(raw_text, N, chunk):
|
|
512 |
n = len(raw_text)
|
513 |
starting_indices = [random.randint(0, n-chunk) for _ in range(N)]
|
514 |
sub_sequences = [raw_text[i:i+chunk] for i in starting_indices]
|
515 |
-
chain = QAGenerationChain.from_llm(ChatOpenAI(temperature=0, model_name='gpt-4
|
516 |
eval_set = []
|
517 |
for i, b in enumerate(sub_sequences):
|
518 |
try:
|
@@ -570,7 +570,7 @@ def embed_text(query,_docsearch):
|
|
570 |
|
571 |
# llm = OpenAI(temperature=0)
|
572 |
chat_llm = ChatOpenAI(streaming=True,
|
573 |
-
model_name = 'gpt-4
|
574 |
callbacks=[StdOutCallbackHandler()],
|
575 |
verbose=True,
|
576 |
temperature=0
|
|
|
512 |
n = len(raw_text)
|
513 |
starting_indices = [random.randint(0, n-chunk) for _ in range(N)]
|
514 |
sub_sequences = [raw_text[i:i+chunk] for i in starting_indices]
|
515 |
+
chain = QAGenerationChain.from_llm(ChatOpenAI(temperature=0, model_name='gpt-4'))
|
516 |
eval_set = []
|
517 |
for i, b in enumerate(sub_sequences):
|
518 |
try:
|
|
|
570 |
|
571 |
# llm = OpenAI(temperature=0)
|
572 |
chat_llm = ChatOpenAI(streaming=True,
|
573 |
+
model_name = 'gpt-4',
|
574 |
callbacks=[StdOutCallbackHandler()],
|
575 |
verbose=True,
|
576 |
temperature=0
|