Spaces:
Sleeping
Sleeping
Decrease tokens by sending less messages
Browse files- functions.py +8 -2
- prompt.txt +0 -1
functions.py
CHANGED
@@ -23,7 +23,7 @@ def get_main_data():
|
|
23 |
|
24 |
authors = ['Sofia', 'Eliza', 'Sindy', 'Carlos', 'Andres', 'Adriana', 'Carolina', 'Valeria']
|
25 |
model = "gpt-4"
|
26 |
-
temperature_values = [0.8, 1.0
|
27 |
|
28 |
return scores_parameters, authors, model, temperature_values
|
29 |
|
@@ -106,13 +106,19 @@ def handle_call(msg_history: gr.State, temperature: gr.State):
|
|
106 |
def get_ai_answer(msg: str, msg_history: gr.State, temperature: gr.State):
|
107 |
"""
|
108 |
Returns the response given by the model, all the message history so far and the seconds
|
109 |
-
the api took to retrieve such response.
|
|
|
110 |
"""
|
111 |
msg_history.append({"role": "user", "content": msg})
|
112 |
response, needed_time = handle_call(msg_history, temperature)
|
113 |
AI_response = response["choices"][0]["message"]["content"]
|
114 |
msg_history.append({'role': 'assistant', 'content': AI_response})
|
115 |
|
|
|
|
|
|
|
|
|
|
|
116 |
return AI_response, msg_history, needed_time
|
117 |
|
118 |
|
|
|
23 |
|
24 |
authors = ['Sofia', 'Eliza', 'Sindy', 'Carlos', 'Andres', 'Adriana', 'Carolina', 'Valeria']
|
25 |
model = "gpt-4"
|
26 |
+
temperature_values = [0.8, 1.0]
|
27 |
|
28 |
return scores_parameters, authors, model, temperature_values
|
29 |
|
|
|
106 |
def get_ai_answer(msg: str, msg_history: gr.State, temperature: gr.State):
|
107 |
"""
|
108 |
Returns the response given by the model, all the message history so far and the seconds
|
109 |
+
the api took to retrieve such response. It also removes some messages in the message history
|
110 |
+
so only the last n (keep) are used (costs are cheaper)
|
111 |
"""
|
112 |
msg_history.append({"role": "user", "content": msg})
|
113 |
response, needed_time = handle_call(msg_history, temperature)
|
114 |
AI_response = response["choices"][0]["message"]["content"]
|
115 |
msg_history.append({'role': 'assistant', 'content': AI_response})
|
116 |
|
117 |
+
keep = 3 # Number of messages to keep
|
118 |
+
if len(msg_history) > (2 * keep) + 1: # last +1 is due to the system message
|
119 |
+
msg_history.pop(1)
|
120 |
+
msg_history.pop(1)
|
121 |
+
|
122 |
return AI_response, msg_history, needed_time
|
123 |
|
124 |
|
prompt.txt
CHANGED
@@ -67,7 +67,6 @@ Manejo de situaciones no contempladas:
|
|
67 |
|
68 |
Informacion importante:
|
69 |
- Antes de responder, aseg煤rate de que tu mensaje:
|
70 |
-
* Tenga 50 palabras o menos.
|
71 |
* Refleje la personalidad de Roomie.
|
72 |
* Est茅 alineado con los escenarios proporcionados.
|
73 |
* Sea coherente con el tono y estilo de comunicaci贸n definidos anteriormente.
|
|
|
67 |
|
68 |
Informacion importante:
|
69 |
- Antes de responder, aseg煤rate de que tu mensaje:
|
|
|
70 |
* Refleje la personalidad de Roomie.
|
71 |
* Est茅 alineado con los escenarios proporcionados.
|
72 |
* Sea coherente con el tono y estilo de comunicaci贸n definidos anteriormente.
|