Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -20,21 +20,17 @@ def respond(
|
|
20 |
if val[1]:
|
21 |
messages.append({"role": "assistant", "content": val[1]})
|
22 |
|
|
|
23 |
messages.append({"role": "user", "content": message})
|
24 |
|
25 |
-
|
|
|
26 |
|
27 |
-
|
28 |
-
|
29 |
-
max_tokens=max_tokens,
|
30 |
-
stream=True,
|
31 |
-
temperature=temperature,
|
32 |
-
top_p=top_p,
|
33 |
-
):
|
34 |
-
token = message.choices[0].delta.content
|
35 |
|
36 |
-
|
37 |
-
|
38 |
|
39 |
|
40 |
"""
|
|
|
20 |
if val[1]:
|
21 |
messages.append({"role": "assistant", "content": val[1]})
|
22 |
|
23 |
+
# Add the user's current message to the conversation context
|
24 |
messages.append({"role": "user", "content": message})
|
25 |
|
26 |
+
# The pipeline expects a list of texts, so we will generate a response
|
27 |
+
input_text = "\n".join([msg["content"] for msg in messages if msg["role"] == "user"])
|
28 |
|
29 |
+
# Generate the response
|
30 |
+
response = pipe(input_text, max_length=max_tokens, temperature=temperature, top_p=top_p)
|
|
|
|
|
|
|
|
|
|
|
|
|
31 |
|
32 |
+
# Return the generated response
|
33 |
+
yield response[0]['generated_text']
|
34 |
|
35 |
|
36 |
"""
|