Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -64,7 +64,7 @@ class MultiAgentConversationalSystem:
|
|
64 |
|
65 |
try:
|
66 |
response = await self.client.chat.completions.create(
|
67 |
-
model="grok-
|
68 |
messages=[{"role": "user", "content": prompt}]
|
69 |
)
|
70 |
agent_response = {
|
@@ -115,7 +115,7 @@ class MultiAgentConversationalSystem:
|
|
115 |
async def manager_summary(self, policy):
|
116 |
try:
|
117 |
response = await self.client.chat.completions.create(
|
118 |
-
model="grok-
|
119 |
messages=[{"role": "user", "content": f"""Summarized this.\n\n{policy}"""}],
|
120 |
stream=False
|
121 |
)
|
@@ -149,7 +149,7 @@ class MultiAgentConversationalSystem:
|
|
149 |
|
150 |
try:
|
151 |
response = await self.client.chat.completions.create(
|
152 |
-
model="grok-
|
153 |
messages=conversation_history,
|
154 |
stream=True
|
155 |
)
|
@@ -194,7 +194,7 @@ def predict(message, history, policy_summary):
|
|
194 |
print("history_openai_format:", history_openai_format)
|
195 |
|
196 |
response = simple_client.chat.completions.create(
|
197 |
-
model='grok-
|
198 |
messages=history_openai_format,
|
199 |
temperature=0.6,
|
200 |
stream=True
|
|
|
64 |
|
65 |
try:
|
66 |
response = await self.client.chat.completions.create(
|
67 |
+
model="grok-2-1212",
|
68 |
messages=[{"role": "user", "content": prompt}]
|
69 |
)
|
70 |
agent_response = {
|
|
|
115 |
async def manager_summary(self, policy):
|
116 |
try:
|
117 |
response = await self.client.chat.completions.create(
|
118 |
+
model="grok-2-1212",
|
119 |
messages=[{"role": "user", "content": f"""Summarized this.\n\n{policy}"""}],
|
120 |
stream=False
|
121 |
)
|
|
|
149 |
|
150 |
try:
|
151 |
response = await self.client.chat.completions.create(
|
152 |
+
model="grok-2-1212",
|
153 |
messages=conversation_history,
|
154 |
stream=True
|
155 |
)
|
|
|
194 |
print("history_openai_format:", history_openai_format)
|
195 |
|
196 |
response = simple_client.chat.completions.create(
|
197 |
+
model='grok-2-1212',
|
198 |
messages=history_openai_format,
|
199 |
temperature=0.6,
|
200 |
stream=True
|