Spaces:
Runtime error
Runtime error
Commit
·
004e770
1
Parent(s):
e053720
updates
Browse files
app.py
CHANGED
@@ -31,12 +31,54 @@ creds = ServiceAccountCredentials.from_json_keyfile_dict(GOOGLE_SHEETS_CERTIFICA
|
|
31 |
client = gspread.authorize(creds)
|
32 |
|
33 |
models = [
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
34 |
BaseModel(
|
35 |
name="mosaicml/mpt-7b",
|
36 |
endpoint="mpt-7b",
|
37 |
namespace="tenant-chairesearch-test",
|
38 |
generation_params={
|
39 |
-
'temperature': 0.
|
40 |
'repetition_penalty': 1.0,
|
41 |
'max_new_tokens': 128,
|
42 |
'top_k': 10,
|
@@ -50,7 +92,7 @@ models = [
|
|
50 |
endpoint="mpt-7b-storywriter",
|
51 |
namespace="tenant-chairesearch-test",
|
52 |
generation_params={
|
53 |
-
'temperature': 0.
|
54 |
'repetition_penalty': 1.0,
|
55 |
'max_new_tokens': 128,
|
56 |
'top_k': 10,
|
@@ -64,7 +106,7 @@ models = [
|
|
64 |
endpoint="mpt-7b-chat",
|
65 |
namespace="tenant-chairesearch-test",
|
66 |
generation_params={
|
67 |
-
'temperature': 0.
|
68 |
'repetition_penalty': 1.0,
|
69 |
'max_new_tokens': 128,
|
70 |
'top_k': 10,
|
@@ -78,7 +120,7 @@ models = [
|
|
78 |
endpoint="redpajama-base-7b",
|
79 |
namespace="tenant-chairesearch-test",
|
80 |
generation_params={
|
81 |
-
'temperature': 0.
|
82 |
'repetition_penalty': 1.0,
|
83 |
'max_new_tokens': 128,
|
84 |
'top_k': 10,
|
@@ -92,7 +134,7 @@ models = [
|
|
92 |
endpoint="redpajama-chat-7b",
|
93 |
namespace="tenant-chairesearch-test",
|
94 |
generation_params={
|
95 |
-
'temperature': 0.
|
96 |
'repetition_penalty': 1.0,
|
97 |
'max_new_tokens': 64,
|
98 |
'top_k': 10,
|
|
|
31 |
client = gspread.authorize(creds)
|
32 |
|
33 |
models = [
|
34 |
+
BaseModel(
|
35 |
+
name="pygmalion-13b",
|
36 |
+
endpoint="pygmalion-13b",
|
37 |
+
namespace="tenant-chaiml-guanaco",
|
38 |
+
generation_params={
|
39 |
+
'temperature': 0.7,
|
40 |
+
'repetition_penalty': 1.0,
|
41 |
+
'max_new_tokens': 128,
|
42 |
+
'top_k': 10,
|
43 |
+
'top_p': 0.9,
|
44 |
+
'do_sample': True,
|
45 |
+
'eos_token_id': 13,
|
46 |
+
}
|
47 |
+
),
|
48 |
+
BaseModel(
|
49 |
+
name="lmsys/vicuna-7b-delta-v1.1",
|
50 |
+
endpoint="vicuna-7b",
|
51 |
+
namespace="tenant-chairesearch-test",
|
52 |
+
generation_params={
|
53 |
+
'temperature': 0.7,
|
54 |
+
'repetition_penalty': 1.0,
|
55 |
+
'max_new_tokens': 128,
|
56 |
+
'top_k': 10,
|
57 |
+
'top_p': 0.9,
|
58 |
+
'do_sample': True,
|
59 |
+
'eos_token_id': 13,
|
60 |
+
}
|
61 |
+
),
|
62 |
+
BaseModel(
|
63 |
+
name="PygmalionAI/pygmalion-7b",
|
64 |
+
endpoint="pygmalion-7b",
|
65 |
+
namespace="tenant-chairesearch-test",
|
66 |
+
generation_params={
|
67 |
+
'temperature': 0.7,
|
68 |
+
'repetition_penalty': 1.0,
|
69 |
+
'max_new_tokens': 128,
|
70 |
+
'top_k': 10,
|
71 |
+
'top_p': 0.9,
|
72 |
+
'do_sample': True,
|
73 |
+
'eos_token_id': 13,
|
74 |
+
}
|
75 |
+
),
|
76 |
BaseModel(
|
77 |
name="mosaicml/mpt-7b",
|
78 |
endpoint="mpt-7b",
|
79 |
namespace="tenant-chairesearch-test",
|
80 |
generation_params={
|
81 |
+
'temperature': 0.7,
|
82 |
'repetition_penalty': 1.0,
|
83 |
'max_new_tokens': 128,
|
84 |
'top_k': 10,
|
|
|
92 |
endpoint="mpt-7b-storywriter",
|
93 |
namespace="tenant-chairesearch-test",
|
94 |
generation_params={
|
95 |
+
'temperature': 0.7,
|
96 |
'repetition_penalty': 1.0,
|
97 |
'max_new_tokens': 128,
|
98 |
'top_k': 10,
|
|
|
106 |
endpoint="mpt-7b-chat",
|
107 |
namespace="tenant-chairesearch-test",
|
108 |
generation_params={
|
109 |
+
'temperature': 0.7,
|
110 |
'repetition_penalty': 1.0,
|
111 |
'max_new_tokens': 128,
|
112 |
'top_k': 10,
|
|
|
120 |
endpoint="redpajama-base-7b",
|
121 |
namespace="tenant-chairesearch-test",
|
122 |
generation_params={
|
123 |
+
'temperature': 0.7,
|
124 |
'repetition_penalty': 1.0,
|
125 |
'max_new_tokens': 128,
|
126 |
'top_k': 10,
|
|
|
134 |
endpoint="redpajama-chat-7b",
|
135 |
namespace="tenant-chairesearch-test",
|
136 |
generation_params={
|
137 |
+
'temperature': 0.7,
|
138 |
'repetition_penalty': 1.0,
|
139 |
'max_new_tokens': 64,
|
140 |
'top_k': 10,
|