Spaces:
Sleeping
Sleeping
Update vlog4chat.py
Browse files- vlog4chat.py +2 -4
vlog4chat.py
CHANGED
@@ -51,8 +51,6 @@ Chat History:
|
|
51 |
Follow Up Input: {question}
|
52 |
Standalone question:"""
|
53 |
|
54 |
-
os.environ["TOGETHER_API_KEY"] = "48bf2536f85b599c7d5d7f9921cc9ee7056f40ed535fd2174d061e1b9abcf8af"
|
55 |
-
|
56 |
def get_prompt(instruction, new_system_prompt=DEFAULT_SYSTEM_PROMPT ):
|
57 |
SYSTEM_PROMPT = B_SYS + new_system_prompt + E_SYS
|
58 |
prompt_template = B_INST + SYSTEM_PROMPT + instruction + E_INST
|
@@ -70,7 +68,7 @@ class TogetherLLM(LLM):
|
|
70 |
model: str = "togethercomputer/llama-2-70b-chat"
|
71 |
"""model endpoint to use"""
|
72 |
|
73 |
-
together_api_key: str = os.
|
74 |
"""Together API key"""
|
75 |
|
76 |
temperature: float = 0.7
|
@@ -131,7 +129,7 @@ class Vlogger4chat :
|
|
131 |
def init_llm(self):
|
132 |
print('\033[1;33m' + "Initializing LLM Reasoner...".center(50, '-') + '\033[0m')
|
133 |
self.llm = TogetherLLM(
|
134 |
-
model= os.getenv(
|
135 |
temperature=0.1,
|
136 |
max_tokens=768
|
137 |
)
|
|
|
51 |
Follow Up Input: {question}
|
52 |
Standalone question:"""
|
53 |
|
|
|
|
|
54 |
def get_prompt(instruction, new_system_prompt=DEFAULT_SYSTEM_PROMPT ):
|
55 |
SYSTEM_PROMPT = B_SYS + new_system_prompt + E_SYS
|
56 |
prompt_template = B_INST + SYSTEM_PROMPT + instruction + E_INST
|
|
|
68 |
model: str = "togethercomputer/llama-2-70b-chat"
|
69 |
"""model endpoint to use"""
|
70 |
|
71 |
+
together_api_key: str = os.getenv["TOGETHER_API_KEY"]
|
72 |
"""Together API key"""
|
73 |
|
74 |
temperature: float = 0.7
|
|
|
129 |
def init_llm(self):
|
130 |
print('\033[1;33m' + "Initializing LLM Reasoner...".center(50, '-') + '\033[0m')
|
131 |
self.llm = TogetherLLM(
|
132 |
+
model= os.getenv("Your_Model_Name"),
|
133 |
temperature=0.1,
|
134 |
max_tokens=768
|
135 |
)
|