Spaces:
Paused
Paused
Update app.py
Browse files
app.py
CHANGED
|
@@ -26,7 +26,7 @@ max_new_tokens=1000
|
|
| 26 |
top_p=0.92
|
| 27 |
repetition_penalty=1.7
|
| 28 |
|
| 29 |
-
model_name = "Inagua/code-model"
|
| 30 |
|
| 31 |
llm = LLM(model_name, max_model_len=4096)
|
| 32 |
|
|
@@ -134,7 +134,8 @@ class MistralChatBot:
|
|
| 134 |
self.system_prompt = system_prompt
|
| 135 |
|
| 136 |
def predict(self, user_message, context):
|
| 137 |
-
detailed_prompt = """### Question ###\n""" + user_message + "\n\n### Contexte ###\n" + context + "\n\n### Formule ###\n"
|
|
|
|
| 138 |
prompts = [detailed_prompt]
|
| 139 |
outputs = llm.generate(prompts, sampling_params, use_tqdm = False)
|
| 140 |
generated_text = outputs[0].outputs[0].text
|
|
|
|
| 26 |
top_p=0.92
|
| 27 |
repetition_penalty=1.7
|
| 28 |
|
| 29 |
+
model_name = "Inagua/code-model-2"
|
| 30 |
|
| 31 |
llm = LLM(model_name, max_model_len=4096)
|
| 32 |
|
|
|
|
| 134 |
self.system_prompt = system_prompt
|
| 135 |
|
| 136 |
def predict(self, user_message, context):
|
| 137 |
+
#detailed_prompt = """### Question ###\n""" + user_message + "\n\n### Contexte ###\n" + context + "\n\n### Formule ###\n"
|
| 138 |
+
detailed_prompt = """### Question ###\n""" + user_message + "\n\n### Field ###\n"
|
| 139 |
prompts = [detailed_prompt]
|
| 140 |
outputs = llm.generate(prompts, sampling_params, use_tqdm = False)
|
| 141 |
generated_text = outputs[0].outputs[0].text
|