Spaces:
Sleeping
Sleeping
Tuchuanhuhuhu
commited on
Commit
•
86018c8
1
Parent(s):
60d042c
增加context上限滑动条
Browse files- ChuanhuChatbot.py +10 -1
- modules/base_model.py +4 -0
- modules/models.py +3 -0
ChuanhuChatbot.py
CHANGED
@@ -189,13 +189,21 @@ with gr.Blocks(css=customCSS, theme=small_and_beautiful_theme) as demo:
|
|
189 |
value="",
|
190 |
lines=1,
|
191 |
)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
192 |
max_tokens_slider = gr.Slider(
|
193 |
minimum=1,
|
194 |
maximum=32768,
|
195 |
value=1000,
|
196 |
step=1,
|
197 |
interactive=True,
|
198 |
-
label="max
|
199 |
)
|
200 |
presence_penalty_slider = gr.Slider(
|
201 |
minimum=-2.0,
|
@@ -379,6 +387,7 @@ with gr.Blocks(css=customCSS, theme=small_and_beautiful_theme) as demo:
|
|
379 |
downloadFile.change(**load_history_from_file_args)
|
380 |
|
381 |
# Advanced
|
|
|
382 |
temperature_slider.change(current_model.value.set_temperature, [temperature_slider], None)
|
383 |
top_p_slider.change(current_model.value.set_top_p, [top_p_slider], None)
|
384 |
n_choices_slider.change(current_model.value.set_n_choices, [n_choices_slider], None)
|
|
|
189 |
value="",
|
190 |
lines=1,
|
191 |
)
|
192 |
+
max_context_length_slider = gr.Slider(
|
193 |
+
minimum=1,
|
194 |
+
maximum=32768,
|
195 |
+
value=2000,
|
196 |
+
step=1,
|
197 |
+
interactive=True,
|
198 |
+
label="max context",
|
199 |
+
)
|
200 |
max_tokens_slider = gr.Slider(
|
201 |
minimum=1,
|
202 |
maximum=32768,
|
203 |
value=1000,
|
204 |
step=1,
|
205 |
interactive=True,
|
206 |
+
label="max generations",
|
207 |
)
|
208 |
presence_penalty_slider = gr.Slider(
|
209 |
minimum=-2.0,
|
|
|
387 |
downloadFile.change(**load_history_from_file_args)
|
388 |
|
389 |
# Advanced
|
390 |
+
max_context_length_slider.change(current_model.value.set_token_upper_limit, [max_context_length_slider], None)
|
391 |
temperature_slider.change(current_model.value.set_temperature, [temperature_slider], None)
|
392 |
top_p_slider.change(current_model.value.set_top_p, [top_p_slider], None)
|
393 |
n_choices_slider.change(current_model.value.set_n_choices, [n_choices_slider], None)
|
modules/base_model.py
CHANGED
@@ -357,6 +357,10 @@ class BaseLLMModel:
|
|
357 |
def recover(self):
|
358 |
self.interrupted = False
|
359 |
|
|
|
|
|
|
|
|
|
360 |
def set_temperature(self, new_temperature):
|
361 |
self.temperature = new_temperature
|
362 |
|
|
|
357 |
def recover(self):
|
358 |
self.interrupted = False
|
359 |
|
360 |
+
def set_token_upper_limit(self, new_upper_limit):
|
361 |
+
self.token_upper_limit = new_upper_limit
|
362 |
+
print(f"token上限设置为{new_upper_limit}")
|
363 |
+
|
364 |
def set_temperature(self, new_temperature):
|
365 |
self.temperature = new_temperature
|
366 |
|
modules/models.py
CHANGED
@@ -97,6 +97,9 @@ class OpenAIClient(BaseLLMModel):
|
|
97 |
logging.error(f"获取API使用情况失败:" + str(e))
|
98 |
return STANDARD_ERROR_MSG + ERROR_RETRIEVE_MSG
|
99 |
|
|
|
|
|
|
|
100 |
@shared.state.switching_api_key # 在不开启多账号模式的时候,这个装饰器不会起作用
|
101 |
def _get_response(self, stream=False):
|
102 |
openai_api_key = self.api_key
|
|
|
97 |
logging.error(f"获取API使用情况失败:" + str(e))
|
98 |
return STANDARD_ERROR_MSG + ERROR_RETRIEVE_MSG
|
99 |
|
100 |
+
def set_token_upper_limit(self, new_upper_limit):
|
101 |
+
pass
|
102 |
+
|
103 |
@shared.state.switching_api_key # 在不开启多账号模式的时候,这个装饰器不会起作用
|
104 |
def _get_response(self, stream=False):
|
105 |
openai_api_key = self.api_key
|