BridgeEight commited on
Commit
167b9ae
1 Parent(s): b031b71

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +5 -5
app.py CHANGED
@@ -1,8 +1,8 @@
1
  from lmdeploy.serve.gradio.turbomind_coupled import *
2
  from lmdeploy.messages import TurbomindEngineConfig
3
 
4
- backend_config = TurbomindEngineConfig(max_batch_size=1, cache_max_entry_count=0.05, model_format='awq')
5
- model_path = 'internlm/internlm2-chat-20b-4bits'
6
 
7
  InterFace.async_engine = AsyncEngine(
8
  model_path=model_path,
@@ -80,13 +80,13 @@ with gr.Blocks(css=CSS, theme=THEME) as demo:
80
  with gr.Row():
81
  request_output_len = gr.Slider(1,
82
  2048,
83
- value=512,
84
  step=1,
85
  label='Maximum new tokens')
86
- top_p = gr.Slider(0.01, 1, value=0.8, step=0.01, label='Top_p')
87
  temperature = gr.Slider(0.01,
88
  1.5,
89
- value=0.7,
90
  step=0.01,
91
  label='Temperature')
92
 
 
1
  from lmdeploy.serve.gradio.turbomind_coupled import *
2
  from lmdeploy.messages import TurbomindEngineConfig
3
 
4
+ backend_config = TurbomindEngineConfig(max_batch_size=1, cache_max_entry_count=0.05)#, model_format='awq')
5
+ model_path = 'internlm/internlm2-math-7b'
6
 
7
  InterFace.async_engine = AsyncEngine(
8
  model_path=model_path,
 
80
  with gr.Row():
81
  request_output_len = gr.Slider(1,
82
  2048,
83
+ value=1024,
84
  step=1,
85
  label='Maximum new tokens')
86
+ top_p = gr.Slider(0.01, 1, value=1.0, step=0.01, label='Top_p')
87
  temperature = gr.Slider(0.01,
88
  1.5,
89
+ value=0.01,
90
  step=0.01,
91
  label='Temperature')
92