inflaton commited on
Commit
c815d49
1 Parent(s): b67e38d

hide param settings by default

Browse files
Files changed (2) hide show
  1. .env.example +2 -2
  2. app.py +43 -40
.env.example CHANGED
@@ -17,7 +17,7 @@ HF_PIPELINE_DEVICE_TYPE=
17
  # LOAD_QUANTIZED_MODEL=8bit
18
 
19
  CHAT_HISTORY_ENABLED=true
20
-
21
  PDF_FILE_BASE_URL=https://ai-engd.netlify.app/pdfs/books/
22
 
23
  # if unset, default to "hkunlp/instructor-xl"
@@ -66,6 +66,6 @@ TOKENIZERS_PARALLELISM=true
66
 
67
  # env variables for ingesting source PDF files
68
  SOURCE_PDFS_PATH="./data/pdfs/"
69
-
70
  CHUNCK_SIZE=1024
71
  CHUNK_OVERLAP=512
 
17
  # LOAD_QUANTIZED_MODEL=8bit
18
 
19
  CHAT_HISTORY_ENABLED=true
20
+ SHOW_PARAM_SETTINGS=false
21
  PDF_FILE_BASE_URL=https://ai-engd.netlify.app/pdfs/books/
22
 
23
  # if unset, default to "hkunlp/instructor-xl"
 
66
 
67
  # env variables for ingesting source PDF files
68
  SOURCE_PDFS_PATH="./data/pdfs/"
69
+ SOURCE_URLS=
70
  CHUNCK_SIZE=1024
71
  CHUNK_OVERLAP=512
app.py CHANGED
@@ -31,7 +31,9 @@ n_threds = int(os.environ.get("NUMBER_OF_CPU_CORES") or "4")
31
  index_path = os.environ.get("FAISS_INDEX_PATH") or os.environ.get("CHROMADB_INDEX_PATH")
32
  using_faiss = os.environ.get("FAISS_INDEX_PATH") is not None
33
  llm_model_type = os.environ.get("LLM_MODEL_TYPE")
34
- chat_history_enabled = os.environ.get("CHAT_HISTORY_ENABLED") or "true"
 
 
35
 
36
  streaming_enabled = True # llm_model_type in ["openai", "llamacpp"]
37
 
@@ -85,7 +87,7 @@ def qa(chatbot):
85
 
86
  with start_blocking_portal() as portal:
87
  chat_history = []
88
- if chat_history_enabled == "true":
89
  for i in range(len(chatbot) - 1):
90
  element = chatbot[i]
91
  item = (element[0] or "", element[1] or "")
@@ -151,44 +153,45 @@ with gr.Blocks(css=customCSS, theme=small_and_beautiful_theme) as demo:
151
  min_width=70,
152
  ):
153
  clearBtn = gr.Button("Clear")
154
- with gr.Column():
155
- with gr.Column(
156
- min_width=50,
157
- ):
158
- with gr.Tab(label="Parameter Setting"):
159
- gr.Markdown("# Parameters")
160
- top_p = gr.Slider(
161
- minimum=-0,
162
- maximum=1.0,
163
- value=0.95,
164
- step=0.05,
165
- # interactive=True,
166
- label="Top-p",
167
- )
168
- temperature = gr.Slider(
169
- minimum=0.1,
170
- maximum=2.0,
171
- value=0,
172
- step=0.1,
173
- # interactive=True,
174
- label="Temperature",
175
- )
176
- max_new_tokens = gr.Slider(
177
- minimum=0,
178
- maximum=2048,
179
- value=2048,
180
- step=8,
181
- # interactive=True,
182
- label="Max Generation Tokens",
183
- )
184
- max_context_length_tokens = gr.Slider(
185
- minimum=0,
186
- maximum=4096,
187
- value=4096,
188
- step=128,
189
- # interactive=True,
190
- label="Max Context Tokens",
191
- )
 
192
  gr.Markdown(description)
193
 
194
  def chat(user_message, history):
 
31
  index_path = os.environ.get("FAISS_INDEX_PATH") or os.environ.get("CHROMADB_INDEX_PATH")
32
  using_faiss = os.environ.get("FAISS_INDEX_PATH") is not None
33
  llm_model_type = os.environ.get("LLM_MODEL_TYPE")
34
+ chat_history_enabled = os.environ.get("CHAT_HISTORY_ENABLED") == "true"
35
+ show_param_settings = os.environ.get("SHOW_PARAM_SETTINGS") == "true"
36
+
37
 
38
  streaming_enabled = True # llm_model_type in ["openai", "llamacpp"]
39
 
 
87
 
88
  with start_blocking_portal() as portal:
89
  chat_history = []
90
+ if chat_history_enabled:
91
  for i in range(len(chatbot) - 1):
92
  element = chatbot[i]
93
  item = (element[0] or "", element[1] or "")
 
153
  min_width=70,
154
  ):
155
  clearBtn = gr.Button("Clear")
156
+ if show_param_settings:
157
+ with gr.Column():
158
+ with gr.Column(
159
+ min_width=50,
160
+ ):
161
+ with gr.Tab(label="Parameter Setting"):
162
+ gr.Markdown("# Parameters")
163
+ top_p = gr.Slider(
164
+ minimum=-0,
165
+ maximum=1.0,
166
+ value=0.95,
167
+ step=0.05,
168
+ # interactive=True,
169
+ label="Top-p",
170
+ )
171
+ temperature = gr.Slider(
172
+ minimum=0.1,
173
+ maximum=2.0,
174
+ value=0,
175
+ step=0.1,
176
+ # interactive=True,
177
+ label="Temperature",
178
+ )
179
+ max_new_tokens = gr.Slider(
180
+ minimum=0,
181
+ maximum=2048,
182
+ value=2048,
183
+ step=8,
184
+ # interactive=True,
185
+ label="Max Generation Tokens",
186
+ )
187
+ max_context_length_tokens = gr.Slider(
188
+ minimum=0,
189
+ maximum=4096,
190
+ value=4096,
191
+ step=128,
192
+ # interactive=True,
193
+ label="Max Context Tokens",
194
+ )
195
  gr.Markdown(description)
196
 
197
  def chat(user_message, history):