hide param settings by default
Browse files- .env.example +2 -2
- app.py +43 -40
.env.example
CHANGED
@@ -17,7 +17,7 @@ HF_PIPELINE_DEVICE_TYPE=
|
|
17 |
# LOAD_QUANTIZED_MODEL=8bit
|
18 |
|
19 |
CHAT_HISTORY_ENABLED=true
|
20 |
-
|
21 |
PDF_FILE_BASE_URL=https://ai-engd.netlify.app/pdfs/books/
|
22 |
|
23 |
# if unset, default to "hkunlp/instructor-xl"
|
@@ -66,6 +66,6 @@ TOKENIZERS_PARALLELISM=true
|
|
66 |
|
67 |
# env variables for ingesting source PDF files
|
68 |
SOURCE_PDFS_PATH="./data/pdfs/"
|
69 |
-
|
70 |
CHUNCK_SIZE=1024
|
71 |
CHUNK_OVERLAP=512
|
|
|
17 |
# LOAD_QUANTIZED_MODEL=8bit
|
18 |
|
19 |
CHAT_HISTORY_ENABLED=true
|
20 |
+
SHOW_PARAM_SETTINGS=false
|
21 |
PDF_FILE_BASE_URL=https://ai-engd.netlify.app/pdfs/books/
|
22 |
|
23 |
# if unset, default to "hkunlp/instructor-xl"
|
|
|
66 |
|
67 |
# env variables for ingesting source PDF files
|
68 |
SOURCE_PDFS_PATH="./data/pdfs/"
|
69 |
+
SOURCE_URLS=
|
70 |
CHUNCK_SIZE=1024
|
71 |
CHUNK_OVERLAP=512
|
app.py
CHANGED
@@ -31,7 +31,9 @@ n_threds = int(os.environ.get("NUMBER_OF_CPU_CORES") or "4")
|
|
31 |
index_path = os.environ.get("FAISS_INDEX_PATH") or os.environ.get("CHROMADB_INDEX_PATH")
|
32 |
using_faiss = os.environ.get("FAISS_INDEX_PATH") is not None
|
33 |
llm_model_type = os.environ.get("LLM_MODEL_TYPE")
|
34 |
-
chat_history_enabled = os.environ.get("CHAT_HISTORY_ENABLED")
|
|
|
|
|
35 |
|
36 |
streaming_enabled = True # llm_model_type in ["openai", "llamacpp"]
|
37 |
|
@@ -85,7 +87,7 @@ def qa(chatbot):
|
|
85 |
|
86 |
with start_blocking_portal() as portal:
|
87 |
chat_history = []
|
88 |
-
if chat_history_enabled
|
89 |
for i in range(len(chatbot) - 1):
|
90 |
element = chatbot[i]
|
91 |
item = (element[0] or "", element[1] or "")
|
@@ -151,44 +153,45 @@ with gr.Blocks(css=customCSS, theme=small_and_beautiful_theme) as demo:
|
|
151 |
min_width=70,
|
152 |
):
|
153 |
clearBtn = gr.Button("Clear")
|
154 |
-
|
155 |
-
with gr.Column(
|
156 |
-
|
157 |
-
|
158 |
-
|
159 |
-
gr.
|
160 |
-
|
161 |
-
|
162 |
-
|
163 |
-
|
164 |
-
|
165 |
-
|
166 |
-
|
167 |
-
|
168 |
-
|
169 |
-
|
170 |
-
|
171 |
-
|
172 |
-
|
173 |
-
|
174 |
-
|
175 |
-
|
176 |
-
|
177 |
-
|
178 |
-
|
179 |
-
|
180 |
-
|
181 |
-
|
182 |
-
|
183 |
-
|
184 |
-
|
185 |
-
|
186 |
-
|
187 |
-
|
188 |
-
|
189 |
-
|
190 |
-
|
191 |
-
|
|
|
192 |
gr.Markdown(description)
|
193 |
|
194 |
def chat(user_message, history):
|
|
|
31 |
index_path = os.environ.get("FAISS_INDEX_PATH") or os.environ.get("CHROMADB_INDEX_PATH")
|
32 |
using_faiss = os.environ.get("FAISS_INDEX_PATH") is not None
|
33 |
llm_model_type = os.environ.get("LLM_MODEL_TYPE")
|
34 |
+
chat_history_enabled = os.environ.get("CHAT_HISTORY_ENABLED") == "true"
|
35 |
+
show_param_settings = os.environ.get("SHOW_PARAM_SETTINGS") == "true"
|
36 |
+
|
37 |
|
38 |
streaming_enabled = True # llm_model_type in ["openai", "llamacpp"]
|
39 |
|
|
|
87 |
|
88 |
with start_blocking_portal() as portal:
|
89 |
chat_history = []
|
90 |
+
if chat_history_enabled:
|
91 |
for i in range(len(chatbot) - 1):
|
92 |
element = chatbot[i]
|
93 |
item = (element[0] or "", element[1] or "")
|
|
|
153 |
min_width=70,
|
154 |
):
|
155 |
clearBtn = gr.Button("Clear")
|
156 |
+
if show_param_settings:
|
157 |
+
with gr.Column():
|
158 |
+
with gr.Column(
|
159 |
+
min_width=50,
|
160 |
+
):
|
161 |
+
with gr.Tab(label="Parameter Setting"):
|
162 |
+
gr.Markdown("# Parameters")
|
163 |
+
top_p = gr.Slider(
|
164 |
+
minimum=-0,
|
165 |
+
maximum=1.0,
|
166 |
+
value=0.95,
|
167 |
+
step=0.05,
|
168 |
+
# interactive=True,
|
169 |
+
label="Top-p",
|
170 |
+
)
|
171 |
+
temperature = gr.Slider(
|
172 |
+
minimum=0.1,
|
173 |
+
maximum=2.0,
|
174 |
+
value=0,
|
175 |
+
step=0.1,
|
176 |
+
# interactive=True,
|
177 |
+
label="Temperature",
|
178 |
+
)
|
179 |
+
max_new_tokens = gr.Slider(
|
180 |
+
minimum=0,
|
181 |
+
maximum=2048,
|
182 |
+
value=2048,
|
183 |
+
step=8,
|
184 |
+
# interactive=True,
|
185 |
+
label="Max Generation Tokens",
|
186 |
+
)
|
187 |
+
max_context_length_tokens = gr.Slider(
|
188 |
+
minimum=0,
|
189 |
+
maximum=4096,
|
190 |
+
value=4096,
|
191 |
+
step=128,
|
192 |
+
# interactive=True,
|
193 |
+
label="Max Context Tokens",
|
194 |
+
)
|
195 |
gr.Markdown(description)
|
196 |
|
197 |
def chat(user_message, history):
|