Update app.py
Browse files
app.py
CHANGED
@@ -75,9 +75,8 @@ def model_inference(
|
|
75 |
|
76 |
|
77 |
with gr.Blocks(fill_height=True) as demo:
|
78 |
-
gr.Markdown("## SmolVLM
|
79 |
-
gr.Markdown("Play with [HuggingFaceTB/SmolVLM](https://huggingface.co/HuggingFaceTB/SmolVLM) in this demo. To get started, upload an image and text or try one of the examples.")
|
80 |
-
gr.Markdown("**Disclaimer:** SmolVLM does not include an RLHF alignment stage, so it may not consistently follow prompts or handle complex tasks. However, this doesn't mean it is incapable of doing so. Adding a prefix to the assistant's response, such as Let's think step for a reasoning question or `<html>` for HTML code generation, can significantly improve the output in practice. You could also play with the parameters such as the temperature in non-greedy mode.")
|
81 |
with gr.Column():
|
82 |
image_input = gr.Image(label="Upload your Image", type="pil", scale=1)
|
83 |
query_input = gr.Textbox(label="Prompt")
|
|
|
75 |
|
76 |
|
77 |
with gr.Blocks(fill_height=True) as demo:
|
78 |
+
gr.Markdown("## SmolVLM")
|
79 |
+
gr.Markdown("Play with [HuggingFaceTB/SmolVLM-Instruct](https://huggingface.co/HuggingFaceTB/SmolVLM-Instruct) in this demo. To get started, upload an image and text or try one of the examples.")
|
|
|
80 |
with gr.Column():
|
81 |
image_input = gr.Image(label="Upload your Image", type="pil", scale=1)
|
82 |
query_input = gr.Textbox(label="Prompt")
|