File size: 1,157 Bytes
590af54
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
import subprocess

if __name__ == '__main__':
    backend_comand = ['python3', 'src/demo/seed_llama_flask.py', '--image_transform', 'configs/processer/qwen_448_transform.yaml', \
                      '--tokenizer', 'configs/tokenizer/clm_llama_tokenizer_224loc_anyres.yaml', \
                      '--llm', 'configs/clm_models/llm_seed_x_i.yaml', \
                      '--visual_encoder', 'configs/visual_encoder/qwen_vitg_448.yaml', \
                      '--sd_adapter', 'configs/sdxl_adapter/sdxl_qwen_vit_resampler_l4_q64_pretrain_no_normalize.yaml', \
                      '--agent', 'configs/clm_models/agent_seed_x_i.yaml', \
                      '--diffusion_path', 'https://huggingface.co/stabilityai/stable-diffusion-xl-base-1.0', \
                      '--port', '7890', '--llm_device', 'cuda:0', '--vit_sd_device', 'cuda:0', '--multi_resolution', 'True', '--has_bbox']
    
    frontend_comand = ['python3', 'src/demo/seed_llama_gradio.py', '--server_port', '7860', '--request_address', 'http://127.0.0.1:7890/generate']
    
    backend_proc = subprocess.Popen(backend_comand)
    
    frontend_proc = subprocess.Popen(frontend_comand)