|
--- |
|
license: apache-2.0 |
|
base_model: |
|
- Qwen/Qwen2-VL-7B-Instruct |
|
--- |
|
|
|
|
|
## How to use it |
|
|
|
1. To get the Code: |
|
``` |
|
git clone https://github.com/HimariO/llama.cpp.git |
|
cd llama.cpp |
|
git switch qwen2-vl |
|
``` |
|
2. nano Makefile #(to add llama-qwen2vl-cli) |
|
|
|
``` |
|
diff --git a/Makefile b/Makefile |
|
index 8a903d7e..51403be2 100644 |
|
--- a/Makefile |
|
+++ b/Makefile |
|
@@ -1485,6 +1485,14 @@ libllava.a: examples/llava/llava.cpp \ |
|
$(OBJ_ALL) |
|
$(CXX) $(CXXFLAGS) -static -fPIC -c $< -o $@ -Wno-cast-qual |
|
|
|
+llama-qwen2vl-cli: examples/llava/qwen2vl-cli.cpp \ |
|
+ examples/llava/llava.cpp \ |
|
+ examples/llava/llava.h \ |
|
+ examples/llava/clip.cpp \ |
|
+ examples/llava/clip.h \ |
|
+ $(OBJ_ALL) |
|
+ $(CXX) $(CXXFLAGS) $< $(filter-out %.h $<,$^) -o $@ $(LDFLAGS) -Wno-cast-qual |
|
+ |
|
llama-llava-cli: examples/llava/llava-cli.cpp \ |
|
examples/llava/llava.cpp \ |
|
examples/llava/llava.h \ |
|
|
|
``` |
|
|
|
3. Metal Build |
|
``` |
|
cmake . -DGGML_CUDA=ON -DCMAKE_CUDA_COMPILER=$(which nvcc) -DTCNN_CUDA_ARCHITECTURES=61 |
|
make -j35 |
|
``` |
|
|
|
4. RUN |
|
``` |
|
./bin/llama-qwen2vl-cli -m ./thomas-yanxin/Qwen2-VL-7B-GGUF/Qwen2-VL-7B-GGUF-Q4_K_M.gguf --mmproj ./thomas-yanxin/Qwen2-VL-7B-GGUF/qwen2vl-vision.gguf -p "Describe the image" --image "./thomas-yanxin/Qwen2-VL-7B-GGUF/1.png" |
|
``` |
|
|
|
|