Edit model card

Llama 3.1 Vision by Capx AI

image/png

Directions to Run Inference:

Minimum requirements to run Inference is an A100 40GB GPU.

  • Clone our fork of the Bunny by BAAI repository here: https://github.com/adarshxs/Capx-Llama-3.1-Carrot
  • Create a conda virtual environment
    conda create -n capx python=3.10
    conda activate capx
    
  • Install the following
    pip install --upgrade pip  # enable PEP 660 support
    pip install transformers
    pip install torch torchvision xformers --index-url https://download.pytorch.org/whl/cu118
    
    # Installing APEX
    pip install ninja
    git clone https://github.com/NVIDIA/apex
    cd apex
    pip install -v --disable-pip-version-check --no-cache-dir --no-build-isolation --global-option="--cpp_ext" --global-option="--cuda_ext" ./
    cd ..
    
    # Installing Flash Attn
    pip install packaging
    pip install flash-attn --no-build-isolation
    
    # Clone the inference Repo
    git clone https://github.com/adarshxs/Capx-Llama3.1-Vision
    cd Capx-Llama3.1-Vision
    pip install -e .
    
  • Run cli server:
    python -m bunny.serve.cli \
      --model-path Capx/Llama-3.1-Vision \
      --model-type llama3.1-8b \
      --image-file /path/to/image \
      --conv-mode llama
    

We thank the amazing team at BAAI, for their Bunny project, upon which this was built and Meta AI for their Llama 3.1 model!

Downloads last month
27
Safetensors
Model size
8.48B params
Tensor type
FP16
·
Inference Examples
Unable to determine this model's library. Check the docs .

Model tree for Capx/Llama-3.1-Carrot

Finetuned
(400)
this model