metadata
language:
- en
license: cc-by-4.0
tags:
- llava
datasets:
- taesiri/video-game-question-answering
- taesiri/glitch-llava-game-qa-dataset-wip
- taesiri/GameplayCaptions-GPT-4V-V2
- taesiri/video-game-question-answering-mixtral-8x7b-instruct-v0-1
inference: false
pipeline_tag: image-text-to-text
LLaVA-VideoGameVQA - Work In Progress - Model Card
Model details
Model type: LLaVA is an open-source chatbot trained by fine-tuning LLaMA/Vicuna on GPT-generated multimodal instruction-following data. It is an auto-regressive language model, based on the transformer architecture.
Model date: LLaVA-v1.5-13B-LoRA was trained in December 2023.
LoRA Weights
- Checkpoint 1 trained on
28K
question-answering pairs. Base Model:liuhaotian/llava-v1.5-13b
- Checkpoint 5 trained on
74K
question-answering pairs. Base Model:liuhaotian/llava-v1.5-13b
- Checkpoint 8 trained on
185K
question-answering pairs. Base Model:liuhaotian/llava-v1.5-13b
How to run
python -m llava.serve.model_worker --host 0.0.0.0 --controller http://localhost:10000 --port 40000 --worker http://localhost:40000 --model-path ./lora-checkpoints-8 --model-base liuhaotian/llava-v1.5-13b