RaushanTurganbay HF staff commited on
Commit
976b0b8
1 Parent(s): c767ad8

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -3
README.md CHANGED
@@ -3,6 +3,8 @@ language:
3
  - en
4
  license: llama2
5
  pipeline_tag: image-text-to-text
 
 
6
  ---
7
 
8
  # LLaVA-NeXT-Video Model Card
@@ -18,7 +20,7 @@ LLaVA-Next-Video is an open-source chatbot trained by fine-tuning LLM on multimo
18
  The model is a current SOTA among open-source models on [VideoMME bench](https://arxiv.org/abs/2405.21075).
19
  Base LLM: [lmsys/vicuna-7b-v1.5](https://huggingface.co/lmsys/vicuna-13b-v1.5)
20
 
21
- <img src="http://drive.google.com/uc?export=view&id=1fVg-r5MU3NoHlTpD7_lYPEBWH9R8na_4">
22
 
23
 
24
  **Model date:**
@@ -231,5 +233,4 @@ If you find our paper and code useful in your research:
231
  month={January},
232
  year={2024}
233
  }
234
- ```
235
-
 
3
  - en
4
  license: llama2
5
  pipeline_tag: image-text-to-text
6
+ datasets:
7
+ - lmms-lab/VideoChatGPT
8
  ---
9
 
10
  # LLaVA-NeXT-Video Model Card
 
20
  The model is a current SOTA among open-source models on [VideoMME bench](https://arxiv.org/abs/2405.21075).
21
  Base LLM: [lmsys/vicuna-7b-v1.5](https://huggingface.co/lmsys/vicuna-13b-v1.5)
22
 
23
+ ![llava_next_video_arch](demo.png)
24
 
25
 
26
  **Model date:**
 
233
  month={January},
234
  year={2024}
235
  }
236
+ ```