StarCycle osanseviero commited on
Commit
da0e35f
1 Parent(s): 598ad31

Specify right model card metadata (#1)

Browse files

- Specify right model card metadata (d328c9e5e46df5628dcba8c53ae28fbffe963f11)


Co-authored-by: Omar Sanseviero <osanseviero@users.noreply.huggingface.co>

Files changed (1) hide show
  1. README.md +3 -0
README.md CHANGED
@@ -1,5 +1,8 @@
1
  ---
2
  license: apache-2.0
 
 
 
3
  ---
4
  ## Model
5
  llava-clip-internlm2-1_8b-pretrain-v1 is a LLaVA checkpoint finetuned from [internlm2-chat-1_8b](https://huggingface.co/internlm/internlm2-chat-1_8b) and [CLIP-ViT-Large-patch14-336](https://huggingface.co/openai/clip-vit-large-patch14-336) with [LLaVA-Pretrain](liuhaotian/LLaVA-Pretrain) by [Xtuner](https://github.com/InternLM/xtuner). The pretraining phase took 16 hours on a single Nvidia A6000 ada GPU.
 
1
  ---
2
  license: apache-2.0
3
+ tags:
4
+ - llava
5
+ pipeline_tag: image-text-to-text
6
  ---
7
  ## Model
8
  llava-clip-internlm2-1_8b-pretrain-v1 is a LLaVA checkpoint finetuned from [internlm2-chat-1_8b](https://huggingface.co/internlm/internlm2-chat-1_8b) and [CLIP-ViT-Large-patch14-336](https://huggingface.co/openai/clip-vit-large-patch14-336) with [LLaVA-Pretrain](liuhaotian/LLaVA-Pretrain) by [Xtuner](https://github.com/InternLM/xtuner). The pretraining phase took 16 hours on a single Nvidia A6000 ada GPU.