LLaVa-1.5 is a series of vision-language models (VLMs) trained on a variety of visual instruction datasets.