Edit model card

Model Card for Model ID

Fine-tuned multimodal LLaVA model for TikZ diagram generation using hand-drawn sketches.

How to Get Started with the Model

from transformers import pipeline
from PIL import Image
import requests

pipe = pipeline("image-to-text", model="waleko/TikZ-llava-1.5-7b")

url = "https://waleko.github.io/data/image.jpg"
image = Image.open(requests.get(url, stream=True).raw)
prompt = "Assistant helps to write down the TikZ code for the user's image. USER: <image>\nWrite down the TikZ code to draw the diagram shown in the image. ASSISTANT: "

print(pipe(image, prompt=prompt)[0]['generated_text'])

Training Details

Training Data

Trained on synthetic TikZ-short-code dataset.

Downloads last month
24
Safetensors
Model size
7.06B params
Tensor type
FP16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Dataset used to train waleko/TikZ-llava-1.5-7b

Space using waleko/TikZ-llava-1.5-7b 1