Model Trained Using AutoTrain

This model was fined tuned on 10K + Indian food recipes , the base model is TinyLlama/TinyLlama-1.1B-Chat-v0.6 (https://huggingface.co/TinyLlama/TinyLlama-1.1B-Chat-v0.6 )

Dataset used for this Model is here https://huggingface.co/datasets/VishalMysore/cookGPT

CookGPT: Your AI-Based Chef

CookGPT is an innovative AI-based chef that combines the charm of traditional cooking with the efficiency of modern technology. Whether you're a culinary enthusiast, a busy professional, or someone looking for culinary inspiration, CookGPT is designed to make your cooking experience delightful, personalized, and effortless.

Original Source code and all the related information is here

https://github.com/vishalmysore/cookGPT

How to use

from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline

model = AutoModelForCausalLM.from_pretrained("VishalMysore/cookgptlama")
tokenizer = AutoTokenizer.from_pretrained("VishalMysore/cookgptlama")

or you can load it in 8bit precision

from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline

model_id = "VishalMysore/cookgptlama"

tokenizer = AutoTokenizer.from_pretrained(model_id)
model_8bit = AutoModelForCausalLM.from_pretrained(model_id, device_map="auto",load_in_8bit=True)
print(model_8bit.get_memory_footprint())

Then use pipeline to query and interact


pipe = pipeline("text-generation", model=model, tokenizer=tokenizer)

# We use the tokenizer's chat template to format each message - see https://huggingface.co/docs/transformers/main/en/chat_templating
messages = [
    {
        "role": "system",
        "content": "you are an expert chef in Indian recipe",
    },
    {"role": "user", "content": "give me receipe for paneer butter masala with cook time diet and cusine and instructions"},
]
prompt = pipe.tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
outputs = pipe(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
print(outputs[0]["generated_text"])

Complete notebook for training and querying is here https://github.com/vishalmysore/AI/blob/main/Llama_CookGPT_AutoTrain_LLM.ipynb

Downloads last month
13
Safetensors
Model size
1.1B params
Tensor type
FP16
ยท
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Spaces using VishalMysore/cookgptlama 2