Edit model card

Gemma-2b-it-finetuned-python-codes

This model card corresponds to the 2B finetuned version of the Gemma-2b-it model. You can visit the model card of the 2B Gemma Instruct.

Author: Dishank Shah

Description

GifPC-2b (Gemma-2b-it-finetuned-python-codes) LLM is trained on a dataset containing Python code snippets. This specialized training aimed to enhance Gemma-2b-it's understanding of Python syntax, semantics, and common programming patterns. With this finetuning, Gemma-2b-it is now proficient in not only comprehending Python code but also capable of assisting in debugging tasks. Users can leverage its trained knowledge to seek guidance on Python-related issues, understand code logic, and identify potential errors within their programs. This specialized Gemma-2b-it variant serves as a valuable tool for programmers seeking assistance and guidance in Python programming and debugging tasks.

Usage

Below we share some code snippets on how to get quickly started with running the model. First make sure to pip install -U transformers, then copy the snippet from the section that is relevant for your usecase.

Running the model on Google Colab CPU

from transformers import AutoTokenizer, AutoModelForCausalLM

model_name = "shahdishank/gemma-2b-it-finetune-python-codes"
HUGGING_FACE_TOKEN = "YOUR_TOKEN"
tokenizer = AutoTokenizer.from_pretrained(model_name, token="HUGGING_FACE_TOKEN")
model = AutoModelForCausalLM.from_pretrained(model_name, token="HUGGING_FACE_TOKEN")

prompt_template = """\
  user:\n{query} \n\n assistant:\n
  """
prompt = prompt_template.format(query="write a simple python function") # write your query here

input_ids = tokenizer(prompt, return_tensors="pt", add_special_tokens=True)
outputs = model.generate(**input_ids, max_new_tokens=2000, do_sample=True, pad_token_id=tokenizer.eos_token_id)
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(response)

Model Data

Data used for model training python-codes-25k.

Training Dataset

These models were trained on a dataset of text data that includes a wide variety of python codes. Here are the key components:

  • Instruction: The instructional task to be performed / User input.
  • Input: Very short, introductive part of AI response or empty.
  • Output: Python code that accomplishes the task.
  • Text: All fields combined together.

This diverse data source is crucial for training a powerful language model that can handle a wide variety of different tasks.

Usage

This LLM can be used for:

  • Code generation
  • Debugging
  • Learn and understand various python coding styles
Downloads last month
18
Safetensors
Model size
2.51B params
Tensor type
FP16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Dataset used to train shahdishank/gemma-2b-it-finetune-python-codes