Text Generation
Transformers
PyTorch
mpt
Composer
MosaicML
llm-foundry
custom_code
text-generation-inference

Provide fine-tuning example notebook using hf transformers

#22
by MakerMotion - opened

Can anyone provide a example fine-tuning notebook with custom data using hf transformers. Specifically I wonder if 'labels' are shifted automatically like GPT2 model or how do I provide the 'labels' to the model on training time.

@MakerMotion Did you find an answer to this?

@zachblank I think so. Because MPT is not fully implemented on this version of HF, I took a look at their model repo and in this file https://huggingface.co/mosaicml/mpt-7b-instruct/blob/main/modeling_mpt.py on forward() function; it seems like if you provide a labels argument it shifts automatically. [line 244]

@MakerMotion Thanks! Do you have an example notebook you could share? I'm new at this and still trying to wrap my head around it. Thanks!

Closing as stale

abhi-mosaic changed discussion status to closed

Is there a concrete Notebook example of taking the MPT-7B-Instruct model and fine-tuning it with an HF dataset for example multi_news dataset for news summarization?

  • How to prepare the dataset/ prompt
  • How to freeze layers and have a small number of trainable parameters (given that you don't have LoRA support yet)

Sign up or log in to comment