# wizardLM-LlaMA-LoRA-7B A LoRA trained on the WizardLM dataset, with a LlaMA 7B as the basemodel. ## Instruction example It was trained with the alpaca-short template, without any inputs, so prompt as follows: ``` ### Instruction: Write a poem about the transformers Python library. Mention the word "large language models" in that poem. ### Response: I'm not sure if this is what you meant, but here goes! The Transformers are large language models that help us make sense of text. They take our sentences and turn them into vectors, which can be used to find similarities between texts. We use these for things like search engines or spam filters; they also have uses in machine learning too. ``` ## Trained with the following params ``` base_model: /root/alpaca-lora/llama-7b-hf data_path: victor123/evol_instruct_70k output_dir: /loras/wizardLM-lama-lora batch_size: 64 micro_batch_size: 8 num_epochs: 3 learning_rate: 2e-05 cutoff_len: 2048 val_set_size: 2000 lora_r: 16 lora_alpha: 16 lora_dropout: 0.05 lora_target_modules: ['q_proj', 'k_proj', 'v_proj', 'o_proj'] train_on_inputs: True add_eos_token: False group_by_length: True wandb_project: wandb_run_name: wandb_watch: wandb_log_model: resume_from_checkpoint: False prompt template: alpaca_short ``` ## Training Details - Trained with https://github.com/tloen/alpaca-lora. Note: ince the `victor123/evol_instruct_70k` dataset only contains instruction and output, comment out the line `data_point["input"],` around line 151 in `alpaca-lora\finetune.py` - Trained on [RunPod](https://runpod.io?ref=qgrfwczf ) community cloud with 1x A100 80GB vram (Note: less GPU was needed) - Took 14:47:39 to train 3 epochs - Cost around $37 to train ## Evaluation - No evaluation has been done on this model. If someone wants to share I would happily pull. - Empirically it looks promising for complex instruction following.