Spaces:
Sleeping
Sleeping
File size: 16,155 Bytes
17c1e65 0824054 17c1e65 0824054 17c1e65 18c14e3 17c1e65 0c4272d 17c1e65 9ba7b62 17c1e65 9485a43 17c1e65 9485a43 17c1e65 9485a43 17c1e65 9485a43 17c1e65 9485a43 17c1e65 9485a43 17c1e65 9485a43 17c1e65 9485a43 17c1e65 9485a43 17c1e65 9485a43 17c1e65 9485a43 17c1e65 9485a43 17c1e65 9485a43 17c1e65 9485a43 17c1e65 9485a43 17c1e65 9485a43 17c1e65 9485a43 17c1e65 9485a43 17c1e65 9485a43 17c1e65 9485a43 17c1e65 9485a43 17c1e65 9485a43 17c1e65 9485a43 17c1e65 9485a43 17c1e65 9485a43 17c1e65 9485a43 17c1e65 9485a43 17c1e65 18c14e3 17c1e65 9485a43 17c1e65 18c14e3 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 |
# Main Fine-Tuning Script for LLaMA-2-chat
# This script is the central executable for fine-tuning large language models, specifically designed for the LLaMA2 Chat
# model.
# It encompasses the entire process of fine-tuning, starting from data preparation to the final model training.
# The script leverages the 'FinetuningDataHandler' class for data loading, inspection, preparation, and splitting.
# This ensures that the dataset is correctly processed and prepared for effective training.
# The fine-tuning process is managed by the Finetuner class, which handles the training of the model using specific
# training arguments and datasets. Advanced configurations for Quantized Low-Rank Adaptation (QLoRA) and Parameter
# Efficient Fine-Tuning (PEFT) are utilized to optimize the training process on limited hardware resources.
# The script is designed to be executed as a standalone process, providing an end-to-end solution for fine-tuning
# LLMs. It is a part of a larger project aimed at optimizing the performance of language model to adapt to
# OK-VQA dataset.
# Ensure all dependencies are installed and the required files are in place before running this script.
# The configurations for the fine-tuning process are defined in the 'my_model/config/fine_tuning_config.py' file.
# ---------- Please run this file for the full fine-tuning process to start ----------#
# ---------- Please ensure this is run on a GPU ----------#
import torch
from transformers import AutoModelForCausalLM, AutoTokenizer, TrainingArguments, TRANSFORMERS_CACHE
from trl import SFTTrainer
from datasets import Dataset, load_dataset
from peft import LoraConfig, prepare_model_for_kbit_training, PeftModel
import my_model.config.fine_tuning_config as config
from typing import List
import bitsandbytes # only on GPU
import gc
import os
import shutil
from my_model.LLAMA2.LLAMA2_model import Llama2ModelManager
from fine_tuning_data_handler import FinetuningDataHandler
class QLoraConfig:
"""
Configures QLoRA (Quantized Low-Rank Adaptation) parameters for efficient model fine-tuning.
LoRA allows adapting large language models with a minimal number of trainable parameters.
Attributes:
lora_config (LoraConfig): Configuration object for LoRA parameters.
"""
def __init__(self) -> None:
"""
Initializes QLoraConfig with specific LoRA parameters.
"""
# please refer to config file 'my_model/config/fine_tuning_config.py' for QLORA arguments description.
self.lora_config = LoraConfig(
lora_alpha=config.LORA_ALPHA,
lora_dropout=config.LORA_DROPOUT,
r=config.LORA_R,
bias="none", # bias is already accounted for in LLAMA2 pre-trained model layers.
task_type="CAUSAL_LM",
target_modules=['up_proj', 'down_proj', 'k_proj', 'q_proj', 'v_proj', 'o_proj'] # modules for fine-tuning.
)
class Finetuner:
"""
The Finetuner class manages the fine-tuning process of a pre-trained language model using specific
training arguments and datasets. It is designed to adapt a pre-trained model on a specific dataset
to enhance its performance on similar data.
This class not only facilitates the fine-tuning of LLAMA2 but also includes advanced
resource management capabilities. It provides methods for deleting model and trainer objects,
clearing GPU memory, and cleaning up Hugging Face's Transformers cache. These functionalities
make the Finetuner class especially useful in environments with limited computational resources
or when managing multiple models or training sessions.
Additionally, the class supports configurations for Quantized Low-Rank Adaptation (QLoRA)
to fine-tune models with minimal trainable parameters, and Parameter Efficient Fine-Tuning (PEFT)
for training efficiency on limited hardware.
Attributes:
base_model (AutoModelForCausalLM): The pre-trained language model to be fine-tuned.
tokenizer (AutoTokenizer): The tokenizer associated with the model.
train_dataset (Dataset): The dataset used for training.
eval_dataset (Dataset): The dataset used for evaluation.
training_arguments (TrainingArguments): Configuration for training the model.
Key Methods:
- load_LLAMA2_for_finetuning: Loads the LLAMA2 model and tokenizer for fine-tuning.
- train: Trains the model using PEFT configuration.
- delete_model: Deletes a specified model attribute.
- delete_trainer: Deletes a specified trainer object.
- clear_training_resources: Clears GPU memory.
- clear_cache_and_collect_garbage: Clears Transformers cache and performs garbage collection.
- find_all_linear_names: Identifies linear layer names suitable for LoRA application.
- print_trainable_parameters: Prints the number of trainable parameters in the model.
"""
def __init__(self, train_dataset: Dataset, eval_dataset: Dataset) -> None:
"""
Initializes the Finetuner class with the model, tokenizer, and datasets.
Args:
train_dataset (Dataset): The dataset for training the model.
eval_dataset (Dataset): The dataset for evaluating the model.
"""
self.base_model, self.tokenizer = self.load_LLAMA2_for_finetuning()
self.merged_model = None
self.train_dataset = train_dataset
self.eval_dataset = eval_dataset
# please refer to config file 'my_model/config/fine_tuning_config.py' for training arguments description.
self.training_arguments = TrainingArguments(
output_dir=config.OUTPUT_DIR,
num_train_epochs=config.NUM_TRAIN_EPOCHS,
per_device_train_batch_size=config.PER_DEVICE_TRAIN_BATCH_SIZE,
per_device_eval_batch_size=config.PER_DEVICE_EVAL_BATCH_SIZE,
gradient_accumulation_steps=config.GRADIENT_ACCUMULATION_STEPS,
fp16=config.FP16,
bf16=config.BF16,
evaluation_strategy=config.Evaluation_STRATEGY,
eval_steps=config.EVALUATION_STEPS,
max_grad_norm=config.MAX_GRAD_NORM,
learning_rate=config.LEARNING_RATE,
weight_decay=config.WEIGHT_DECAY,
optim=config.OPTIM,
lr_scheduler_type=config.LR_SCHEDULER_TYPE,
max_steps=config.MAX_STEPS,
warmup_ratio=config.WARMUP_RATIO,
group_by_length=config.GROUP_BY_LENGTH,
save_steps=config.SAVE_STEPS,
logging_steps=config.LOGGING_STEPS,
report_to="tensorboard"
)
def load_LLAMA2_for_finetuning(self) -> Tuple[AutoModelForCausalLM, AutoTokenizer]:
"""
Loads the LLAMA2 model and tokenizer, specifically configured for fine-tuning.
Returns:
Tuple[AutoModelForCausalLM, AutoTokenizer]: The loaded model and tokenizer.
"""
llm_manager = Llama2ModelManager()
base_model, tokenizer = llm_manager.load_model_and_tokenizer(for_fine_tuning=True)
return base_model, tokenizer
def find_all_linear_names(self) -> List[str]:
"""
Identifies all linear layer names in the model that are suitable for applying LoRA.
Returns:
List[str]: A list of linear layer names.
"""
cls = bitsandbytes.nn.Linear4bit
lora_module_names = set()
for name, module in self.base_model.named_modules():
if isinstance(module, cls):
names = name.split('.')
lora_module_names.add(names[0] if len(names) == 1 else names[-1])
# We dont want to train these two modules to avoid computational overhead.
lora_module_names -= {'lm_head', 'gate_proj'}
return list(lora_module_names)
def print_trainable_parameters(self, use_4bit: bool = False) -> None:
"""
Calculates and prints the number of trainable parameters in the model.
Args:
use_4bit (bool): If true, calculates the parameter count considering 4-bit quantization.
Returns:
List[str]: None.
"""
trainable_params = sum(p.numel() for p in self.base_model.parameters() if p.requires_grad)
if use_4bit:
trainable_params /= 2
total_params = sum(p.numel() for p in self.base_model.parameters())
print(f"All Parameters: {total_params:,d} || Trainable Parameters: {trainable_params:,d} "
f"|| Trainable Parameters %: {100 * trainable_params / total_params:.2f}%")
def train(self, peft_config: LoraConfig) -> None:
"""
Trains the model using the specified PEFT (Progressive Effort Fine-Tuning) configuration.
Args:
peft_config (LoraConfig): Configuration for the PEFT training process.
Returns:
List[str]: None.
"""
self.base_model.config.use_cache = False
# Set the pretraining_tp flag to 1 to enable the use of LoRA (Low-Rank Adapters) layers.
self.base_model.config.pretraining_tp = 1
# Prepare the model for k-bit training by quantizing the weights to 4 bits using bitsandbytes.
self.base_model = prepare_model_for_kbit_training(self.base_model)
self.trainer = SFTTrainer(
model=self.base_model,
train_dataset=self.train_dataset,
eval_dataset=self.eval_dataset,
peft_config=peft_config,
dataset_text_field='text',
max_seq_length=config.MAX_TOKEN_COUNT,
tokenizer=self.tokenizer,
args=self.training_arguments,
packing=config.PACKING
)
self.trainer.train()
def save_model(self) -> None:
"""
Saves the fine-tuned model to the specified directory.
This method saves the model weights and configuration of the fine-tuned model.
The save directory and filename are determined by the configuration provided in
the 'fine_tuning_config.py' file. It is useful for persisting the fine-tuned model
for later use or evaluation.
The saved model can be easily loaded using Hugging Face's model loading utilities.
Returns:
None
"""
self.fine_tuned_adapter_name = config.ADAPTER_SAVE_NAME
self.trainer.model.save_pretrained(self.fine_tuned_adapter_name)
def merge_weights(self) -> None:
"""
Merges the weights of the fine-tuned adapter with the base model.
This method integrates the fine-tuned adapter weights into the base model,
resulting in a single consolidated model. The merged model can then be used
for inference or further training.
After merging, the weights of the adapter are no longer separate from the
base model, enabling more efficient storage and deployment. The merged model
is stored in the 'self.merged_model' attribute of the Finetuner class.
Returns:
None
"""
self.merged_model = PeftModel.from_pretrained(self.base_model, self.fine_tuned_adapter_name)
self.merged_model = self.merged_model.merge_and_unload()
def delete_model(self, model_name: str) -> None:
"""
Deletes a specified model attribute.
Args:
model_name (str): The name of the model attribute to delete.
Returns:
None
"""
try:
if hasattr(self, model_name) and getattr(self, model_name) is not None:
delattr(self, model_name)
print(f"Model '{model_name}' has been deleted.")
else:
print(f"Warning: Model '{model_name}' has already been cleared or does not exist.")
except Exception as e:
print(f"Error occurred while deleting model '{model_name}': {str(e)}")
def delete_trainer(self, trainer_name: str) -> None:
"""
Deletes a specified trainer object.
Args:
trainer_name (str): The name of the trainer object to delete.
Returns:
None
"""
try:
if hasattr(self, trainer_name) and getattr(self, trainer_name) is not None:
delattr(self, trainer_name)
print(f"Trainer object '{trainer_name}' has been deleted.")
else:
print(f"Warning: Trainer object '{trainer_name}' has already been cleared or does not exist.")
except Exception as e:
print(f"Error occurred while deleting trainer object '{trainer_name}': {str(e)}")
def clear_training_resources(self) -> None:
"""
Clears GPU memory.
Returns:
None
"""
try:
if torch.cuda.is_available():
torch.cuda.empty_cache()
print("GPU memory has been cleared.")
except Exception as e:
print(f"Error occurred while clearing GPU memory: {str(e)}")
def clear_cache_and_collect_garbage(self) -> None:
"""
Clears Hugging Face's Transformers cache and runs garbage collection.
Returns:
None
"""
try:
if os.path.exists(TRANSFORMERS_CACHE):
shutil.rmtree(TRANSFORMERS_CACHE, ignore_errors=True)
print("Transformers cache has been cleared.")
gc.collect()
print("Garbage collection has been executed.")
except Exception as e:
print(f"Error occurred while clearing cache and collecting garbage: {str(e)}")
def fine_tune(save_fine_tuned_adapter: bool = False, merge: bool = False, delete_trainer_after_fine_tune: bool = False) -> AutoModelForCausalLM:
"""
Conducts the fine-tuning process of a pre-trained language model using specified configurations.
This function encompasses the complete workflow of fine-tuning, including data handling, training,
and optional steps like saving the fine-tuned model and merging weights.
The function initiates by preparing the training and evaluation datasets using the `FinetuningDataHandler`.
It then sets up the QLoRA configuration for the fine-tuning process. The actual training is carried out by
the `Finetuner` class. Post training, based on the arguments, the function can save the fine-tuned model,
merge the adapter weights with the base model, and clean up resources by deleting the trainer object.
Args:
save_fine_tuned_adapter (bool): If True, saves the fine-tuned adapter after training.
merge (bool): If True, merges the weights of the fine-tuned adapter into the base model.
delete_trainer_after_fine_tune (bool): If True, deletes the trainer object after fine-tuning to free up resources.
Returns:
AutoModelForCausalLM: The fine-tuned model after the fine-tuning process. This could be either the merged model
or the trained model based on the provided arguments.
"""
data_handler = FinetuningDataHandler()
fine_tuning_data_train, fine_tuning_data_eval = data_handler.inspect_prepare_split_data()
qlora = QLoraConfig()
peft_config = qlora.lora_config
tuner = Finetuner(fine_tuning_data_train, fine_tuning_data_eval)
tuner.train(peft_config=peft_config)
if save_fine_tuned_adapter:
tuner.save_model()
if merge:
tuner.merge_weights()
if delete_trainer_after_fine_tune:
tuner.delete_trainer("trainer")
tuner.delete_model("base_model") # We always delete this as it is not required after the merger.
if save_fine_tuned_adapter:
tuner.save_model()
if tuner.merged_model is not None:
return tuner.merged_model
else:
return tuner.trainer.model
if __name__ == "__main__":
# uncomment to run the fine-tuning process.
#fine_tune()
|