metadata
language: pt
library_name: peft
datasets: Weni/zeroshot-3.0.3
pipeline_tag: zero-shot-classification
training_arguments:
output_dir: >-
./Zeroshot/01-12-23-NousResearch-Nous-Hermes-Llama2-13b_multilang-dataset-3.0.3-portuguese-2_epochs-10_batch_2/checkpoints/
overwrite_output_dir: false
do_train: false
do_eval: true
do_predict: false
evaluation_strategy: epoch
prediction_loss_only: false
per_device_train_batch_size: 2
per_device_eval_batch_size: 8
gradient_accumulation_steps: 2
eval_accumulation_steps: 1
eval_delay: 0
learning_rate: 0.0004
weight_decay: 0.01
adam_beta1: 0.9
adam_beta2: 0.999
adam_epsilon: 1.e-8
max_grad_norm: 0.3
num_train_epochs: 10
max_steps: -1
lr_scheduler_type: cosine
warmup_ratio: 0.1
warmup_steps: 0
log_level: passive
log_level_replica: warning
log_on_each_node: true
logging_dir: >-
./Zeroshot/01-12-23-NousResearch-Nous-Hermes-Llama2-13b_multilang-dataset-3.0.3-portuguese-2_epochs-10_batch_2/checkpoints/runs/Dec01_21-53-07_fd10189bb234
logging_strategy: steps
logging_first_step: false
logging_steps: 500
logging_nan_inf_filter: true
save_strategy: epoch
save_steps: 500
save_total_limit: 5
save_safetensors: true
save_on_each_node: false
no_cuda: false
use_mps_device: false
seed: 42
jit_mode_eval: false
use_ipex: false
bf16: false
fp16: true
fp16_opt_level: O1
half_precision_backend: auto
bf16_full_eval: false
fp16_full_eval: false
local_rank: 0
tpu_metrics_debug: false
debug: []
dataloader_drop_last: false
dataloader_num_workers: 0
past_index: -1
run_name: >-
./Zeroshot/01-12-23-NousResearch-Nous-Hermes-Llama2-13b_multilang-dataset-3.0.3-portuguese-2_epochs-10_batch_2/checkpoints/
disable_tqdm: false
remove_unused_columns: true
load_best_model_at_end: true
metric_for_best_model: eval_loss
greater_is_better: false
ignore_data_skip: false
sharded_ddp: []
fsdp: []
fsdp_min_num_params: 0
fsdp_config:
fsdp_min_num_params: 0
xla: false
xla_fsdp_grad_ckpt: false
label_smoothing_factor: 0
optim: adamw_torch
adafactor: false
group_by_length: false
length_column_name: length
report_to:
- tensorboard
dataloader_pin_memory: true
skip_memory_metrics: true
use_legacy_prediction_loop: false
push_to_hub: true
hub_model_id: Weni/ZeroShot-2.2.1-Llama2-13b-Multilanguage-3.0.3
hub_strategy: all_checkpoints
hub_token: <HUB_TOKEN>
hub_private_repo: false
gradient_checkpointing: true
include_inputs_for_metrics: false
fp16_backend: auto
push_to_hub_token: <PUSH_TO_HUB_TOKEN>
mp_parameters: ''
auto_find_batch_size: false
full_determinism: false
ray_scope: last
ddp_timeout: 1800
torch_compile: false
datatset:
name: Weni/zeroshot-3.0.3
Training Procedure:
Training Hyperparameters:
output_dir: >-
./Zeroshot/01-12-23-NousResearch-Nous-Hermes-Llama2-13b_multilang-dataset-3.0.3-portuguese-2_epochs-10_batch_2/checkpoints/
overwrite_output_dir: false
do_train: false
do_eval: true
do_predict: false
evaluation_strategy: epoch
prediction_loss_only: false
per_device_train_batch_size: 2
per_device_eval_batch_size: 8
gradient_accumulation_steps: 2
eval_accumulation_steps: 1
eval_delay: 0
learning_rate: 0.0004
weight_decay: 0.01
adam_beta1: 0.9
adam_beta2: 0.999
adam_epsilon: 1.e-8
max_grad_norm: 0.3
num_train_epochs: 10
max_steps: -1
lr_scheduler_type: cosine
warmup_ratio: 0.1
warmup_steps: 0
log_level: passive
log_level_replica: warning
log_on_each_node: true
logging_dir: >-
./Zeroshot/01-12-23-NousResearch-Nous-Hermes-Llama2-13b_multilang-dataset-3.0.3-portuguese-2_epochs-10_batch_2/checkpoints/runs/Dec01_21-53-07_fd10189bb234
logging_strategy: steps
logging_first_step: false
logging_steps: 500
logging_nan_inf_filter: true
save_strategy: epoch
save_steps: 500
save_total_limit: 5
save_safetensors: true
save_on_each_node: false
no_cuda: false
use_mps_device: false
seed: 42
jit_mode_eval: false
use_ipex: false
bf16: false
fp16: true
fp16_opt_level: O1
half_precision_backend: auto
bf16_full_eval: false
fp16_full_eval: false
local_rank: 0
tpu_metrics_debug: false
debug: []
dataloader_drop_last: false
dataloader_num_workers: 0
past_index: -1
run_name: >-
./Zeroshot/01-12-23-NousResearch-Nous-Hermes-Llama2-13b_multilang-dataset-3.0.3-portuguese-2_epochs-10_batch_2/checkpoints/
disable_tqdm: false
remove_unused_columns: true
load_best_model_at_end: true
metric_for_best_model: eval_loss
greater_is_better: false
ignore_data_skip: false
sharded_ddp: []
fsdp: []
fsdp_min_num_params: 0
fsdp_config:
fsdp_min_num_params: 0
xla: false
xla_fsdp_grad_ckpt: false
label_smoothing_factor: 0
optim: adamw_torch
adafactor: false
group_by_length: false
length_column_name: length
report_to:
- tensorboard
dataloader_pin_memory: true
skip_memory_metrics: true
use_legacy_prediction_loop: false
push_to_hub: true
hub_model_id: Weni/ZeroShot-2.2.1-Llama2-13b-Multilanguage-3.0.3
hub_strategy: all_checkpoints
hub_token: <HUB_TOKEN>
hub_private_repo: false
gradient_checkpointing: true
include_inputs_for_metrics: false
fp16_backend: auto
push_to_hub_token: <PUSH_TO_HUB_TOKEN>
mp_parameters: ''
auto_find_batch_size: false
full_determinism: false
ray_scope: last
ddp_timeout: 1800
torch_compile: false
Training data:
name: Weni/zeroshot-3.0.3
Training processing: |-
dataset = dataset.shuffle(seed=55)
dataset = dataset['train'].train_test_split(test_size=0.1)
training_regime: >-
### Training Hyperparameters- output_dir:
./Zeroshot/01-12-23-NousResearch-Nous-Hermes-Llama2-13b_multilang-dataset-3.0.3-portuguese-2_epochs-10_batch_2/checkpoints/
- overwrite_output_dir: False
- do_train: False
- do_eval: True
- do_predict: False
- evaluation_strategy: epoch
- prediction_loss_only: False
- per_device_train_batch_size: 2
- per_device_eval_batch_size: 8
- per_gpu_train_batch_size: None
- per_gpu_eval_batch_size: None
- gradient_accumulation_steps: 2
- eval_accumulation_steps: 1
- eval_delay: 0
- learning_rate: 0.0004
- weight_decay: 0.01
- adam_beta1: 0.9
- adam_beta2: 0.999
- adam_epsilon: 1e-08
- max_grad_norm: 0.3
- num_train_epochs: 10
- max_steps: -1
- lr_scheduler_type: cosine
- warmup_ratio: 0.1
- warmup_steps: 0
- log_level: passive
- log_level_replica: warning
- log_on_each_node: True
- logging_dir:
./Zeroshot/01-12-23-NousResearch-Nous-Hermes-Llama2-13b_multilang-dataset-3.0.3-portuguese-2_epochs-10_batch_2/checkpoints/runs/Dec01_21-53-07_fd10189bb234
- logging_strategy: steps
- logging_first_step: False
- logging_steps: 500
- logging_nan_inf_filter: True
- save_strategy: epoch
- save_steps: 500
- save_total_limit: 5
- save_safetensors: True
- save_on_each_node: False
- no_cuda: False
- use_mps_device: False
- seed: 42
- data_seed: None
- jit_mode_eval: False
- use_ipex: False
- bf16: False
- fp16: True
- fp16_opt_level: O1
- half_precision_backend: auto
- bf16_full_eval: False
- fp16_full_eval: False
- tf32: None
- local_rank: 0
- ddp_backend: None
- tpu_num_cores: None
- tpu_metrics_debug: False
- debug: []
- dataloader_drop_last: False
- eval_steps: None
- dataloader_num_workers: 0
- past_index: -1
- run_name:
./Zeroshot/01-12-23-NousResearch-Nous-Hermes-Llama2-13b_multilang-dataset-3.0.3-portuguese-2_epochs-10_batch_2/checkpoints/
- disable_tqdm: False
- remove_unused_columns: True
- label_names: None
- load_best_model_at_end: True
- metric_for_best_model: eval_loss
- greater_is_better: False
- ignore_data_skip: False
- sharded_ddp: []
- fsdp: []
- fsdp_min_num_params: 0
- fsdp_config: {'fsdp_min_num_params': 0, 'xla': False, 'xla_fsdp_grad_ckpt':
False}
- fsdp_transformer_layer_cls_to_wrap: None
- deepspeed: None
- label_smoothing_factor: 0.0
- optim: adamw_torch
- optim_args: None
- adafactor: False
- group_by_length: False
- length_column_name: length
- report_to: ['tensorboard']
- ddp_find_unused_parameters: None
- ddp_bucket_cap_mb: None
- ddp_broadcast_buffers: None
- dataloader_pin_memory: True
- skip_memory_metrics: True
- use_legacy_prediction_loop: False
- push_to_hub: True
- resume_from_checkpoint: None
- hub_model_id: Weni/ZeroShot-2.2.1-Llama2-13b-Multilanguage-3.0.3
- hub_strategy: all_checkpoints
- hub_token: <HUB_TOKEN>
- hub_private_repo: False
- gradient_checkpointing: True
- include_inputs_for_metrics: False
- fp16_backend: auto
- push_to_hub_model_id: None
- push_to_hub_organization: None
- push_to_hub_token: <PUSH_TO_HUB_TOKEN>
- mp_parameters:
- auto_find_batch_size: False
- full_determinism: False
- torchdynamo: None
- ray_scope: last
- ddp_timeout: 1800
- torch_compile: False
- torch_compile_backend: None
- torch_compile_mode: None
- xpu_backend: None
training_data:
name: Weni/zeroshot-3.0.3
'preprocessing ': |-
dataset = dataset.shuffle(seed=55)
dataset = dataset['train'].train_test_split(test_size=0.1)
preprocessing: |-
dataset = dataset.shuffle(seed=55)
dataset = dataset['train'].train_test_split(test_size=0.1)
base_model: NousResearch/Nous-Hermes-Llama2-13b
Training Hyperparameters
- evaluation_strategy: epoch
- prediction_loss_only: False
- per_device_train_batch_size: 2
- per_device_eval_batch_size: 8
- per_gpu_train_batch_size: None
- per_gpu_eval_batch_size: None
- gradient_accumulation_steps: 2
- eval_accumulation_steps: 1
- eval_delay: 0
- learning_rate: 0.0004
- weight_decay: 0.01
- adam_beta1: 0.9
- adam_beta2: 0.999
- adam_epsilon: 1e-08
- max_grad_norm: 0.3
- num_train_epochs: 10
- max_steps: -1
- lr_scheduler_type: cosine
- warmup_ratio: 0.1
- warmup_steps: 0
- log_level: passive
- log_level_replica: warning
- log_on_each_node: True
- logging_strategy: steps
- logging_first_step: False
- logging_steps: 500
- logging_nan_inf_filter: True
- save_strategy: epoch
- save_steps: 500
- save_total_limit: 5
- save_safetensors: True
- save_on_each_node: False
- no_cuda: False
- use_mps_device: False
- seed: 42
- data_seed: None
- jit_mode_eval: False
- use_ipex: False
- bf16: False
- fp16: True
- fp16_opt_level: O1
- half_precision_backend: auto
- bf16_full_eval: False
- fp16_full_eval: False
- tf32: None
- local_rank: 0
- ddp_backend: None
- tpu_num_cores: None
- tpu_metrics_debug: False
- debug: []
- dataloader_drop_last: False
- eval_steps: None
- dataloader_num_workers: 0
- past_index: -1
- remove_unused_columns: True
- label_names: None
- load_best_model_at_end: True
- metric_for_best_model: eval_loss
- greater_is_better: False
- ignore_data_skip: False
- sharded_ddp: []
- fsdp: []
- fsdp_min_num_params: 0
- fsdp_config: {'fsdp_min_num_params': 0, 'xla': False, 'xla_fsdp_grad_ckpt': False}
- fsdp_transformer_layer_cls_to_wrap: None
- deepspeed: None
- label_smoothing_factor: 0.0
- optim: adamw_torch
- optim_args: None
- adafactor: False
- group_by_length: False
- ddp_find_unused_parameters: None
- ddp_bucket_cap_mb: None
- ddp_broadcast_buffers: None
- dataloader_pin_memory: True
- skip_memory_metrics: True
- use_legacy_prediction_loop: False
- push_to_hub: True
- resume_from_checkpoint: None
- hub_strategy: all_checkpoints
- gradient_checkpointing: True
Training procedure
The following bitsandbytes
quantization config was used during training:
load_in_8bit: False
load_in_4bit: True
llm_int8_threshold: 6.0
llm_int8_skip_modules: None
llm_int8_enable_fp32_cpu_offload: False
llm_int8_has_fp16_weight: False
bnb_4bit_quant_type: nf4
bnb_4bit_use_double_quant: True
bnb_4bit_compute_dtype: bfloat16
Framework versions
- PEFT 0.4.0