nlparabic commited on
Commit
25ea50e
1 Parent(s): 19b7a96

Training in progress, epoch 19

Browse files
Files changed (3) hide show
  1. egy_training_log.txt +149 -0
  2. model.safetensors +1 -1
  3. training_args.bin +1 -1
egy_training_log.txt CHANGED
@@ -321,3 +321,152 @@ INFO:root:Epoch 18.0: Train Loss = 1.4159, Eval Loss = 1.979936957359314
321
  INFO:absl:Using default tokenizer.
322
  INFO:__main__:*** Evaluate ***
323
  INFO:absl:Using default tokenizer.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
321
  INFO:absl:Using default tokenizer.
322
  INFO:__main__:*** Evaluate ***
323
  INFO:absl:Using default tokenizer.
324
+ WARNING:__main__:Process rank: 0, device: cuda:0, n_gpu: 1, distributed training: False, 16-bits training: False
325
+ INFO:__main__:Training/evaluation parameters TrainingArguments(
326
+ _n_gpu=1,
327
+ accelerator_config={'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None, 'use_configured_state': False},
328
+ adafactor=False,
329
+ adam_beta1=0.9,
330
+ adam_beta2=0.999,
331
+ adam_epsilon=1e-08,
332
+ auto_find_batch_size=False,
333
+ batch_eval_metrics=False,
334
+ bf16=False,
335
+ bf16_full_eval=False,
336
+ data_seed=None,
337
+ dataloader_drop_last=False,
338
+ dataloader_num_workers=0,
339
+ dataloader_persistent_workers=False,
340
+ dataloader_pin_memory=True,
341
+ dataloader_prefetch_factor=None,
342
+ ddp_backend=None,
343
+ ddp_broadcast_buffers=None,
344
+ ddp_bucket_cap_mb=None,
345
+ ddp_find_unused_parameters=None,
346
+ ddp_timeout=1800,
347
+ debug=[],
348
+ deepspeed=None,
349
+ disable_tqdm=False,
350
+ dispatch_batches=None,
351
+ do_eval=True,
352
+ do_predict=False,
353
+ do_train=True,
354
+ eval_accumulation_steps=None,
355
+ eval_delay=0,
356
+ eval_do_concat_batches=True,
357
+ eval_on_start=False,
358
+ eval_steps=None,
359
+ eval_strategy=IntervalStrategy.EPOCH,
360
+ eval_use_gather_object=False,
361
+ evaluation_strategy=epoch,
362
+ fp16=False,
363
+ fp16_backend=auto,
364
+ fp16_full_eval=False,
365
+ fp16_opt_level=O1,
366
+ fsdp=[],
367
+ fsdp_config={'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False},
368
+ fsdp_min_num_params=0,
369
+ fsdp_transformer_layer_cls_to_wrap=None,
370
+ full_determinism=False,
371
+ gradient_accumulation_steps=1,
372
+ gradient_checkpointing=False,
373
+ gradient_checkpointing_kwargs=None,
374
+ greater_is_better=False,
375
+ group_by_length=False,
376
+ half_precision_backend=auto,
377
+ hub_always_push=False,
378
+ hub_model_id=None,
379
+ hub_private_repo=False,
380
+ hub_strategy=HubStrategy.EVERY_SAVE,
381
+ hub_token=<HUB_TOKEN>,
382
+ ignore_data_skip=False,
383
+ include_inputs_for_metrics=False,
384
+ include_num_input_tokens_seen=False,
385
+ include_tokens_per_second=False,
386
+ jit_mode_eval=False,
387
+ label_names=None,
388
+ label_smoothing_factor=0.0,
389
+ learning_rate=5e-05,
390
+ length_column_name=length,
391
+ load_best_model_at_end=True,
392
+ local_rank=0,
393
+ log_level=passive,
394
+ log_level_replica=warning,
395
+ log_on_each_node=True,
396
+ logging_dir=/home/iais_marenpielka/Bouthaina/results_fixed/runs/Aug25_15-20-25_lmgpu-node-09,
397
+ logging_first_step=False,
398
+ logging_nan_inf_filter=True,
399
+ logging_steps=500,
400
+ logging_strategy=IntervalStrategy.STEPS,
401
+ lr_scheduler_kwargs={},
402
+ lr_scheduler_type=SchedulerType.LINEAR,
403
+ max_grad_norm=1.0,
404
+ max_steps=-1,
405
+ metric_for_best_model=loss,
406
+ mp_parameters=,
407
+ neftune_noise_alpha=None,
408
+ no_cuda=False,
409
+ num_train_epochs=20.0,
410
+ optim=OptimizerNames.ADAMW_TORCH,
411
+ optim_args=None,
412
+ optim_target_modules=None,
413
+ output_dir=/home/iais_marenpielka/Bouthaina/results_fixed,
414
+ overwrite_output_dir=False,
415
+ past_index=-1,
416
+ per_device_eval_batch_size=8,
417
+ per_device_train_batch_size=8,
418
+ prediction_loss_only=False,
419
+ push_to_hub=True,
420
+ push_to_hub_model_id=None,
421
+ push_to_hub_organization=None,
422
+ push_to_hub_token=<PUSH_TO_HUB_TOKEN>,
423
+ ray_scope=last,
424
+ remove_unused_columns=True,
425
+ report_to=[],
426
+ restore_callback_states_from_checkpoint=False,
427
+ resume_from_checkpoint=None,
428
+ run_name=/home/iais_marenpielka/Bouthaina/results_fixed,
429
+ save_on_each_node=False,
430
+ save_only_model=False,
431
+ save_safetensors=True,
432
+ save_steps=500,
433
+ save_strategy=IntervalStrategy.EPOCH,
434
+ save_total_limit=None,
435
+ seed=42,
436
+ skip_memory_metrics=True,
437
+ split_batches=None,
438
+ tf32=None,
439
+ torch_compile=False,
440
+ torch_compile_backend=None,
441
+ torch_compile_mode=None,
442
+ torch_empty_cache_steps=None,
443
+ torchdynamo=None,
444
+ tpu_metrics_debug=False,
445
+ tpu_num_cores=None,
446
+ use_cpu=False,
447
+ use_ipex=False,
448
+ use_legacy_prediction_loop=False,
449
+ use_mps_device=False,
450
+ warmup_ratio=0.0,
451
+ warmup_steps=500,
452
+ weight_decay=0.0,
453
+ )
454
+ INFO:__main__:Checkpoint detected, resuming training at /home/iais_marenpielka/Bouthaina/results_fixed/checkpoint-8298. To avoid this behavior, change the `--output_dir` or add `--overwrite_output_dir` to train from scratch.
455
+ INFO:datasets.builder:Using custom data configuration default-93ed01be52df6f6e
456
+ INFO:datasets.info:Loading Dataset Infos from /home/iais_marenpielka/Bouthaina/miniconda3/lib/python3.12/site-packages/datasets/packaged_modules/text
457
+ INFO:datasets.builder:Generating dataset text (/home/iais_marenpielka/.cache/huggingface/datasets/text/default-93ed01be52df6f6e/0.0.0/96636a050ef51804b84abbfd4f4ad440e01153c24b86293eb5c3b300a41f9101)
458
+ INFO:datasets.builder:Downloading and preparing dataset text/default to /home/iais_marenpielka/.cache/huggingface/datasets/text/default-93ed01be52df6f6e/0.0.0/96636a050ef51804b84abbfd4f4ad440e01153c24b86293eb5c3b300a41f9101...
459
+ INFO:datasets.download.download_manager:Downloading took 0.0 min
460
+ INFO:datasets.download.download_manager:Checksum Computation took 0.0 min
461
+ INFO:datasets.builder:Generating train split
462
+ INFO:datasets.builder:Generating validation split
463
+ INFO:datasets.utils.info_utils:Unable to verify splits sizes.
464
+ INFO:datasets.builder:Dataset text downloaded and prepared to /home/iais_marenpielka/.cache/huggingface/datasets/text/default-93ed01be52df6f6e/0.0.0/96636a050ef51804b84abbfd4f4ad440e01153c24b86293eb5c3b300a41f9101. Subsequent calls will reuse this data.
465
+ INFO:datasets.arrow_dataset:Caching processed dataset at /home/iais_marenpielka/.cache/huggingface/datasets/text/default-93ed01be52df6f6e/0.0.0/96636a050ef51804b84abbfd4f4ad440e01153c24b86293eb5c3b300a41f9101/cache-4cda59a599643701.arrow
466
+ INFO:datasets.arrow_dataset:Caching processed dataset at /home/iais_marenpielka/.cache/huggingface/datasets/text/default-93ed01be52df6f6e/0.0.0/96636a050ef51804b84abbfd4f4ad440e01153c24b86293eb5c3b300a41f9101/cache-d82ef9a45800c64f.arrow
467
+ WARNING:__main__:The tokenizer picked seems to have a very large `model_max_length` (1000000000000000019884624838656). Using block_size=768 instead. You can change that default value by passing --block_size xxx.
468
+ INFO:datasets.arrow_dataset:Caching processed dataset at /home/iais_marenpielka/.cache/huggingface/datasets/text/default-93ed01be52df6f6e/0.0.0/96636a050ef51804b84abbfd4f4ad440e01153c24b86293eb5c3b300a41f9101/cache-038f8e8385bf6638.arrow
469
+ INFO:datasets.arrow_dataset:Caching processed dataset at /home/iais_marenpielka/.cache/huggingface/datasets/text/default-93ed01be52df6f6e/0.0.0/96636a050ef51804b84abbfd4f4ad440e01153c24b86293eb5c3b300a41f9101/cache-51f1e2b6546273ed.arrow
470
+ WARNING:accelerate.utils.other:Detected kernel version 5.4.0, which is below the recommended minimum of 5.5.0; this can cause the process to hang. It is recommended to upgrade the kernel to the minimum version or higher.
471
+ WARNING:root:Epoch 19.0: No losses recorded yet.
472
+ INFO:absl:Using default tokenizer.
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:eebba2f2ff833170f3124d5c7a39e6a8e26d5c480423329231080b523195e5df
3
  size 539221632
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6ddab2f4bc4af17da079c503fa60da366561c567ff9ddbd6521efa3060c19558
3
  size 539221632
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:07e81171e3677bc3d4ccc43b272fd78fc5a1df3b2eae1e5cb15081224873b8d5
3
  size 5240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:92e8fb7e384c0e60dc19bccc81d5ba66ecd494477a99a468184ef66231010c0a
3
  size 5240