96abhishekarora commited on
Commit
2065bfc
1 Parent(s): 1b4230e

Modified validation and training for linktransformer model

Browse files
Files changed (4) hide show
  1. .gitattributes +1 -0
  2. LT_training_config.json +2 -2
  3. README.md +6 -6
  4. pytorch_model.bin +1 -1
.gitattributes CHANGED
@@ -34,3 +34,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
  pytorch_model.bin filter=lfs diff=lfs merge=lfs -text
 
 
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
  pytorch_model.bin filter=lfs diff=lfs merge=lfs -text
37
+ .git/lfs/objects/8a/f5/8af50c97be5450dea1d079b45919e05e992d4d72babaa4c09a7d64793982cd3e filter=lfs diff=lfs merge=lfs -text
LT_training_config.json CHANGED
@@ -1,13 +1,13 @@
1
  {
2
  "model_save_dir": "models",
3
  "model_save_name": "linkage_kn_en_familyname",
4
- "opt_model_description": "This model was trained on a dataset consisting of 8138064 people and their family id \n It was trained for 10 epochs using other defaults that can be found in the repo's LinkTransformer config file - LT_training_config.json \n ",
5
  "opt_model_lang": [
6
  "kn",
7
  "en"
8
  ],
9
  "train_batch_size": 64,
10
- "num_epochs": 10,
11
  "warm_up_perc": 1,
12
  "learning_rate": 2e-06,
13
  "val_perc": 0.01,
 
1
  {
2
  "model_save_dir": "models",
3
  "model_save_name": "linkage_kn_en_familyname",
4
+ "opt_model_description": "This model was trained on a dataset consisting of 12105132 people and their family id. 50% of the names are alo transliterated. \n It was trained for 6 epochs using other defaults that can be found in the repo's LinkTransformer config file - LT_training_config.json \n ",
5
  "opt_model_lang": [
6
  "kn",
7
  "en"
8
  ],
9
  "train_batch_size": 64,
10
+ "num_epochs": 6,
11
  "warm_up_perc": 1,
12
  "learning_rate": 2e-06,
13
  "val_perc": 0.01,
README.md CHANGED
@@ -24,8 +24,8 @@ This model has been fine-tuned on the model : bert-base-multilingual-cased. It i
24
  - en.
25
 
26
 
27
- This model was trained on a dataset consisting of 8138064 people and their family id
28
- It was trained for 10 epochs using other defaults that can be found in the repo's LinkTransformer config file - LT_training_config.json
29
 
30
 
31
  ## Usage (LinkTransformer)
@@ -99,7 +99,7 @@ The model was trained with the parameters:
99
 
100
  **DataLoader**:
101
 
102
- `torch.utils.data.dataloader.DataLoader` of length 113924 with parameters:
103
  ```
104
  {'batch_size': 64, 'sampler': 'torch.utils.data.dataloader._InfiniteConstantSampler', 'batch_sampler': 'torch.utils.data.sampler.BatchSampler'}
105
  ```
@@ -111,8 +111,8 @@ The model was trained with the parameters:
111
  Parameters of the fit()-Method:
112
  ```
113
  {
114
- "epochs": 10,
115
- "evaluation_steps": 11393,
116
  "evaluator": "sentence_transformers.evaluation.SequentialEvaluator.SequentialEvaluator",
117
  "max_grad_norm": 1,
118
  "optimizer_class": "<class 'torch.optim.adamw.AdamW'>",
@@ -121,7 +121,7 @@ Parameters of the fit()-Method:
121
  },
122
  "scheduler": "WarmupLinear",
123
  "steps_per_epoch": null,
124
- "warmup_steps": 1139240,
125
  "weight_decay": 0.01
126
  }
127
  ```
 
24
  - en.
25
 
26
 
27
+ This model was trained on a dataset consisting of 12105132 people and their family id. 50% of the names are alo transliterated.
28
+ It was trained for 6 epochs using other defaults that can be found in the repo's LinkTransformer config file - LT_training_config.json
29
 
30
 
31
  ## Usage (LinkTransformer)
 
99
 
100
  **DataLoader**:
101
 
102
+ `torch.utils.data.dataloader.DataLoader` of length 186000 with parameters:
103
  ```
104
  {'batch_size': 64, 'sampler': 'torch.utils.data.dataloader._InfiniteConstantSampler', 'batch_sampler': 'torch.utils.data.sampler.BatchSampler'}
105
  ```
 
111
  Parameters of the fit()-Method:
112
  ```
113
  {
114
+ "epochs": 6,
115
+ "evaluation_steps": 18600,
116
  "evaluator": "sentence_transformers.evaluation.SequentialEvaluator.SequentialEvaluator",
117
  "max_grad_norm": 1,
118
  "optimizer_class": "<class 'torch.optim.adamw.AdamW'>",
 
121
  },
122
  "scheduler": "WarmupLinear",
123
  "steps_per_epoch": null,
124
+ "warmup_steps": 1116000,
125
  "weight_decay": 0.01
126
  }
127
  ```
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8af50c97be5450dea1d079b45919e05e992d4d72babaa4c09a7d64793982cd3e
3
  size 711477737
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b729ce841823b35b8245111c07d098e2efaa1e7a93d2e5fe85d5de2bd7806181
3
  size 711477737