Thalesian commited on
Commit
a63b0e0
β€’
1 Parent(s): 68062e3

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +52 -7
README.md CHANGED
@@ -6,27 +6,72 @@ model-index:
6
  results: []
7
  ---
8
 
9
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
10
- should probably proofread and complete it, then remove this comment. -->
11
 
12
- # t5-small-p-l-akk-en-20240910-174859
 
13
 
14
- This model was trained from scratch on the None dataset.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
15
 
16
  ## Model description
17
 
18
- More information needed
 
 
 
 
 
 
 
 
 
 
 
 
 
 
19
 
20
  ## Intended uses & limitations
21
 
22
- More information needed
23
 
24
  ## Training and evaluation data
25
 
26
- More information needed
27
 
28
  ## Training procedure
29
 
 
 
 
 
 
 
 
 
 
30
  ### Training hyperparameters
31
 
32
  The following hyperparameters were used during training:
 
6
  results: []
7
  ---
8
 
9
+ A model for translating cuneiform to english using [google's t5-small](https://huggingface.co/google-t5/t5-small) as a baseline.
 
10
 
11
+ - Akkadian: π’„Ώ π’ˆΎ π’Œ— π’ƒΆ π’Œ“ 𒐉 π’†š π’€€ π’ˆΎ 𒆳 π’†Έ π’„­ 𒇻 𒁺 π’…… 𒆳 𒁀 π’€€ 𒍝 𒆳 π’Š“ π’…ˆ 𒁀 π’‡· π’€€ 𒆳 𒁲 𒁺 π’€€ π’†· π’€€ 𒁲 π’Œ· π’ˆ¨ π’Œ π’‰Œ 𒃻 π’…† 𒁲 π’€€ 𒇉 π’Š’ π’Œ‘ π’Š’ π’Š­ 𒆳 π’ˆ¨ π’„΄ π’Š‘ 𒀝 π’‹€ π’Š© π’†· π’‹’ 𒉑 𒃻 π’‹— π’ˆ¨ π’Œ π’‹— 𒉑 π’Œ‘ π’ŠΊ 𒍝 π’€€ π’€€ π’ˆΎ π’Œ· π’…€ π’€Έ π’‹© π’Œ’ π’†·'
12
+ - English: 'in the month kislimu the fourth day i marched to the land habhu i conquered the lands bazu sarbaliu and didualu together with the cities on the banks of the river ruru of the land mehru i brought forth their booty and possessions and brought them to my city assur' Prediction: 'in the mo nth tammuz iv i conquered the land s que and que i conquered the land s que and bi t yakin i conquered the cities f ro m the river i conquered and plundered the cities on the bo rd er of the land elam'
13
 
14
+ Note that the training loss does not reflect full training - this model was trained at expanding context sizes (56 -> 512) restricted to complete sequences. It was trained on cuneiform -> English, transliteration, and grouping in both directions to reinforce itself. It is an instruct model, so it requires a request to intepret data.
15
+
16
+
17
+ # akk-111m
18
+
19
+ This model was trained from scratch on the [Akkademia dataset](https://github.com/gaigutherz/Akkademia).
20
+ It achieves the following categorical cross-entropy results on the evaluation set (512 tokens):
21
+ - Loss: 0.0753
22
+
23
+ Cuneiform -> English Bleu score
24
+ - 500 tokens: 38.91
25
+ - 100 tokens: 43.13
26
+
27
+ Transliterated -> English Bleu score
28
+ - 500 tokens: 37.02
29
+ - 100 tokens: 41.67
30
+
31
+ Cuneiform -> Transliteration Bleu score
32
+ - 500 tokens: 94.31
33
+ - 100 tokens: 94.36
34
+
35
+ Cuneiform -> Transliteration Accuracy
36
+ - 100 tokens: 50% (note a single missed character significantly decreases accuracy in seq2seq models, see Bleu score for positional flexibility)
37
 
38
  ## Model description
39
 
40
+ This is an instruct model, meaning it is capable of multiple tasks. It is intended for primarily translation + transliteration, but it can also be used for reverse translation as well.
41
+
42
+ ###Translation Instrutions:
43
+ - "Translate Akkadian cuneiform to English" + cuneiform signs -> English
44
+ - "Translate Akkadian simple transliteration to English" + simple transliteration -> English
45
+ - "Translate Akkadian grouped transliteration to English" + transliteration with spacial symbols -> English
46
+ - "Translate English to Akkadian cuneiform" + English -> Akkadian cuneiform signs
47
+ - "Translate English to simple Akkadian transliteration" + English -> Akkadian simple transliteration with no special symbols
48
+ - "Translate English to grouped Akkadian transliteration" + English -> Akkadian transliteration grouped into words with special symbols
49
+
50
+ ###Transliteration Instructions:
51
+ - "Transliterate Akkadian cuneiform to simple Latin Characters" + cuneiform signs -> transliteration with no special symbols
52
+ - "Transliterate Akkadian cuneiform to grouped Latin characters" + cuneiform signs -> transliteration with special symbols/subscripts
53
+ - "Group Akkadian transliteration into likely words" + simple transliteration -> transliteration with special symbols/subscripts
54
+
55
 
56
  ## Intended uses & limitations
57
 
58
+ This model is designed to facilitate the translation/transliteration of Akkadian cuneiform. It may have limited facility in the reverse (e.g. translate English to Akkadian cuneiform) but these use cases are untested.
59
 
60
  ## Training and evaluation data
61
 
62
+ Data was used from the [Akkademia project](https://github.com/gaigutherz/Akkademia), previously published in [PNAS Nexus](https://academic.oup.com/pnasnexus/article/2/5/pgad096/7147349). More information on the training data, as well as the test and validation splits, can be found on both the GitHub and published methodology.
63
 
64
  ## Training procedure
65
 
66
+ Because of the unequal distribution of data (many short sequences + long sequences) data was trained with different padded lengths:
67
+ An initial few epochs with a max length of 56 tokens
68
+ A follow-up series of epochs at 128 tokens
69
+ The same for 256 tokens
70
+ A final 5 epochs for 512 tokens
71
+
72
+ The origional t5-small model had its tokens and embedding layers expanded by the additional linguistic data. Cuneiform symbols were split by spaces to be fed directly into the model, following the instructions detailed above.
73
+
74
+
75
  ### Training hyperparameters
76
 
77
  The following hyperparameters were used during training: