Abinaya commited on
Commit
ba4ba04
1 Parent(s): fe5e94f

News Event fine tuned model

Browse files

Microsoft Phi1.5 model fine tuned on Event Detection Dataset <href https://cs.uns.edu.ar/~mmaisonnave/resources/maisonnave2020improving.pdf >

Files changed (1) hide show
  1. README.md +10 -57
README.md CHANGED
@@ -1,58 +1,11 @@
1
  ---
2
- license: mit
3
- library_name: peft
4
- tags:
5
- - trl
6
- - sft
7
- - generated_from_trainer
8
- base_model: microsoft/phi-1_5
9
- model-index:
10
- - name: phi-1_5-finetuned-news-events
11
- results: []
12
- ---
13
-
14
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
- should probably proofread and complete it, then remove this comment. -->
16
-
17
- # phi-1_5-finetuned-news-events
18
-
19
- This model is a fine-tuned version of [microsoft/phi-1_5](https://huggingface.co/microsoft/phi-1_5) on an unknown dataset.
20
- It achieves the following results on the evaluation set:
21
- - Loss: 2.3737
22
-
23
- ## Model description
24
-
25
- More information needed
26
-
27
- ## Intended uses & limitations
28
-
29
- More information needed
30
-
31
- ## Training and evaluation data
32
-
33
- More information needed
34
-
35
- ## Training procedure
36
-
37
- ### Training hyperparameters
38
-
39
- The following hyperparameters were used during training:
40
- - learning_rate: 0.0002
41
- - train_batch_size: 4
42
- - eval_batch_size: 8
43
- - seed: 42
44
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
- - lr_scheduler_type: cosine
46
- - training_steps: 800
47
-
48
- ### Training results
49
-
50
-
51
-
52
- ### Framework versions
53
-
54
- - PEFT 0.10.0
55
- - Transformers 4.38.2
56
- - Pytorch 2.2.1+cu121
57
- - Datasets 2.18.0
58
- - Tokenizers 0.15.2
 
1
  ---
2
+ license: apache-2.0
3
+ datasets:
4
+ - cnn_dailymail
5
+ language:
6
+ - en
7
+ metrics:
8
+ - accuracy
9
+ library_name: adapter-transformers
10
+ pipeline_tag: text-generation
11
+ ---