ihk commited on
Commit
51c0d06
·
1 Parent(s): f312c3c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -3
README.md CHANGED
@@ -16,9 +16,12 @@ _Nesta, the UK's innovation agency, has been scraping online job adverts since 2
16
 
17
  _Although we are unable to share the raw data openly, we aim to open source **our models, algorithms and tools** so that anyone can use them for their own research and analysis._
18
 
 
 
19
  This model is pre-trained from a `distilbert-base-uncased` checkpoint on 100k sentences from scraped online job postings as part of the Open Jobs Observatory.
20
 
21
- 🖨️ Use
 
22
  To use the model:
23
 
24
  ```
@@ -33,7 +36,7 @@ An example use is as follows:
33
  text = "Would you like to join a major [MASK] company?"
34
  model(text, top_k=3)
35
 
36
- >> [{'score': 0.1886572688817978,
37
  'token': 13859,
38
  'token_str': 'pharmaceutical',
39
  'sequence': 'would you like to join a major pharmaceutical company?'},
@@ -46,7 +49,8 @@ model(text, top_k=3)
46
  'token_str': 'construction',
47
  'sequence': 'would you like to join a major construction company?'}]
48
 
49
- ⚖️ Training results
 
50
  The fine-tuning metrics are as follows:
51
 
52
  - eval_loss: 2.5871026515960693
 
16
 
17
  _Although we are unable to share the raw data openly, we aim to open source **our models, algorithms and tools** so that anyone can use them for their own research and analysis._
18
 
19
+ ## 📟 About
20
+
21
  This model is pre-trained from a `distilbert-base-uncased` checkpoint on 100k sentences from scraped online job postings as part of the Open Jobs Observatory.
22
 
23
+ ## 🖨️ Use
24
+
25
  To use the model:
26
 
27
  ```
 
36
  text = "Would you like to join a major [MASK] company?"
37
  model(text, top_k=3)
38
 
39
+ [{'score': 0.1886572688817978,
40
  'token': 13859,
41
  'token_str': 'pharmaceutical',
42
  'sequence': 'would you like to join a major pharmaceutical company?'},
 
49
  'token_str': 'construction',
50
  'sequence': 'would you like to join a major construction company?'}]
51
 
52
+ ## ⚖️ Training results
53
+
54
  The fine-tuning metrics are as follows:
55
 
56
  - eval_loss: 2.5871026515960693