Update README.md
Browse files
README.md
CHANGED
@@ -16,9 +16,12 @@ _Nesta, the UK's innovation agency, has been scraping online job adverts since 2
|
|
16 |
|
17 |
_Although we are unable to share the raw data openly, we aim to open source **our models, algorithms and tools** so that anyone can use them for their own research and analysis._
|
18 |
|
|
|
|
|
19 |
This model is pre-trained from a `distilbert-base-uncased` checkpoint on 100k sentences from scraped online job postings as part of the Open Jobs Observatory.
|
20 |
|
21 |
-
🖨️ Use
|
|
|
22 |
To use the model:
|
23 |
|
24 |
```
|
@@ -33,7 +36,7 @@ An example use is as follows:
|
|
33 |
text = "Would you like to join a major [MASK] company?"
|
34 |
model(text, top_k=3)
|
35 |
|
36 |
-
|
37 |
'token': 13859,
|
38 |
'token_str': 'pharmaceutical',
|
39 |
'sequence': 'would you like to join a major pharmaceutical company?'},
|
@@ -46,7 +49,8 @@ model(text, top_k=3)
|
|
46 |
'token_str': 'construction',
|
47 |
'sequence': 'would you like to join a major construction company?'}]
|
48 |
|
49 |
-
⚖️ Training results
|
|
|
50 |
The fine-tuning metrics are as follows:
|
51 |
|
52 |
- eval_loss: 2.5871026515960693
|
|
|
16 |
|
17 |
_Although we are unable to share the raw data openly, we aim to open source **our models, algorithms and tools** so that anyone can use them for their own research and analysis._
|
18 |
|
19 |
+
## 📟 About
|
20 |
+
|
21 |
This model is pre-trained from a `distilbert-base-uncased` checkpoint on 100k sentences from scraped online job postings as part of the Open Jobs Observatory.
|
22 |
|
23 |
+
## 🖨️ Use
|
24 |
+
|
25 |
To use the model:
|
26 |
|
27 |
```
|
|
|
36 |
text = "Would you like to join a major [MASK] company?"
|
37 |
model(text, top_k=3)
|
38 |
|
39 |
+
[{'score': 0.1886572688817978,
|
40 |
'token': 13859,
|
41 |
'token_str': 'pharmaceutical',
|
42 |
'sequence': 'would you like to join a major pharmaceutical company?'},
|
|
|
49 |
'token_str': 'construction',
|
50 |
'sequence': 'would you like to join a major construction company?'}]
|
51 |
|
52 |
+
## ⚖️ Training results
|
53 |
+
|
54 |
The fine-tuning metrics are as follows:
|
55 |
|
56 |
- eval_loss: 2.5871026515960693
|