|
--- |
|
language: |
|
- fa |
|
pipeline_tag: feature-extraction |
|
--- |
|
This is the Persian word2vec embedding model trained with CBOW algorithm on the wikipedia data. |
|
|
|
In order to use this model in Hezar you can simply use this piece of code: |
|
```bash |
|
pip install hezar |
|
``` |
|
```python |
|
from hezar import Embedding |
|
|
|
w2v = Embedding.load("hezarai/word2vec-cbow-fa-wikipedia") |
|
# Get embedding vector |
|
vector = w2v("هزار") |
|
# Find the word that doesn't match with the rest |
|
doesnt_match = w2v.doesnt_match(["خانه", "اتاق", "ماشین"]) |
|
# Find the top-n most similar words to the given word |
|
most_similar = w2v.most_similar("هزار", top_n=5) |
|
# Find the cosine similarity value between two words |
|
similarity = w2v.similarity("مهندس", "دکتر") |
|
``` |