agentlans's picture
Update README.md
1d753ae verified
metadata
language: en
license: mit
library_name: transformers
tags:
  - generated_from_trainer
  - text-classification
  - fill-mask
  - embeddings
metrics:
  - accuracy
model-index:
  - name: snowflake-arctic-embed-xs-zyda-2
    results:
      - task:
          type: text-classification
          name: Text Classification
        dataset:
          name: Zyphra/Zyda-2 (subset)
          type: Zyphra/Zyda-2
        metrics:
          - type: accuracy
            value: 0.4676
            name: Accuracy
base_model: Snowflake/snowflake-arctic-embed-xs

snowflake-arctic-embed-xs-zyda-2

Model Description

This model is a fine-tuned version of Snowflake/snowflake-arctic-embed-xs on a subset of the Zyphra/Zyda-2 dataset. It was trained using the Masked Language Modeling (MLM) objective to enhance its understanding of the English language.

Performance

The model achieves the following results on the evaluation set:

  • Loss: 3.0689
  • Accuracy: 0.4676

Intended Uses & Limitations

This model is designed to be used and finetuned for the following tasks:

  • Text embedding
  • Text classification
  • Fill-in-the-blank tasks

Limitations:

  • English language only
  • May be inaccurate for specialized jargon, dialects, slang, code, and LaTeX

Training Data

The model was trained on the first 300 000 rows of the Zyphra/Zyda-2 dataset. 5% of that data was used for validation.

Training Procedure

Hyperparameters

The following hyperparameters were used during training:

  • Learning rate: 5e-05
  • Train batch size: 8
  • Eval batch size: 8
  • Seed: 42
  • Optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • Learning rate scheduler: Linear
  • Number of epochs: 1.0

Framework Versions

  • Transformers: 4.44.2
  • PyTorch: 2.5.1+cu124
  • Datasets: 3.1.0
  • Tokenizers: 0.19.1

Usage Examples

Masked Language Modeling

from transformers import pipeline

unmasker = pipeline('fill-mask', model='agentlans/snowflake-arctic-embed-xs-zyda-2')
result = unmasker("[MASK] is the capital of France.")
print(result)

Text Embedding

from transformers import AutoTokenizer, AutoModel
import torch

model_name = "agentlans/snowflake-arctic-embed-xs-zyda-2"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModel.from_pretrained(model_name)

text = "Example sentence for embedding."
inputs = tokenizer(text, return_tensors='pt')
with torch.no_grad():
    outputs = model(**inputs)

embeddings = outputs.last_hidden_state.mean(dim=1)
print(embeddings)

Ethical Considerations and Bias

As this model is trained on a subset of the Zyda-2 dataset, it may inherit biases present in that data. Users should be aware of potential biases and evaluate the model's output critically, especially for sensitive applications.

Additional Information

For more details about the base model, please refer to Snowflake/snowflake-arctic-embed-xs.