|
--- |
|
license: mit |
|
datasets: |
|
- pcuenq/oxford-pets |
|
metrics: |
|
- accuracy |
|
pipeline_tag: image-classification |
|
--- |
|
|
|
# CLIP ViT Base Patch32 Fine-tuned on Oxford Pets |
|
|
|
This model is a fine-tuned version of OpenAI's CLIP model on the Oxford Pets dataset, intended for pets classification. |
|
|
|
## Training Information |
|
|
|
- **Model Name**: openai/clip-vit-base-patch32 |
|
- **Dataset**: oxford-pets |
|
- **Training Epochs**: 4 |
|
- **Batch Size**: 256 |
|
- **Learning Rate**: 3e-6 |
|
- **Test Accuracy**: 93.74% |
|
|
|
## Parameters Information |
|
|
|
Trainable params: 151.2773M || All params: 151.2773M || Trainable%: 100.00% |
|
|
|
## Bias, Risks, and Limitations |
|
|
|
Refer to the original [CLIP repository](https://huggingface.co/openai/clip-vit-base-patch32). |
|
|
|
## License |
|
[MIT] |