Italian Psychology DistilGPT-2

This model is a fine-tuned version of the GPT-2 language model, trained on a dataset of Italian psychology articles. It is capable of generating human-like text on topics related to psychology and mental health in Italian language.

Model details

The base model used for fine-tuning is the GPT-2 model, with a transformer architecture and 1250M parameters. The fine-tuning dataset consists of approximately 10,000 Italian psychology articles.

Example usage

from transformers import pipeline

nlp = pipeline("text-generation", model="misterkilgore/distilgpt2-psy-ita")

generated_text = nlp("Le cause del disturbo d'ansia nei bambini sono", max_length=100)
print(generated_text)

Limitations and bias

This model has been trained on a dataset of Italian psychology articles and may not perform well on other types of text or in other languages. Additionally, the dataset used to fine-tune the model may contain biases and limitations, which will be reflected in the generated text.

Dataset

The dataset used to fine-tune this model is composed of Italian psychology articles. It contains various topics on mental health and psychology, but some limitations and biases may be present. This model is meant to be used only for research and educational purposes.

Training data

The training data is composed of Italian psychology articles. Fine-tuning was performed on this dataset to adapt the base GPT-2 model to the specific topic of psychology and mental health in Italian language.

Downloads last month
19
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.