--- datasets: - imdb language: - en library_name: transformers pipeline_tag: text-classification tags: - movies - gpt2 - sentiment-analysis - fine-tuned --- # Fine-tuned GPT-2 Model for IMDb Movie Review Sentiment Analysis ## Model Description This is a GPT-2 model fine-tuned on the IMDb movie review dataset for sentiment analysis. It classifies a movie review text into two classes: "positive" or "negative". ## Intended Uses & Limitations This model is intended to be used for binary sentiment analysis of English movie reviews. It can determine whether a review is positive or negative. It should not be used for languages other than English, or for text with ambiguous sentiment. ## How to Use Here's a simple way to use this model: ```python from transformers import GPT2Tokenizer, GPT2ForSequenceClassification tokenizer = GPT2Tokenizer.from_pretrained("hipnologo/gpt2-imdb-finetune") model = GPT2ForSequenceClassification.from_pretrained("hipnologo/gpt2-imdb-finetune") text = "Your review text here!" # encoding the input text input_ids = tokenizer.encode(text, return_tensors="pt") # Move the input_ids tensor to the same device as the model input_ids = input_ids.to(model.device) # getting the logits logits = model(input_ids).logits # getting the predicted class predicted_class = logits.argmax(-1).item() print(f"The sentiment predicted by the model is: {'Positive' if predicted_class == 1 else 'Negative'}") ``` ## Training Procedure The model was trained using the 'Trainer' class from the transformers library, with a learning rate of 2e-5, batch size of 1, and 3 training epochs. ## Fine-tuning Details The model was fine-tuned using the IMDb movie review dataset.