TOSBertV2 / README.md
CodeHima's picture
Update README.md
6f4c86f verified
metadata
license: mit
language:
  - en
metrics:
  - accuracy
widget:
  - text: >-
      You have the right to use CommunityConnect for its intended purpose of
      connecting with others, sharing content responsibly, and engaging in
      constructive dialogue. You are responsible for the content you post and
      must respect the rights and privacy of others.
    example_title: Fair Clause
  - text: ' We reserve the right to suspend, terminate, or restrict your access to the platform at any time and for any reason, without prior notice or explanation. This includes but is not limited to violations of our community guidelines or terms of service, as determined solely by ConnectWorld.'
    example_title: Unfair Clause
library_name: transformers
pipeline_tag: text-classification
tags:
  - nlp
  - bert
  - TOS

BertTOS v2: Terms of Service Unfairness Classifier

Model Details

  • Model Name: BertTOS v2
  • Model Type: Fine-tuned BERT for sequence classification
  • Version: 2.0
  • Language(s): English
  • License: [MIT]
  • Developer: [Himanshu Mohanty]

Model Description

BertTOS v2 is a fine-tuned BERT model designed to classify clauses in Terms of Service (ToS) documents based on their unfairness level. This model can help users identify potentially problematic clauses in legal documents, particularly in the context of consumer protection.

Task

The model performs multi-class classification on individual sentences or clauses, categorizing them into three levels of unfairness:

  1. Clearly Fair
  2. Potentially Unfair
  3. Clearly Unfair

Training Data

The model was trained on the CodeHima/TOS_Dataset dataset, which contains annotated sentences from Terms of Service documents. Each sentence is labeled with one of the three unfairness levels.

Model Architecture

  • Base Model: BERT (bert-base-uncased)
  • Fine-tuning: Sequence classification head
  • Input: Tokenized text (max length 512 tokens)
  • Output: Probabilities for each unfairness level

Performance

The model's performance metrics on the test set:

  • Accuracy: [0.8795761078998073]
  • F1 Score (weighted): [0.885282]
  • Precision (weighted): [0.883729]
  • Recall (weighted): [0.889157]

Limitations

  • The model is trained on English language ToS documents and may not perform well on other languages or legal contexts.
  • Performance may vary depending on the specific wording and context of clauses.
  • The model should be used as a tool to assist human judgment, not as a definitive legal assessment.

Ethical Considerations

  • This model is intended to help identify potentially unfair clauses, but it should not be considered as legal advice.
  • Users should be aware of potential biases in the training data and model predictions.
  • The model's output should be reviewed by legal professionals for critical applications.

How to Use

You can use this model directly with the Hugging Face transformers library:

from transformers import AutoTokenizer, AutoModelForSequenceClassification
import torch

# Load model and tokenizer
model_name = "YourHuggingFaceUsername/TOSBertV2"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForSequenceClassification.from_pretrained(model_name)

# Function to predict unfairness level
def predict_unfairness(text):
    inputs = tokenizer(text, return_tensors="pt", truncation=True, padding=True, max_length=512)
    
    model.eval()
    with torch.no_grad():
        outputs = model(**inputs)
    
    probabilities = torch.softmax(outputs.logits, dim=-1).squeeze()
    predicted_class = torch.argmax(probabilities).item()
    
    label_mapping = {0: 'clearly_fair', 1: 'potentially_unfair', 2: 'clearly_unfair'}
    predicted_label = label_mapping[predicted_class]
    
    return predicted_label, probabilities.tolist()

# Example usage
clause = "The company reserves the right to change these terms at any time without notice."
predicted_label, probabilities = predict_unfairness(clause)

print(f"Predicted unfairness level: {predicted_label}")
print("Probabilities:")
for label, prob in zip(['clearly_fair', 'potentially_unfair', 'clearly_unfair'], probabilities):
    print(f"{label}: {prob:.4f}")

Training

The model was trained using the following hyperparameters:

  • Epochs: 3
  • Batch Size: 16
  • Learning Rate: [ ]
  • Optimizer: AdamW
  • Weight Decay: 0.01

Citation

If you use this model in your research, please cite:

@misc{TOSBertV2,
  author = {Himanshu Mohanty},
  title = {TOSBertV2: is a fine-tuned BERT model designed to classify clauses in Terms of Service},
  year = {2024},
  publisher = {Hugging Face},
  journal = {Hugging Face Model Hub},
  howpublished = {\url{https://huggingface.co/CodeHima/TOSBertV2}}
}