scfengv's picture
Update README
b900f3a
metadata
license: mit
datasets:
  - scfengv/TVL-game-layer-dataset
language:
  - zh
metrics:
  - accuracy
base_model:
  - google-bert/bert-base-chinese
pipeline_tag: text-classification
tags:
  - multi-label
model-index:
  - name: scfengv/TVL_GameLayerClassifier
    results:
      - task:
          type: multi-label text-classification
        dataset:
          name: scfengv/TVL-game-layer-dataset
          type: scfengv/TVL-game-layer-dataset
        metrics:
          - name: Accuracy
            type: Accuracy
            value: 0.985764
          - name: F1 score (Micro)
            type: F1 score (Micro)
            value: 0.993132
          - name: F1 score (Macro)
            type: F1 score (Macro)
            value: 0.993694

Model Details of TVL_GameLayerClassifier

Base Model

This model is fine-tuned from google-bert/bert-base-chinese.

Model Architecture

  • Type: BERT-based text classification model
  • Hidden Size: 768
  • Number of Layers: 12
  • Number of Attention Heads: 12
  • Intermediate Size: 3072
  • Max Sequence Length: 512
  • Vocabulary Size: 21,128

Key Components

  1. Embeddings

    • Word Embeddings
    • Position Embeddings
    • Token Type Embeddings
    • Layer Normalization
  2. Encoder

    • 12 layers of:
      • Self-Attention Mechanism
      • Intermediate Dense Layer
      • Output Dense Layer
      • Layer Normalization
  3. Pooler

    • Dense layer for sentence representation
  4. Classifier

    • Output layer with 5 classes

Training Hyperparameters

The model was trained using the following hyperparameters:

Learning rate: 1e-05
Batch size: 32
Number of epochs: 10
Optimizer: Adam
Loss function: torch.nn.BCEWithLogitsLoss()

Training Infrastructure

  • Hardware Type: NVIDIA Quadro RTX8000
  • Library: PyTorch
  • Hours used: 2hr 13mins

Model Parameters

  • Total parameters: ~102M (estimated)
  • All parameters are in 32-bit floating point (F32) format

Input Processing

  • Uses BERT tokenization
  • Supports sequences up to 512 tokens

Output

  • 5-class multi-label classification

Performance Metrics

  • Accuracy score: 0.985764
  • F1 score (Micro): 0.993132
  • F1 score (Macro): 0.993694

Training Dataset

This model was trained on the scfengv/TVL-game-layer-dataset.

Testing Dataset

Usage

import torch
from transformers import BertForSequenceClassification, BertTokenizer

model = BertForSequenceClassification.from_pretrained("scfengv/TVL_GameLayerClassifier")
tokenizer = BertTokenizer.from_pretrained("scfengv/TVL_GameLayerClassifier")

# Prepare your text
text = "Your text here" ## Please refer to Dataset
inputs = tokenizer(text, return_tensors = "pt", padding = True, truncation = True, max_length = 512)

# Make prediction
with torch.no_grad():
    outputs = model(**inputs)
    predictions = torch.sigmoid(outputs.logits)

# Print predictions
print(predictions)

Additional Notes

  • This model is specifically designed for TVL Game layer classification tasks.
  • It's based on the Chinese BERT model, indicating it's optimized for Chinese text.

For more detailed information about the model architecture or usage, please refer to the BERT documentation and the specific fine-tuning process used for this classifier.