language:
- en
pipeline_tag: text-classification
Span NLI BERT (base)
This is a BERT-base model (bert-base-uncased
) fine-tuned on the ContractNLI dataset (non-disclosure agreements) with the Span NLI BERT model architecture,
from ContractNLI: A Dataset for Document-level Natural Language Inference for Contracts (Koreeda and Manning, 2021).
For a hypothesis, the Span NLI BERT model predicts NLI labels and identifies evidence for documents as premises. Spans of documents should be pre-annotated; evidence is always full sentences or items in an enumerated list in the document.
For details of the architecture and usage of the relevant training/testing scripts, check out the paper and their Github repo.
This model is fine-tuned according to the hyperparameters in data/conf_base.yml
in their repo,
which differs from their hyperparameters that produced the best dev scores as noted in the Appendix of the paper.