Datasets:

Modalities:
Text
Languages:
English
ArXiv:
Libraries:
Datasets
License:
lewtun HF staff commited on
Commit
45fc294
1 Parent(s): 3fe19b3

Add URL to RAFT viewer

Browse files
Files changed (1) hide show
  1. README.md +2 -0
README.md CHANGED
@@ -63,6 +63,8 @@ task_ids:
63
 
64
  The Real-world Annotated Few-shot Tasks (RAFT) dataset is an aggregation of English-language datasets found in the real world. Associated with each dataset is a binary or multiclass classification task, intended to improve our understanding of how language models perform on tasks that have concrete, real-world value. Only 50 labeled examples are provided in each dataset.
65
 
 
 
66
  ### Supported Tasks and Leaderboards
67
 
68
  - `text-classification`: Each subtask in RAFT is a text classification task, and the provided train and test sets can be used to submit to the [RAFT Leaderboard](https://huggingface.co/spaces/ought/raft-leaderboard) To prevent overfitting and tuning on a held-out test set, the leaderboard is only evaluated once per week. Each task has its macro-f1 score calculated, then those scores are averaged to produce the overall leaderboard score.
 
63
 
64
  The Real-world Annotated Few-shot Tasks (RAFT) dataset is an aggregation of English-language datasets found in the real world. Associated with each dataset is a binary or multiclass classification task, intended to improve our understanding of how language models perform on tasks that have concrete, real-world value. Only 50 labeled examples are provided in each dataset.
65
 
66
+ You can explore the dataset [here](https://huggingface.co/spaces/ought/raft-viewer?dataset=raft&config=ade_corpus_v2).
67
+
68
  ### Supported Tasks and Leaderboards
69
 
70
  - `text-classification`: Each subtask in RAFT is a text classification task, and the provided train and test sets can be used to submit to the [RAFT Leaderboard](https://huggingface.co/spaces/ought/raft-leaderboard) To prevent overfitting and tuning on a held-out test set, the leaderboard is only evaluated once per week. Each task has its macro-f1 score calculated, then those scores are averaged to produce the overall leaderboard score.