autotrain-data-processor commited on
Commit
61f32dc
1 Parent(s): 3352993

Processed data from AutoTrain data processor ([2023-05-03 12:43 ]

Browse files
README.md ADDED
@@ -0,0 +1,53 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ task_categories:
3
+ - text-classification
4
+
5
+ ---
6
+ # AutoTrain Dataset for project: sachin-test-summarizer
7
+
8
+ ## Dataset Description
9
+
10
+ This dataset has been automatically processed by AutoTrain for project sachin-test-summarizer.
11
+
12
+ ### Languages
13
+
14
+ The BCP-47 code for the dataset's language is unk.
15
+
16
+ ## Dataset Structure
17
+
18
+ ### Data Instances
19
+
20
+ A sample from this dataset looks as follows:
21
+
22
+ ```json
23
+ [
24
+ {
25
+ "text": "\u2212 Scope 3: Optional scope that includes indirect emissions associated with the goods and services supply chain produced outside the organization. Included are emissions from the transport of products from our logistics centres to stores (downstream) performed by external logistics operators (air, land and sea transport) as well as the emissions associated with electricity consumption in franchise stores.",
26
+ "target": 1
27
+ },
28
+ {
29
+ "text": "The Group is not aware of any noise pollution that could negatively impact the environment, nor is it aware of any impact on biodiversity. With regards to land use, the Group is only a commercial user, and the Group is not aware of any local constraints with regards to water supply. The Group does not believe that it is at risk with regards to climate change in the near-or mid-term.",
30
+ "target": 0
31
+ }
32
+ ]
33
+ ```
34
+
35
+ ### Dataset Fields
36
+
37
+ The dataset has the following fields (also called "features"):
38
+
39
+ ```json
40
+ {
41
+ "text": "Value(dtype='string', id=None)",
42
+ "target": "ClassLabel(names=['0', '1', '2'], id=None)"
43
+ }
44
+ ```
45
+
46
+ ### Dataset Splits
47
+
48
+ This dataset is split into a train and validation split. The split sizes are as follow:
49
+
50
+ | Split name | Num samples |
51
+ | ------------ | ------------------- |
52
+ | train | 1000 |
53
+ | valid | 320 |
processed/dataset_dict.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"splits": ["train", "valid"]}
processed/train/data-00000-of-00001.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d251d4962e92b9dab31438e3665838eeec4eb204d77b269b3c9076313119dd78
3
+ size 491960
processed/train/dataset_info.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "citation": "",
3
+ "description": "AutoTrain generated dataset",
4
+ "features": {
5
+ "text": {
6
+ "dtype": "string",
7
+ "_type": "Value"
8
+ },
9
+ "target": {
10
+ "names": [
11
+ "0",
12
+ "1",
13
+ "2"
14
+ ],
15
+ "_type": "ClassLabel"
16
+ }
17
+ },
18
+ "homepage": "",
19
+ "license": "",
20
+ "splits": {
21
+ "train": {
22
+ "name": "train",
23
+ "num_bytes": 491359,
24
+ "num_examples": 1000,
25
+ "dataset_name": null
26
+ }
27
+ }
28
+ }
processed/train/state.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_data_files": [
3
+ {
4
+ "filename": "data-00000-of-00001.arrow"
5
+ }
6
+ ],
7
+ "_fingerprint": "8ae14147116ebc3e",
8
+ "_format_columns": [
9
+ "target",
10
+ "text"
11
+ ],
12
+ "_format_kwargs": {},
13
+ "_format_type": null,
14
+ "_output_all_columns": false,
15
+ "_split": null
16
+ }
processed/valid/data-00000-of-00001.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a1fbea8c411ededd6184ec94ad6802c498b2f184e508e1d9e0cd0938a8687a9a
3
+ size 174872
processed/valid/dataset_info.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "citation": "",
3
+ "description": "AutoTrain generated dataset",
4
+ "features": {
5
+ "text": {
6
+ "dtype": "string",
7
+ "_type": "Value"
8
+ },
9
+ "target": {
10
+ "names": [
11
+ "0",
12
+ "1",
13
+ "2"
14
+ ],
15
+ "_type": "ClassLabel"
16
+ }
17
+ },
18
+ "homepage": "",
19
+ "license": "",
20
+ "splits": {
21
+ "valid": {
22
+ "name": "valid",
23
+ "num_bytes": 174265,
24
+ "num_examples": 320,
25
+ "dataset_name": null
26
+ }
27
+ }
28
+ }
processed/valid/state.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_data_files": [
3
+ {
4
+ "filename": "data-00000-of-00001.arrow"
5
+ }
6
+ ],
7
+ "_fingerprint": "980c493c3e60fa87",
8
+ "_format_columns": [
9
+ "target",
10
+ "text"
11
+ ],
12
+ "_format_kwargs": {},
13
+ "_format_type": null,
14
+ "_output_all_columns": false,
15
+ "_split": null
16
+ }