Datasets:

Modalities:
Text
Formats:
parquet
Languages:
English
ArXiv:
Libraries:
Datasets
Dask
License:
Qingyun commited on
Commit
23c01d2
1 Parent(s): 4697ac2

Upload dataset

Browse files
CC-MAIN-2014-10/train-00000-of-00008.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b804f61392839f69a997080c569f06aeaa29c330b3e9e188cc976720808fd344
3
+ size 407152634
CC-MAIN-2014-10/train-00001-of-00008.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b6448c8f60d23fcaf03330d8ed1269ae518ca11a27e417cedfa564bb01cb617d
3
+ size 406676018
CC-MAIN-2014-10/train-00002-of-00008.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:24631111dd081731198468506d92375cecf90bac14b4cd230a625ea3b964510d
3
+ size 408166490
CC-MAIN-2014-10/train-00003-of-00008.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:94eeb9f12c75637ea2003aceba8529fe6c6ce73f0460b5826d929d797c839200
3
+ size 405844536
CC-MAIN-2014-10/train-00004-of-00008.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9cbece418d2426155bac89b94811d309bef094945f80c1997a208d6d8b6b6d3a
3
+ size 407071632
CC-MAIN-2014-10/train-00005-of-00008.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e4c92a8e4bdca4bae06a953d98a02f5ad0fcff5b3ea7a0027f01166d3c1184e4
3
+ size 408141045
CC-MAIN-2014-10/train-00006-of-00008.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cc4523fc443cb068426af3feb7f8f3b2102e569bcd5ca1561e70d67e1b476fee
3
+ size 408271277
CC-MAIN-2014-10/train-00007-of-00008.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c297bd9ea92e45b2fac79d2bb5fcceb62fbaa360989b49941047fba838786652
3
+ size 407915929
README.md CHANGED
@@ -112,6 +112,58 @@ dataset_info:
112
  num_examples: 3091537
113
  download_size: 6965036866
114
  dataset_size: 15282078925
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
115
  configs:
116
  - config_name: CC-MAIN-2013-20
117
  data_files:
@@ -121,6 +173,10 @@ configs:
121
  data_files:
122
  - split: train
123
  path: CC-MAIN-2013-48/train-*
 
 
 
 
124
  ---
125
 
126
  We are uploading the dataset files ~
 
112
  num_examples: 3091537
113
  download_size: 6965036866
114
  dataset_size: 15282078925
115
+ - config_name: CC-MAIN-2014-10
116
+ features:
117
+ - name: general_metadata
118
+ struct:
119
+ - name: domain
120
+ sequence: string
121
+ - name: fluency_prob
122
+ dtype: float64
123
+ - name: id
124
+ dtype: string
125
+ - name: non_advertisement_prob
126
+ dtype: float64
127
+ - name: politics_prob
128
+ dtype: float64
129
+ - name: porn_prob
130
+ dtype: float64
131
+ - name: toxic_prob
132
+ dtype: float64
133
+ - name: url
134
+ dtype: string
135
+ - name: images
136
+ sequence: string
137
+ - name: texts
138
+ sequence: string
139
+ - name: metadata
140
+ list:
141
+ - name: aesthetic_prob
142
+ dtype: float64
143
+ - name: bytes
144
+ dtype: int64
145
+ - name: d_hash
146
+ dtype: string
147
+ - name: d_hash_dup_count
148
+ dtype: int64
149
+ - name: height
150
+ dtype: int64
151
+ - name: img_url_sha
152
+ dtype: string
153
+ - name: p_hash
154
+ dtype: string
155
+ - name: p_hash_dup_count
156
+ dtype: int64
157
+ - name: unsafe_prob
158
+ dtype: float64
159
+ - name: width
160
+ dtype: int64
161
+ splits:
162
+ - name: train
163
+ num_bytes: 7227087609
164
+ num_examples: 1390034
165
+ download_size: 3259239561
166
+ dataset_size: 7227087609
167
  configs:
168
  - config_name: CC-MAIN-2013-20
169
  data_files:
 
173
  data_files:
174
  - split: train
175
  path: CC-MAIN-2013-48/train-*
176
+ - config_name: CC-MAIN-2014-10
177
+ data_files:
178
+ - split: train
179
+ path: CC-MAIN-2014-10/train-*
180
  ---
181
 
182
  We are uploading the dataset files ~