Datasets:

Formats:
parquet
ArXiv:
Libraries:
Datasets
Dask
filtered-wit / wit_create_meta.py
ARKseal's picture
Upload wit_create_meta.py
c38ca74
raw
history blame contribute delete
942 Bytes
from pathlib import Path
import pandas as pd
import torch
import ujson
import webdataset as wds
from tqdm import tqdm
def load_json(json):
return ujson.loads(json)
load_map = {
'json': load_json,
}
def get_glob(path):
return sorted(Path('.').absolute().glob(path))
def chunker(l, n):
for i in range(0, len(l), n):
yield l[i:i + n]
def func(glob_path, combine_all=True):
glob = get_glob(glob_path)
for file in tqdm(glob, position=0):
ds = wds.WebDataset(str(file)).map_dict(**load_map).to_tuple('json')
metas = pd.DataFrame([meta[0]
for meta in tqdm(ds, position=1, leave=False)])
metas.to_parquet(f'{file.stem}.parquet')
if combine_all:
combined = pd.concat(pd.read_parquet(
f'{file.stem}.parquet') for file in tqdm(glob))
combined.to_parquet('combined_meta.parquet')