|
import pandas as pd |
|
import numpy as np |
|
import os |
|
import subprocess |
|
import sys |
|
from tqdm import tqdm |
|
import timm |
|
import torchvision.transforms as T |
|
from PIL import Image |
|
import torch |
|
|
|
|
|
|
|
CONFIG_PATH = 'models/swinv2_base_w24_b16x4-fp16_fungi+val_res_384_cb_epochs_6.py' |
|
CHECKPOINT_PATH = "models/swinv2_base_w24_b16x4-fp16_fungi+val_res_384_cb_epochs_6_epoch_6_20240514-de00365e.pth" |
|
SCORE_THRESHOLD = 0.2 |
|
|
|
def run_inference(input_csv, output_csv, data_root_path): |
|
"""Load model and dataloader and run inference.""" |
|
|
|
if not data_root_path.endswith('/'): |
|
data_root_path += '/' |
|
data_cfg_opts = [ |
|
f'test_dataloader.dataset.data_root=', |
|
f'test_dataloader.dataset.ann_file={input_csv}', |
|
f'test_dataloader.dataset.data_prefix={data_root_path}'] |
|
|
|
inference = subprocess.Popen([ |
|
'python', '-m', |
|
'tools.test_generate_result_pre-consensus', |
|
CONFIG_PATH, CHECKPOINT_PATH, |
|
output_csv, |
|
'--threshold', str(SCORE_THRESHOLD), |
|
'--no-scores', |
|
'--cfg-options'] + data_cfg_opts) |
|
return_code = inference.wait() |
|
if return_code != 0: |
|
print(f'Inference crashed with exit code {return_code}') |
|
sys.exit(return_code) |
|
print(f'Written {output_csv}') |
|
|
|
|
|
if __name__ == "__main__": |
|
|
|
import zipfile |
|
|
|
with zipfile.ZipFile("/tmp/data/private_testset.zip", 'r') as zip_ref: |
|
zip_ref.extractall("/tmp/data") |
|
|
|
metadata_file_path = "./FungiCLEF2024_TestMetadata.csv" |
|
|
|
run_inference(metadata_file_path, "./submission.csv", "/tmp/data/private_testset") |
|
|