Update requirements.txt
ec3b8de
verified
-
1.52 kB
initial commit
-
247 Bytes
initial commit
-
1.56 kB
Update app.py
-
321 kB
Rename Audio.mp3 to example_audio.mp3
model.pkl
Detected Pickle imports (91)
- "inspect.Parameter",
- "PIL.Image.Image",
- "fastai.data.transforms.Categorize",
- "fastai.vision.augment._WarpCoord",
- "fastai.learner.CastToTensor",
- "inspect.Signature",
- "pathlib.PosixPath",
- "fastai.vision.augment.Flip",
- "fastai.data.transforms.CategoryMap",
- "torch.nn.modules.batchnorm.BatchNorm1d",
- "fastai.vision.learner._resnet_split",
- "fastai.vision.augment.RandomResizedCropGPU",
- "fastcore.basics.fastuple",
- "fastai.torch_core.TensorCategory",
- "torch._tensor._rebuild_from_type_v2",
- "torch.device",
- "fastai.data.core.Datasets",
- "fastai.vision.augment._BrightnessLogit",
- "numpy.dtype",
- "fastcore.foundation.L",
- "torch.nn.modules.pooling.AdaptiveMaxPool2d",
- "fastai.learner.AvgSmoothLoss",
- "torch.Tensor",
- "__builtin__.object",
- "fastai.layers.AdaptiveConcatPool2d",
- "functools.partial",
- "fastcore.transform.Pipeline",
- "fastcore.dispatch.TypeDispatch",
- "fastai.data.core.TfmdDL",
- "__builtin__.float",
- "torch.nn.modules.linear.Linear",
- "__builtin__.set",
- "fastai.imports.noop",
- "fastai.data.transforms.Normalize",
- "fastai.learner.Learner",
- "fastai.layers.Flatten",
- "torch._utils._rebuild_parameter",
- "__builtin__.tuple",
- "fastcore.transform.Transform",
- "fastai.learner.AvgMetric",
- "pathlib.Path",
- "fastai.data.transforms.parent_label",
- "__builtin__.bytes",
- "torch.nn.modules.activation.ReLU",
- "numpy.ndarray",
- "torchvision.models.resnet.resnet50",
- "random.Random",
- "fastai.losses.CrossEntropyLossFlat",
- "torch.nn.modules.batchnorm.BatchNorm2d",
- "fastai.vision.core.PILImage",
- "__builtin__.getattr",
- "inspect._ParameterKind",
- "fastai.torch_core.TensorBase",
- "fastai.learner.Recorder",
- "fastai.data.core.DataLoaders",
- "inspect._empty",
- "fastai.vision.augment._ContrastLogit",
- "fastai.vision.augment.Resize",
- "fastai.torch_core.TensorImage",
- "fastai.vision.augment.flip_mat",
- "fastai.data.transforms.IntToFloatTensor",
- "fastai.optimizer.Adam",
- "fastai.data.load._wif",
- "torch.nn.modules.pooling.AdaptiveAvgPool2d",
- "torch._utils._rebuild_tensor_v2",
- "torch.nn.modules.conv.Conv2d",
- "numpy.core.multiarray.scalar",
- "torch.nn.modules.pooling.MaxPool2d",
- "torch.nn.modules.dropout.Dropout",
- "fastcore.dispatch._TypeDict",
- "_codecs.encode",
- "torchvision.models.resnet.Bottleneck",
- "fastai.data.transforms.ToTensor",
- "fastai.vision.augment.zoom_mat",
- "fastai.learner.AvgLoss",
- "__builtin__.long",
- "collections.OrderedDict",
- "fastai.metrics.accuracy",
- "fastai.vision.augment.Brightness",
- "__builtin__.unicode",
- "torch.FloatStorage",
- "torch.nn.modules.container.Sequential",
- "fastai.callback.progress.ProgressCallback",
- "torch.nn.modules.loss.CrossEntropyLoss",
- "torch.LongStorage",
- "fastai.callback.core.TrainEvalCallback",
- "fastai.data.load._FakeLoader",
- "PIL.Image.Resampling",
- "fastai.vision.augment.rotate_mat",
- "fastai.data.core.TfmdLists",
- "__builtin__.print"
How to fix it?
103 MB
Rename modal.pkl to model.pkl
-
56 Bytes
Update requirements.txt