MIT/ast-finetuned-speech-commands-v2
This is the MIT/ast-finetuned-speech-commands-v2 model converted to OpenVINO, for accellerated inference.
An example of how to do inference on this model:
from optimum.intel.openvino import OVModelForAudioClassification
from transformers import AutoFeatureExtractor, pipeline
# model_id should be set to either a local directory or a model available on the HuggingFace hub.
model_id = "helenai/MIT-ast-finetuned-speech-commands-v2-ov"
feature_extractor = AutoFeatureExtractor.from_pretrained(model_id)
model = OVModelForAudioClassification.from_pretrained(model_id)
pipe = pipeline("audio-classification", model=model, feature_extractor=feature_extractor)
result = pipe("https://datasets-server.huggingface.co/assets/speech_commands/--/v0.01/test/38/audio/audio.mp3")
print(result)
- Downloads last month
- 749
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.