IR_States_faiss / app.py
Hunzla's picture
Update app.py
0920132
from transformers import pipeline
from langchain.vectorstores import FAISS
from langchain.embeddings import HuggingFaceEmbeddings
pipe = pipeline("automatic-speech-recognition", model="openai/whisper-small")
import json
with open("tasks.json", "r",encoding="utf-8") as json_file:
global data
data = json.load(json_file)
def find_index(sentence):
global data
for key, value in data.items():
for i,j in value.items():
for s in j:
if sentence == s:
return i
for x,item in data.items():
texts = []
for key,value in item.items():
for each in value:
print(find_index(each))
texts.append(each)
globals()[f"faiss_{x}"] = FAISS.from_texts(texts,HuggingFaceEmbeddings(model_name="sentence-transformers/all-MiniLM-L6-v2",model_kwargs={'device':'cpu'}))
def transcribe_the_command(audio_path,state):
transcript = pipe(audio_path[1])["text"]
similar = globals()[f"faiss_{state}"].similarity_search(transcript)[0].page_content
print(similar)
reply = find_index(similar)
return reply
import gradio as gr
iface = gr.Interface(
fn=transcribe_the_command,
inputs=[gr.Audio(),gr.Textbox()],
outputs="text",
title="Whisper Small",
description="Realtime demo for intent recognition using a Whisper small model.",
)
iface.launch(share="true")