Spaces:
Running
Running
Update main.py
Browse files
main.py
CHANGED
@@ -693,6 +693,7 @@ async def construction_FCS(romeListArray,settings):
|
|
693 |
async def construction_NCS(romeListArray):
|
694 |
context = await contexte(romeListArray)
|
695 |
emploisST = cl.user_session.get("EmploiST")
|
|
|
696 |
### Mistral Completion ###
|
697 |
client_llm = await IA()
|
698 |
structure = str(modele('Note de composante sectorielle'))
|
@@ -710,11 +711,17 @@ async def construction_NCS(romeListArray):
|
|
710 |
prompt = PromptTemplate(template=template, input_variables=["question","context"])
|
711 |
#llm_chain = LLMChain(prompt=prompt, llm=client_llm)
|
712 |
#completion_NCS = llm_chain.run({"question":question_p,"context":context_p}, callbacks=[StreamingStdOutCallbackHandler()])
|
713 |
-
chain =
|
|
|
|
|
|
|
|
|
|
|
714 |
#completion_NCS = chain.invoke({"question":question_p,"context":context_p})
|
715 |
|
716 |
msg = cl.Message(author="Datapcc : πππ",content="")
|
717 |
-
async for chunk in chain.astream({"question":question_p,"context":context_p}
|
|
|
718 |
await msg.stream_token(chunk)
|
719 |
|
720 |
cl.user_session.set("NCS" + romeListArray[0], msg.content)
|
@@ -926,7 +933,7 @@ async def IA():
|
|
926 |
repo_id = "mistralai/Mixtral-8x7B-Instruct-v0.1"
|
927 |
|
928 |
llm = HuggingFaceEndpoint(
|
929 |
-
repo_id=repo_id, max_new_tokens=5000, temperature=0.7, streaming=True
|
930 |
)
|
931 |
return llm
|
932 |
|
|
|
693 |
async def construction_NCS(romeListArray):
|
694 |
context = await contexte(romeListArray)
|
695 |
emploisST = cl.user_session.get("EmploiST")
|
696 |
+
memory = ConversationBufferMemory(return_messages=True)
|
697 |
### Mistral Completion ###
|
698 |
client_llm = await IA()
|
699 |
structure = str(modele('Note de composante sectorielle'))
|
|
|
711 |
prompt = PromptTemplate(template=template, input_variables=["question","context"])
|
712 |
#llm_chain = LLMChain(prompt=prompt, llm=client_llm)
|
713 |
#completion_NCS = llm_chain.run({"question":question_p,"context":context_p}, callbacks=[StreamingStdOutCallbackHandler()])
|
714 |
+
chain = (
|
715 |
+
RunnablePassthrough.assign(
|
716 |
+
history=RunnableLambda(memory.load_memory_variables) | itemgetter("history")
|
717 |
+
)
|
718 |
+
| prompt | client_llm
|
719 |
+
)
|
720 |
#completion_NCS = chain.invoke({"question":question_p,"context":context_p})
|
721 |
|
722 |
msg = cl.Message(author="Datapcc : πππ",content="")
|
723 |
+
async for chunk in chain.astream({"question":question_p,"context":context_p},
|
724 |
+
config=RunnableConfig(callbacks=[cl.AsyncLangchainCallbackHandler(stream_final_answer=True)]):
|
725 |
await msg.stream_token(chunk)
|
726 |
|
727 |
cl.user_session.set("NCS" + romeListArray[0], msg.content)
|
|
|
933 |
repo_id = "mistralai/Mixtral-8x7B-Instruct-v0.1"
|
934 |
|
935 |
llm = HuggingFaceEndpoint(
|
936 |
+
repo_id=repo_id, max_new_tokens=5000, temperature=0.7, task="text2text-generation", streaming=True
|
937 |
)
|
938 |
return llm
|
939 |
|