Spaces:
Sleeping
Sleeping
fastapi_django_main_live
/
open-interpreter
/interpreter
/terminal_interface
/contributing_conversations.py
import json | |
import os | |
import time | |
from typing import List, TypedDict | |
import pkg_resources | |
import requests | |
from interpreter.terminal_interface.profiles.profiles import write_key_to_profile | |
from interpreter.terminal_interface.utils.display_markdown_message import ( | |
display_markdown_message, | |
) | |
contribute_cache_path = os.path.join( | |
os.path.expanduser("~"), ".cache", "open-interpreter", "contribute.json" | |
) | |
def display_contribution_message(): | |
display_markdown_message( | |
""" | |
--- | |
> We're training an open-source language model. | |
Want to contribute? Run `interpreter --model i` to use our free, hosted model. Conversations with this `i` model will be used for training. | |
""" | |
) | |
time.sleep(1) | |
def display_contributing_current_message(): | |
display_markdown_message( | |
f""" | |
--- | |
> This conversation will be used to train Open Interpreter's open-source language model. | |
""" | |
) | |
def send_past_conversations(interpreter): | |
past_conversations = get_all_conversations(interpreter) | |
if len(past_conversations) > 0: | |
print() | |
print( | |
"We are about to send all previous conversations to Open Interpreter for training an open-source language model. Please make sure these don't contain any private information. Run `interpreter --conversations` to browse them." | |
) | |
print() | |
time.sleep(2) | |
uh = input( | |
"Do we have your permission to send all previous conversations to Open Interpreter? (y/n): " | |
) | |
print() | |
if uh == "y": | |
print("Sending all previous conversations to OpenInterpreter...") | |
contribute_conversations(past_conversations) | |
print() | |
def set_send_future_conversations(interpreter, should_send_future): | |
write_key_to_profile("contribute_conversation", should_send_future) | |
display_markdown_message( | |
""" | |
> Open Interpreter will contribute conversations from now on. Thank you for your help! | |
To change this, run `interpreter --profiles` and edit the `default.yaml` profile so "contribute_conversation" = False. | |
""" | |
) | |
def user_wants_to_contribute_past(): | |
print("\nWould you like to contribute all past conversations?\n") | |
response = input("(y/n) ") | |
return response.lower() == "y" | |
def user_wants_to_contribute_future(): | |
print("\nWould you like to contribute all future conversations?\n") | |
response = input("(y/n) ") | |
return response.lower() == "y" | |
def contribute_conversation_launch_logic(interpreter): | |
contribution_cache = get_contribute_cache_contents() | |
if interpreter.will_contribute: | |
contribute_past_and_future_logic(interpreter, contribution_cache) | |
elif not contribution_cache["displayed_contribution_message"]: | |
display_contribution_message() | |
# don't show the contribution message again no matter what. | |
contribution_cache["displayed_contribution_message"] = True | |
write_to_contribution_cache(contribution_cache) | |
class ContributionCache(TypedDict): | |
displayed_contribution_message: bool | |
asked_to_contribute_past: bool | |
asked_to_contribute_future: bool | |
# modifies the contribution cache! | |
def contribute_past_and_future_logic( | |
interpreter, contribution_cache: ContributionCache | |
): | |
if not contribution_cache["asked_to_contribute_past"]: | |
if user_wants_to_contribute_past(): | |
send_past_conversations(interpreter) | |
contribution_cache["asked_to_contribute_past"] = True | |
if not contribution_cache["asked_to_contribute_future"]: | |
if user_wants_to_contribute_future(): | |
set_send_future_conversations(interpreter, True) | |
contribution_cache["asked_to_contribute_future"] = True | |
display_contributing_current_message() | |
# Returns a {"asked_to_run_contribute": bool, "asked_to_contribute_past": bool} | |
# as the first part of its Tuple, a bool as a second. | |
# Writes the contribution cache file if it doesn't already exist. | |
# The bool is True if the file does not already exist, False if it does. | |
def get_contribute_cache_contents() -> ContributionCache: | |
if not os.path.exists(contribute_cache_path): | |
default_dict: ContributionCache = { | |
"asked_to_contribute_past": False, | |
"displayed_contribution_message": False, | |
"asked_to_contribute_future": False, | |
} | |
with open(contribute_cache_path, "a") as file: | |
file.write(json.dumps(default_dict)) | |
return default_dict | |
else: | |
with open(contribute_cache_path, "r") as file: | |
contribute_cache = json.load(file) | |
return contribute_cache | |
# Takes in a {"asked_to_run_contribute": bool, "asked_to_contribute_past": bool} | |
def write_to_contribution_cache(contribution_cache: ContributionCache): | |
with open(contribute_cache_path, "w") as file: | |
json.dump(contribution_cache, file) | |
def get_all_conversations(interpreter) -> List[List]: | |
def is_conversation_path(path: str): | |
_, ext = os.path.splitext(path) | |
return ext == ".json" | |
history_path = interpreter.conversation_history_path | |
all_conversations: List[List] = [] | |
conversation_files = ( | |
os.listdir(history_path) if os.path.exists(history_path) else [] | |
) | |
for mpath in conversation_files: | |
if not is_conversation_path(mpath): | |
continue | |
full_path = os.path.join(history_path, mpath) | |
with open(full_path, "r") as cfile: | |
conversation = json.load(cfile) | |
all_conversations.append(conversation) | |
return all_conversations | |
def is_list_of_lists(l): | |
return isinstance(l, list) and all([isinstance(e, list) for e in l]) | |
def contribute_conversations( | |
conversations: List[List], feedback=None, conversation_id=None | |
): | |
if len(conversations) == 0 or len(conversations[0]) == 0: | |
return None | |
url = "https://api.openinterpreter.com/v0/contribute/" | |
version = pkg_resources.get_distribution("open-interpreter").version | |
payload = { | |
"conversation_id": conversation_id, | |
"conversations": conversations, | |
"oi_version": version, | |
"feedback": feedback, | |
} | |
assert is_list_of_lists( | |
payload["conversations"] | |
), "the contribution payload is not a list of lists!" | |
try: | |
requests.post(url, json=payload) | |
except: | |
# Non blocking | |
pass | |