Spaces:
Runtime error
Runtime error
dev(hansbug): merged
Browse files- README.md +12 -9
- app.py +16 -4
- llmriddles/assets/avatar.png +0 -0
- llmriddles/llms/__init__.py +2 -1
- llmriddles/llms/llm_client.py +54 -0
- llmriddles/llms/llm_server.py +72 -0
- llmriddles/llms/mistral.py +22 -0
- llmriddles/questions/__init__.py +4 -0
- llmriddles/questions/level4.py +97 -0
- llmriddles/questions/level5.py +56 -0
- requirements.txt +3 -1
README.md
CHANGED
@@ -29,30 +29,33 @@ QUESTION_LANG=cn QUESTION_LLM='chatgpt' QUESTION_LLM_KEY=<your API key> python3
|
|
29 |
```shell
|
30 |
QUESTION_LANG=en QUESTION_LLM='chatgpt' QUESTION_LLM_KEY=<your API key> python3 -u app.py
|
31 |
```
|
32 |
-
###
|
33 |
```shell
|
34 |
-
QUESTION_LANG=
|
35 |
-
```
|
36 |
-
### LLaMA2-7b + 英文
|
37 |
-
```shell
|
38 |
-
QUESTION_LANG=en QUESTION_LLM='llama2-7b' python3 -u app.py
|
39 |
```
|
40 |
## :technologist: 为什么制作这个游戏
|
41 |
|
42 |
我们的目标是通过这一游戏,让参与者深入领略到提示工程(prompt engineering)和自然语言处理的令人着迷之处。这个过程将向玩家们展示,如何巧妙地构建提示词(prompts),以及如何运用它们来引发人工智能系统的惊人反应,同时也帮助他们更好地理解深度学习和自然语言处理技术的不可思议之处。
|
43 |
|
44 |
## :raising_hand: 如何提交设计好的关卡
|
45 |
-
|
46 |
[发起 Pull Request](https://github.com/opendilab/LLMRiddles/compare) 向我们提交, 我们会在审核通过后收录至关卡中。
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
47 |
|
48 |
## :writing_hand: 未来计划
|
49 |
|
50 |
- [x] 支持自定义关卡
|
51 |
- [ ] 在线试玩链接
|
52 |
- [ ] Hugging Face Space 链接
|
53 |
-
- [
|
54 |
-
- [ ] 支持Mistral-7B(英文)
|
55 |
- [ ] 支持Baichuan2-7B(中文)
|
|
|
56 |
- [ ] LLM 推理速度优化
|
57 |
|
58 |
|
|
|
29 |
```shell
|
30 |
QUESTION_LANG=en QUESTION_LLM='chatgpt' QUESTION_LLM_KEY=<your API key> python3 -u app.py
|
31 |
```
|
32 |
+
### Mistral-7B-Instruct-v0.1 + 英文
|
33 |
```shell
|
34 |
+
QUESTION_LANG=en QUESTION_LLM='mistral-7b' python3 -u app.py
|
|
|
|
|
|
|
|
|
35 |
```
|
36 |
## :technologist: 为什么制作这个游戏
|
37 |
|
38 |
我们的目标是通过这一游戏,让参与者深入领略到提示工程(prompt engineering)和自然语言处理的令人着迷之处。这个过程将向玩家们展示,如何巧妙地构建提示词(prompts),以及如何运用它们来引发人工智能系统的惊人反应,同时也帮助他们更好地理解深度学习和自然语言处理技术的不可思议之处。
|
39 |
|
40 |
## :raising_hand: 如何提交设计好的关卡
|
41 |
+
如果有好玩的问题或想法,欢迎玩家提交自己的创意,可以
|
42 |
[发起 Pull Request](https://github.com/opendilab/LLMRiddles/compare) 向我们提交, 我们会在审核通过后收录至关卡中。
|
43 |
+
问题的设计格式应包含以下几点:
|
44 |
+
- Pull Request标题,示例:feature(username): 章节X-关卡设计
|
45 |
+
- 希望被提及的ID
|
46 |
+
- 对应章节问题文件的修改
|
47 |
+
- init.py的修改
|
48 |
+
|
49 |
+
完整示例请参考:[提交属于自己的关卡设计]()
|
50 |
|
51 |
## :writing_hand: 未来计划
|
52 |
|
53 |
- [x] 支持自定义关卡
|
54 |
- [ ] 在线试玩链接
|
55 |
- [ ] Hugging Face Space 链接
|
56 |
+
- [x] 支持Mistral-7B-Instruct-v0.1(英文)
|
|
|
57 |
- [ ] 支持Baichuan2-7B(中文)
|
58 |
+
- [ ] 支持LLaMA2-7B(英文)
|
59 |
- [ ] LLM 推理速度优化
|
60 |
|
61 |
|
app.py
CHANGED
@@ -1,5 +1,6 @@
|
|
1 |
import os
|
2 |
import uuid
|
|
|
3 |
|
4 |
import gradio as gr
|
5 |
|
@@ -7,14 +8,20 @@ from llmriddles.questions import QuestionExecutor
|
|
7 |
from llmriddles.questions import list_ordered_questions
|
8 |
|
9 |
_QUESTION_IDS = {}
|
|
|
10 |
_QUESTIONS = list_ordered_questions()
|
11 |
_LANG = os.environ.get('QUESTION_LANG', 'cn')
|
12 |
assert _LANG in ['cn', 'en'], _LANG
|
13 |
_LLM = os.environ.get('QUESTION_LLM', 'chatgpt')
|
14 |
-
assert _LLM in ['chatgpt', '
|
15 |
_LLM_KEY = os.environ.get('QUESTION_LLM_KEY', None)
|
|
|
16 |
|
17 |
if _LANG == "cn":
|
|
|
|
|
|
|
|
|
18 |
title = "完蛋!我被 LLM 拿捏了"
|
19 |
requirement_ph = """
|
20 |
欢迎来到 LLM Riddles!
|
@@ -122,7 +129,7 @@ if __name__ == '__main__':
|
|
122 |
gr_question = gr.TextArea(placeholder=question_ph, label=question_label)
|
123 |
gr_api_key = gr.Text(placeholder=api_ph, label=api_label, type='password', visible=_need_api_key())
|
124 |
with gr.Row():
|
125 |
-
gr_submit = gr.Button(submit_label, interactive=
|
126 |
gr_next = gr.Button(next_label)
|
127 |
|
128 |
with gr.Column():
|
@@ -134,8 +141,11 @@ if __name__ == '__main__':
|
|
134 |
|
135 |
|
136 |
def _next_question(uuid_):
|
|
|
137 |
if not uuid_:
|
138 |
uuid_ = str(uuid.uuid4())
|
|
|
|
|
139 |
global _QUESTION_IDS
|
140 |
_qid = _QUESTION_IDS.get(uuid_, -1)
|
141 |
_qid += 1
|
@@ -143,8 +153,9 @@ if __name__ == '__main__':
|
|
143 |
|
144 |
if _qid >= len(_QUESTIONS):
|
145 |
del _QUESTION_IDS[uuid_]
|
|
|
146 |
return game_cleared_label, '', '', {}, '', \
|
147 |
-
gr.Button(submit_label, interactive=
|
148 |
gr.Button(try_again_label, interactive=True), \
|
149 |
''
|
150 |
else:
|
@@ -187,4 +198,5 @@ if __name__ == '__main__':
|
|
187 |
)
|
188 |
|
189 |
concurrency = int(os.environ.get('CONCURRENCY', os.cpu_count()))
|
190 |
-
|
|
|
|
1 |
import os
|
2 |
import uuid
|
3 |
+
import logging
|
4 |
|
5 |
import gradio as gr
|
6 |
|
|
|
8 |
from llmriddles.questions import list_ordered_questions
|
9 |
|
10 |
_QUESTION_IDS = {}
|
11 |
+
count = 0
|
12 |
_QUESTIONS = list_ordered_questions()
|
13 |
_LANG = os.environ.get('QUESTION_LANG', 'cn')
|
14 |
assert _LANG in ['cn', 'en'], _LANG
|
15 |
_LLM = os.environ.get('QUESTION_LLM', 'chatgpt')
|
16 |
+
assert _LLM in ['chatgpt', 'mistral-7b'], _LLM
|
17 |
_LLM_KEY = os.environ.get('QUESTION_LLM_KEY', None)
|
18 |
+
_DEBUG = os.environ.get('DEBUG', 'false').lower() == 'true'
|
19 |
|
20 |
if _LANG == "cn":
|
21 |
+
if _DEBUG:
|
22 |
+
logging.getLogger().setLevel(logging.INFO)
|
23 |
+
else:
|
24 |
+
logging.getLogger().setLevel(logging.WARNING)
|
25 |
title = "完蛋!我被 LLM 拿捏了"
|
26 |
requirement_ph = """
|
27 |
欢迎来到 LLM Riddles!
|
|
|
129 |
gr_question = gr.TextArea(placeholder=question_ph, label=question_label)
|
130 |
gr_api_key = gr.Text(placeholder=api_ph, label=api_label, type='password', visible=_need_api_key())
|
131 |
with gr.Row():
|
132 |
+
gr_submit = gr.Button(submit_label, interactive=False)
|
133 |
gr_next = gr.Button(next_label)
|
134 |
|
135 |
with gr.Column():
|
|
|
141 |
|
142 |
|
143 |
def _next_question(uuid_):
|
144 |
+
global count
|
145 |
if not uuid_:
|
146 |
uuid_ = str(uuid.uuid4())
|
147 |
+
count += 1
|
148 |
+
logging.info(f'Player {count} starts the game now')
|
149 |
global _QUESTION_IDS
|
150 |
_qid = _QUESTION_IDS.get(uuid_, -1)
|
151 |
_qid += 1
|
|
|
153 |
|
154 |
if _qid >= len(_QUESTIONS):
|
155 |
del _QUESTION_IDS[uuid_]
|
156 |
+
logging.info(f'Player {count} has passed the game now')
|
157 |
return game_cleared_label, '', '', {}, '', \
|
158 |
+
gr.Button(submit_label, interactive=False), \
|
159 |
gr.Button(try_again_label, interactive=True), \
|
160 |
''
|
161 |
else:
|
|
|
198 |
)
|
199 |
|
200 |
concurrency = int(os.environ.get('CONCURRENCY', os.cpu_count()))
|
201 |
+
favicon_path = os.path.join(os.path.dirname(__file__), 'llmriddles', 'assets', 'avatar.png')
|
202 |
+
demo.queue().launch(max_threads=concurrency, favicon_path=favicon_path, share=True)
|
llmriddles/assets/avatar.png
ADDED
llmriddles/llms/__init__.py
CHANGED
@@ -1,2 +1,3 @@
|
|
1 |
-
from .chatgpt import ask_chatgpt
|
2 |
from .base import register_llm, get_llm_fn
|
|
|
|
|
|
|
|
1 |
from .base import register_llm, get_llm_fn
|
2 |
+
from .chatgpt import ask_chatgpt
|
3 |
+
from .mistral import ask_mistral_7b_instruct
|
llmriddles/llms/llm_client.py
ADDED
@@ -0,0 +1,54 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
import time
|
2 |
+
import requests
|
3 |
+
import logging
|
4 |
+
import argparse
|
5 |
+
|
6 |
+
|
7 |
+
class LLMFlaskClient:
|
8 |
+
def __init__(self, ip: str, port: int, max_retry: int = 3):
|
9 |
+
self.ip = ip
|
10 |
+
self.port = port
|
11 |
+
|
12 |
+
self.url_prefix_format = 'http://{}:{}/'
|
13 |
+
self.url = self.url_prefix_format.format(self.ip, self.port)
|
14 |
+
self.max_retry = max_retry
|
15 |
+
|
16 |
+
self.logger = logging.getLogger()
|
17 |
+
self.logger.addHandler(logging.StreamHandler())
|
18 |
+
self.logger.handlers[0].setFormatter(logging.Formatter("%(message)s"))
|
19 |
+
|
20 |
+
def _request(self, name: str, data: dict):
|
21 |
+
for _ in range(self.max_retry):
|
22 |
+
try:
|
23 |
+
self.logger.info(f'{name}\ndata: {data}')
|
24 |
+
response = requests.post(self.url + name, json=data).json()
|
25 |
+
except Exception as e:
|
26 |
+
self.logger.warning('error: ', repr(e))
|
27 |
+
time.sleep(1)
|
28 |
+
continue
|
29 |
+
if response['code'] == 0:
|
30 |
+
return response['output']
|
31 |
+
else:
|
32 |
+
raise Exception(response['error_msg'])
|
33 |
+
raise Exception("Web service failed. Please retry or contact with manager")
|
34 |
+
|
35 |
+
def run(self, message: str) -> str:
|
36 |
+
try:
|
37 |
+
return self._request('ask_llm_for_answer', {'user_text': message})
|
38 |
+
except Exception as e:
|
39 |
+
return f"Error: {repr(e)}"
|
40 |
+
|
41 |
+
|
42 |
+
if __name__ == "__main__":
|
43 |
+
parser = argparse.ArgumentParser()
|
44 |
+
parser.add_argument('--ip', required=True)
|
45 |
+
parser.add_argument('-p', '--port', required=True)
|
46 |
+
parser.add_argument('--debug', action='store_true')
|
47 |
+
args = parser.parse_args()
|
48 |
+
if args.debug:
|
49 |
+
logging.getLogger().setLevel(logging.INFO)
|
50 |
+
else:
|
51 |
+
logging.getLogger().setLevel(logging.WARNING)
|
52 |
+
|
53 |
+
client = LLMFlaskClient(args.ip, args.port)
|
54 |
+
print(client.run('Please concatenate string "1+" and "1=3". Only give me the result without "".'))
|
llmriddles/llms/llm_server.py
ADDED
@@ -0,0 +1,72 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
from transformers import AutoModelForCausalLM, AutoTokenizer
|
2 |
+
from flask import Flask, request
|
3 |
+
import argparse
|
4 |
+
import logging
|
5 |
+
|
6 |
+
|
7 |
+
class LLMInstance:
|
8 |
+
|
9 |
+
def __init__(self, model_path: str, device: str = "cuda"):
|
10 |
+
|
11 |
+
self.model = AutoModelForCausalLM.from_pretrained(model_path)
|
12 |
+
self.tokenizer = AutoTokenizer.from_pretrained(model_path)
|
13 |
+
self.model.to(device)
|
14 |
+
self.device = device
|
15 |
+
|
16 |
+
def query(self, message):
|
17 |
+
try:
|
18 |
+
messages = [
|
19 |
+
{"role": "user", "content": message},
|
20 |
+
]
|
21 |
+
encodeds = self.tokenizer.apply_chat_template(messages, return_tensors="pt")
|
22 |
+
model_inputs = encodeds.to(self.device)
|
23 |
+
|
24 |
+
generated_ids = self.model.generate(model_inputs, max_new_tokens=1000, do_sample=True)
|
25 |
+
decoded = self.tokenizer.batch_decode(generated_ids)
|
26 |
+
|
27 |
+
# output is the string decoded[0] after "[/INST]". There may exist "</s>", delete it.
|
28 |
+
output = decoded[0].split("[/INST]")[1].split("</s>")[0]
|
29 |
+
return {
|
30 |
+
'code': 0,
|
31 |
+
'ret': True,
|
32 |
+
'error_msg': None,
|
33 |
+
'output': output
|
34 |
+
}
|
35 |
+
except Exception as e:
|
36 |
+
return {
|
37 |
+
'code': 1,
|
38 |
+
'ret': False,
|
39 |
+
'error_msg': str(e),
|
40 |
+
'output': None
|
41 |
+
}
|
42 |
+
|
43 |
+
|
44 |
+
def create_app(core):
|
45 |
+
app = Flask(__name__)
|
46 |
+
|
47 |
+
@app.route('/ask_llm_for_answer', methods=['POST'])
|
48 |
+
def ask_llm_for_answer():
|
49 |
+
user_text = request.json['user_text']
|
50 |
+
return core.query(user_text)
|
51 |
+
|
52 |
+
return app
|
53 |
+
|
54 |
+
|
55 |
+
if __name__ == "__main__":
|
56 |
+
parser = argparse.ArgumentParser()
|
57 |
+
parser.add_argument('-m', '--model_path', required=True, default='Mistral-7B-Instruct-v0.1', help='the model path of reward model')
|
58 |
+
parser.add_argument('--ip', default='0.0.0.0')
|
59 |
+
parser.add_argument('-p', '--port', default=8001)
|
60 |
+
parser.add_argument('--debug', action='store_true')
|
61 |
+
args = parser.parse_args()
|
62 |
+
|
63 |
+
if args.debug:
|
64 |
+
logging.getLogger().setLevel(logging.DEBUG)
|
65 |
+
else:
|
66 |
+
logging.getLogger().setLevel(logging.INFO)
|
67 |
+
logging.getLogger().addHandler(logging.StreamHandler())
|
68 |
+
logging.getLogger().handlers[0].setFormatter(logging.Formatter("%(message)s"))
|
69 |
+
|
70 |
+
core = LLMInstance(args.model_path)
|
71 |
+
app = create_app(core)
|
72 |
+
app.run(host=args.ip, port=args.port)
|
llmriddles/llms/mistral.py
ADDED
@@ -0,0 +1,22 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
from functools import lru_cache
|
2 |
+
|
3 |
+
from .base import register_llm
|
4 |
+
from .llm_client import LLMFlaskClient
|
5 |
+
|
6 |
+
|
7 |
+
@lru_cache()
|
8 |
+
def _get_mistral_7b_instruct_server(host: str, port: int):
|
9 |
+
from .llm_server import LLMInstance, create_app
|
10 |
+
core = LLMInstance('Mistral-7B-Instruct-v0.1')
|
11 |
+
app = create_app(core)
|
12 |
+
app.run(host=host, port=port)
|
13 |
+
|
14 |
+
|
15 |
+
def ask_mistral_7b_instruct(message: str, **kwargs):
|
16 |
+
host, port = '0.0.0.0', 8001
|
17 |
+
_get_mistral_7b_instruct_server(host, port)
|
18 |
+
client = LLMFlaskClient(host, port)
|
19 |
+
return client.run(message).strip()
|
20 |
+
|
21 |
+
|
22 |
+
register_llm('mistral-7b', ask_mistral_7b_instruct)
|
llmriddles/questions/__init__.py
CHANGED
@@ -2,8 +2,12 @@ from .executor import QuestionExecutor
|
|
2 |
from .level1 import __file__ as _level1_file_
|
3 |
from .level2 import __file__ as _level2_file_
|
4 |
from .level3 import __file__ as _level3_file_
|
|
|
|
|
5 |
from .question import Question, register_question, list_ordered_questions
|
6 |
|
7 |
_ = _level1_file_
|
8 |
_ = _level2_file_
|
9 |
_ = _level3_file_
|
|
|
|
|
|
2 |
from .level1 import __file__ as _level1_file_
|
3 |
from .level2 import __file__ as _level2_file_
|
4 |
from .level3 import __file__ as _level3_file_
|
5 |
+
from .level4 import __file__ as _level4_file_
|
6 |
+
from .level5 import __file__ as _level5_file_
|
7 |
from .question import Question, register_question, list_ordered_questions
|
8 |
|
9 |
_ = _level1_file_
|
10 |
_ = _level2_file_
|
11 |
_ = _level3_file_
|
12 |
+
_ = _level4_file_
|
13 |
+
_ = _level5_file_
|
llmriddles/questions/level4.py
ADDED
@@ -0,0 +1,97 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
from .question import register_question
|
2 |
+
import re
|
3 |
+
|
4 |
+
|
5 |
+
def check_if_is_number(text: str):
|
6 |
+
try:
|
7 |
+
int(text)
|
8 |
+
return True
|
9 |
+
except ValueError:
|
10 |
+
return False
|
11 |
+
|
12 |
+
|
13 |
+
def get_all_numbers_in_a_sentence(text: str):
|
14 |
+
return [int(i) for i in re.findall(r'\d+', text)]
|
15 |
+
|
16 |
+
|
17 |
+
# CN_TEXT_1 = """
|
18 |
+
# 第四章第一题,请构造一个问题A,它的回复是不同于A的B,然后以B提问能再次得到A。
|
19 |
+
|
20 |
+
# 请在下面的输入框内填写你构造并点击按钮提交。
|
21 |
+
# """
|
22 |
+
|
23 |
+
# EN_TEXT_1 = """
|
24 |
+
# For the first question in chapter 4, please make a question A whose answer is B that is different from A, and then ask B to get A again.
|
25 |
+
|
26 |
+
# Please enter your query below and click the submit button
|
27 |
+
# """
|
28 |
+
|
29 |
+
|
30 |
+
# def _checker_1(question_text: str, user_text: str, answer_text: str, lang: str):
|
31 |
+
# _ = question_text, lang
|
32 |
+
# answer_text = answer_text.strip()
|
33 |
+
# user_text = user_text.strip()
|
34 |
+
# pass
|
35 |
+
|
36 |
+
# register_question({
|
37 |
+
# 'cn': CN_TEXT_1,
|
38 |
+
# 'en': EN_TEXT_1,
|
39 |
+
# }, _checker_1, level=4)
|
40 |
+
|
41 |
+
|
42 |
+
# CN_TEXT_2 = """
|
43 |
+
# 第四章第二题,
|
44 |
+
|
45 |
+
# 请在下面的输入框内填写你构造并点击按钮提交。
|
46 |
+
# """
|
47 |
+
|
48 |
+
# EN_TEXT_2 = """
|
49 |
+
# For the second question in chapter 4,
|
50 |
+
|
51 |
+
# Please enter your query below and click the submit button
|
52 |
+
# """
|
53 |
+
|
54 |
+
|
55 |
+
# def _checker_2(question_text: str, user_text: str, answer_text: str, lang: str):
|
56 |
+
# _ = question_text, lang
|
57 |
+
# answer_text = answer_text.strip()
|
58 |
+
# user_text = user_text.strip()
|
59 |
+
# pass
|
60 |
+
|
61 |
+
# register_question({
|
62 |
+
# 'cn': CN_TEXT_2,
|
63 |
+
# 'en': EN_TEXT_2,
|
64 |
+
# }, _checker_2, level=4)
|
65 |
+
|
66 |
+
|
67 |
+
CN_TEXT_3 = """
|
68 |
+
第四章第三题(自然之密),请输入一个大于一的正整数作为问题,使回答里包含和它刚好相差1的数。
|
69 |
+
|
70 |
+
请在下面的输入框内填写你构造并点击按钮提交。
|
71 |
+
"""
|
72 |
+
|
73 |
+
EN_TEXT_3 = """
|
74 |
+
For the third question in chapter 4, please enter a positive integer greater than one as the question so that the answer contains a number that is exactly 1 different from it.
|
75 |
+
|
76 |
+
Please enter your query below and click the submit button
|
77 |
+
"""
|
78 |
+
|
79 |
+
|
80 |
+
def _checker_3(question_text: str, user_text: str, answer_text: str, lang: str):
|
81 |
+
_ = question_text, lang
|
82 |
+
answer_text = answer_text.strip()
|
83 |
+
user_text = user_text.strip()
|
84 |
+
if not check_if_is_number(question_text):
|
85 |
+
return False, "问题应该是一个正整数" if lang == 'cn' else 'Question should be a positive integer.'
|
86 |
+
elif int(question_text) == 1:
|
87 |
+
return False, "问题应该是一个大于1的正整数" if lang == 'cn' else 'Question should be a positive integer greater than 1.'
|
88 |
+
elif int(question_text)-1 not in get_all_numbers_in_a_sentence(answer_text) or int(question_text)+1 not in get_all_numbers_in_a_sentence(answer_text):
|
89 |
+
return False, "回答中应该包含一个与问题相差1的数字" if lang == 'cn' else 'Answer should contain a number that is exactly 1 different from the question.'
|
90 |
+
else:
|
91 |
+
return True, None
|
92 |
+
|
93 |
+
|
94 |
+
register_question({
|
95 |
+
'cn': CN_TEXT_3,
|
96 |
+
'en': EN_TEXT_3,
|
97 |
+
}, _checker_3, level=4)
|
llmriddles/questions/level5.py
ADDED
@@ -0,0 +1,56 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
from .question import register_question
|
2 |
+
|
3 |
+
def count_words(text: str, contain_punctuation: bool = False):
|
4 |
+
chinese_words = []
|
5 |
+
english_words = []
|
6 |
+
other_words = []
|
7 |
+
temp_english_words = []
|
8 |
+
for char in text:
|
9 |
+
if '\u4e00' <= char <= '\u9fff':
|
10 |
+
chinese_words.append(char)
|
11 |
+
if len(temp_english_words) > 0:
|
12 |
+
english_words.append(''.join(temp_english_words))
|
13 |
+
temp_english_words = []
|
14 |
+
else:
|
15 |
+
if char.isalpha():
|
16 |
+
temp_english_words.append(char)
|
17 |
+
else:
|
18 |
+
if len(temp_english_words) > 0:
|
19 |
+
english_words.append(''.join(temp_english_words))
|
20 |
+
temp_english_words = []
|
21 |
+
other_words.append(char)
|
22 |
+
if contain_punctuation:
|
23 |
+
return len(chinese_words)+len(english_words)+len(other_words)
|
24 |
+
else:
|
25 |
+
return len(chinese_words)+len(english_words)
|
26 |
+
|
27 |
+
CN_TEXT_1 = """
|
28 |
+
第五章第一题(口是心非)请构造一个不少于十个字的问题,使得回答中不包含问题中的任意字符。
|
29 |
+
|
30 |
+
请在下面的输入框内填写你的提问并点击按钮提交。
|
31 |
+
"""
|
32 |
+
|
33 |
+
EN_TEXT_1 = """
|
34 |
+
For the first question in chapter 5, Please construct a question of no less than ten words so that the answer does not contain any of the characters in the question.
|
35 |
+
|
36 |
+
Please enter your query below and click the submit button
|
37 |
+
"""
|
38 |
+
|
39 |
+
def _checker_1(question_text: str, user_text: str, answer_text: str, lang: str):
|
40 |
+
_ = question_text, lang
|
41 |
+
answer_text = answer_text.strip()
|
42 |
+
user_text = user_text.strip()
|
43 |
+
|
44 |
+
if count_words(user_text) < 10:
|
45 |
+
return False, "用户的问题长度应该至少10个字" if lang=='cn' else 'Question should be no less than 10 words.'
|
46 |
+
|
47 |
+
if answer_text == user_text:
|
48 |
+
return True, None
|
49 |
+
else:
|
50 |
+
return False, None
|
51 |
+
|
52 |
+
register_question({
|
53 |
+
'cn': CN_TEXT_1,
|
54 |
+
'en': EN_TEXT_1,
|
55 |
+
}, _checker_1, level=5)
|
56 |
+
|
requirements.txt
CHANGED
@@ -3,4 +3,6 @@ tqdm
|
|
3 |
requests>=2.20
|
4 |
gradio==4.1.1
|
5 |
openai>=1
|
6 |
-
sympy
|
|
|
|
|
|
3 |
requests>=2.20
|
4 |
gradio==4.1.1
|
5 |
openai>=1
|
6 |
+
sympy
|
7 |
+
flask
|
8 |
+
transformers
|