From ca7ff47fcb3105ba3134196161c3f557be693cf2 Mon Sep 17 00:00:00 2001 From: ji-jinlong <61379293+ji-jinlong@users.noreply.github.com> Date: Wed, 1 Nov 2023 16:05:57 +0800 Subject: [PATCH 1/2] =?UTF-8?q?Update=20=E7=90=86=E8=A7=A3PDF=E6=96=87?= =?UTF-8?q?=E6=A1=A3=E5=86=85=E5=AE=B9.py?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- ...7\243PDF\346\226\207\346\241\243\345\206\205\345\256\271.py" | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git "a/crazy_functions/\347\220\206\350\247\243PDF\346\226\207\346\241\243\345\206\205\345\256\271.py" "b/crazy_functions/\347\220\206\350\247\243PDF\346\226\207\346\241\243\345\206\205\345\256\271.py" index f1a89a7ec9..4674b935ed 100644 --- "a/crazy_functions/\347\220\206\350\247\243PDF\346\226\207\346\241\243\345\206\205\345\256\271.py" +++ "b/crazy_functions/\347\220\206\350\247\243PDF\346\226\207\346\241\243\345\206\205\345\256\271.py" @@ -49,7 +49,7 @@ def get_token_num(txt): return len(enc.encode(txt, disallowed_special=())) gpt_say = yield from request_gpt_model_in_new_thread_with_ui_alive(i_say, i_say_show_user, # i_say=真正给chatgpt的提问, i_say_show_user=给用户看的提问 llm_kwargs, chatbot, history=["The main idea of the previous section is?", last_iteration_result], # 迭代上一次的结果 - sys_prompt="Extract the main idea of this section." # 提示 + sys_prompt="Extract the main idea of this section.用中文回答我" # 提示 ) iteration_results.append(gpt_say) last_iteration_result = gpt_say From f016323b8aafe2c4c3d4dd4561fdfc8d7cbf234d Mon Sep 17 00:00:00 2001 From: binary-husky <96192199+binary-husky@users.noreply.github.com> Date: Wed, 1 Nov 2023 22:26:46 +0800 Subject: [PATCH 2/2] =?UTF-8?q?Update=20=E7=90=86=E8=A7=A3PDF=E6=96=87?= =?UTF-8?q?=E6=A1=A3=E5=86=85=E5=AE=B9.py?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- ...7\243PDF\346\226\207\346\241\243\345\206\205\345\256\271.py" | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git "a/crazy_functions/\347\220\206\350\247\243PDF\346\226\207\346\241\243\345\206\205\345\256\271.py" "b/crazy_functions/\347\220\206\350\247\243PDF\346\226\207\346\241\243\345\206\205\345\256\271.py" index 4674b935ed..3b2db998c6 100644 --- "a/crazy_functions/\347\220\206\350\247\243PDF\346\226\207\346\241\243\345\206\205\345\256\271.py" +++ "b/crazy_functions/\347\220\206\350\247\243PDF\346\226\207\346\241\243\345\206\205\345\256\271.py" @@ -49,7 +49,7 @@ def get_token_num(txt): return len(enc.encode(txt, disallowed_special=())) gpt_say = yield from request_gpt_model_in_new_thread_with_ui_alive(i_say, i_say_show_user, # i_say=真正给chatgpt的提问, i_say_show_user=给用户看的提问 llm_kwargs, chatbot, history=["The main idea of the previous section is?", last_iteration_result], # 迭代上一次的结果 - sys_prompt="Extract the main idea of this section.用中文回答我" # 提示 + sys_prompt="Extract the main idea of this section, answer me with Chinese." # 提示 ) iteration_results.append(gpt_say) last_iteration_result = gpt_say