From d01013b6af132e9050ae76db7f92d54e5ca34f71 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?=E6=81=8D=E5=85=AE=E6=83=9A=E5=85=AE?= <101191390+HIllya51@users.noreply.github.com> Date: Tue, 13 Aug 2024 00:58:28 +0800 Subject: [PATCH] Update translatorsetting.json --- .../files/defaultconfig/translatorsetting.json | 16 ++++++++-------- 1 file changed, 8 insertions(+), 8 deletions(-) diff --git a/LunaTranslator/files/defaultconfig/translatorsetting.json b/LunaTranslator/files/defaultconfig/translatorsetting.json index 467c7c82..5de1abe8 100644 --- a/LunaTranslator/files/defaultconfig/translatorsetting.json +++ b/LunaTranslator/files/defaultconfig/translatorsetting.json @@ -204,7 +204,7 @@ "使用自定义promt": false, "自定义promt": "", "api_type": 0, - "流式输出": false, + "流式输出": true, "user_user_prompt": "{sentence}", "use_user_user_prompt": false, "other_args": "{}", @@ -302,7 +302,7 @@ "user_prompt": "", "user_user_prompt": "{sentence}", "use_user_user_prompt": false, - "usingstream": false, + "usingstream": true, "other_args": "{}", "use_other_args": false, "s": "" @@ -374,7 +374,7 @@ "附带上下文个数": 0, "使用自定义promt": false, "自定义promt": "", - "流式输出": false, + "流式输出": true, "user_user_prompt": "{sentence}", "use_user_user_prompt": false, "other_args": "{}", @@ -444,7 +444,7 @@ "SECRET_KEY": "", "使用自定义promt": false, "自定义promt": "", - "流式输出": false, + "流式输出": true, "Temperature": 0.3, "top_p": 0.3, "max_tokens": 128, @@ -530,7 +530,7 @@ "SECRET_KEY": "", "use_user_prompt": false, "user_prompt": "", - "usingstream": false, + "usingstream": true, "Temperature": 0.3, "top_p": 0.3, "max_tokens": 128, @@ -618,7 +618,7 @@ "附带上下文个数": 0, "使用自定义promt": false, "自定义promt": "", - "流式输出": false, + "流式输出": true, "use_user_user_prompt": false, "user_user_prompt": "{sentence}", "other_args": "{}", @@ -1098,7 +1098,7 @@ "repetition_penalty": 1, "frequency_penalty": 0, "fix_degeneration": true, - "流式输出": false, + "流式输出": true, "prompt_version": 0, "s": "" }, @@ -1193,7 +1193,7 @@ "SECRET_KEY": "", "使用自定义promt": false, "自定义promt": "", - "流式输出": false, + "流式输出": true, "Temperature": 0.3, "top_p": 0.3, "max_tokens": 128,