From 0a056a7c5c0be993321174706d8b679e7ffde038 Mon Sep 17 00:00:00 2001 From: dupl Date: Sat, 21 Dec 2024 08:00:37 +0800 Subject: [PATCH 1/2] add gemini-exp-1206, gemini-2.0-flash-thinking-exp-1219 --- app/constant.ts | 9 +++++++++ 1 file changed, 9 insertions(+) diff --git a/app/constant.ts b/app/constant.ts index 429d52b3db4..28700cb1439 100644 --- a/app/constant.ts +++ b/app/constant.ts @@ -325,6 +325,15 @@ const googleModels = [ "gemini-exp-1121", "learnlm-1.5-pro-experimental", "gemini-2.0-flash-exp", + "gemini-1.5-flash", + "gemini-1.5-flash-002", + "gemini-1.5-flash-8b", + "gemini-1.5-flash-exp-0827", + "gemini-1.5-pro", + "gemini-1.5-pro-002", + "gemini-1.5-pro-exp-0827", + "gemini-2.0-flash-thinking-exp-1219", + "gemini-exp-1206", ]; const anthropicModels = [ From 1d8029301b4e5fcfe044a9ac55c53d2fa6d08871 Mon Sep 17 00:00:00 2001 From: dupl Date: Sat, 21 Dec 2024 08:17:58 +0800 Subject: [PATCH 2/2] support glm-4v-flash --- app/client/platforms/openai.ts | 2 +- app/utils.ts | 1 + 2 files changed, 2 insertions(+), 1 deletion(-) diff --git a/app/client/platforms/openai.ts b/app/client/platforms/openai.ts index 15cfb7ca602..b7b37492c82 100644 --- a/app/client/platforms/openai.ts +++ b/app/client/platforms/openai.ts @@ -240,7 +240,7 @@ export class ChatGPTApi implements LLMApi { } // add max_tokens to vision model - if (visionModel) { + if (visionModel && modelConfig.model !== "glm-4v-flash") { requestPayload["max_tokens"] = Math.max(modelConfig.max_tokens, 4000); } } diff --git a/app/utils.ts b/app/utils.ts index 30f95fa02eb..169c6e182b2 100644 --- a/app/utils.ts +++ b/app/utils.ts @@ -265,6 +265,7 @@ export function isVisionModel(model: string) { "learnlm", "qwen-vl", "qwen2-vl", + "glm-4v", ]; const isGpt4Turbo = model.includes("gpt-4-turbo") && !model.includes("preview");