From 79bf76ed721936cd84b998c7e9332850f28aec91 Mon Sep 17 00:00:00 2001 From: yetone Date: Mon, 17 Mar 2025 03:29:44 +0800 Subject: [PATCH] fix: increase max_tokens (#1607) --- lua/avante/config.lua | 24 ++++++++++++------------ 1 file changed, 12 insertions(+), 12 deletions(-) diff --git a/lua/avante/config.lua b/lua/avante/config.lua index f3661aa..cf52279 100644 --- a/lua/avante/config.lua +++ b/lua/avante/config.lua @@ -195,7 +195,7 @@ M._defaults = { model = "gpt-4o", timeout = 30000, -- Timeout in milliseconds temperature = 0, - max_tokens = 8192, + max_tokens = 10240, }, ---@type AvanteSupportedProvider copilot = { @@ -205,7 +205,7 @@ M._defaults = { allow_insecure = false, -- Allow insecure server connections timeout = 30000, -- Timeout in milliseconds temperature = 0, - max_tokens = 8192, + max_tokens = 10240, }, ---@type AvanteAzureProvider azure = { @@ -214,7 +214,7 @@ M._defaults = { api_version = "2024-06-01", timeout = 30000, -- Timeout in milliseconds temperature = 0, - max_tokens = 8192, + max_tokens = 10240, }, ---@type AvanteSupportedProvider claude = { @@ -222,14 +222,14 @@ M._defaults = { model = "claude-3-7-sonnet-20250219", timeout = 30000, -- Timeout in milliseconds temperature = 0, - max_tokens = 8192, + max_tokens = 10240, }, ---@type AvanteSupportedProvider bedrock = { model = "anthropic.claude-3-5-sonnet-20241022-v2:0", timeout = 30000, -- Timeout in milliseconds temperature = 0, - max_tokens = 8192, + max_tokens = 10240, }, ---@type AvanteSupportedProvider gemini = { @@ -237,7 +237,7 @@ M._defaults = { model = "gemini-1.5-flash-latest", timeout = 30000, -- Timeout in milliseconds temperature = 0, - max_tokens = 8192, + max_tokens = 10240, }, ---@type AvanteSupportedProvider vertex = { @@ -245,7 +245,7 @@ M._defaults = { model = "gemini-1.5-flash-002", timeout = 30000, -- Timeout in milliseconds temperature = 0, - max_tokens = 8192, + max_tokens = 10240, }, ---@type AvanteSupportedProvider cohere = { @@ -253,7 +253,7 @@ M._defaults = { model = "command-r-plus-08-2024", timeout = 30000, -- Timeout in milliseconds temperature = 0, - max_tokens = 8192, + max_tokens = 10240, }, ---@type AvanteSupportedProvider ollama = { @@ -261,7 +261,7 @@ M._defaults = { timeout = 30000, -- Timeout in milliseconds options = { temperature = 0, - num_ctx = 8192, + num_ctx = 10240, }, }, ---@type AvanteSupportedProvider @@ -270,7 +270,7 @@ M._defaults = { model = "claude-3-5-sonnet-v2@20241022", timeout = 30000, -- Timeout in milliseconds temperature = 0, - max_tokens = 8192, + max_tokens = 10240, }, ---To add support for custom provider, follow the format below ---See https://github.com/yetone/avante.nvim/wiki#custom-providers for more details @@ -282,7 +282,7 @@ M._defaults = { model = "claude-3-5-haiku-20241022", timeout = 30000, -- Timeout in milliseconds temperature = 0, - max_tokens = 8192, + max_tokens = 10240, }, ---@type AvanteSupportedProvider ["claude-opus"] = { @@ -290,7 +290,7 @@ M._defaults = { model = "claude-3-opus-20240229", timeout = 30000, -- Timeout in milliseconds temperature = 0, - max_tokens = 8192, + max_tokens = 10240, }, ["openai-gpt-4o-mini"] = { __inherited_from = "openai",