fix: increase max_tokens (#1607)

This commit is contained in:
yetone
2025-03-17 03:29:44 +08:00
committed by GitHub
parent dfdab1d41f
commit 79bf76ed72

View File

@@ -195,7 +195,7 @@ M._defaults = {
model = "gpt-4o", model = "gpt-4o",
timeout = 30000, -- Timeout in milliseconds timeout = 30000, -- Timeout in milliseconds
temperature = 0, temperature = 0,
max_tokens = 8192, max_tokens = 10240,
}, },
---@type AvanteSupportedProvider ---@type AvanteSupportedProvider
copilot = { copilot = {
@@ -205,7 +205,7 @@ M._defaults = {
allow_insecure = false, -- Allow insecure server connections allow_insecure = false, -- Allow insecure server connections
timeout = 30000, -- Timeout in milliseconds timeout = 30000, -- Timeout in milliseconds
temperature = 0, temperature = 0,
max_tokens = 8192, max_tokens = 10240,
}, },
---@type AvanteAzureProvider ---@type AvanteAzureProvider
azure = { azure = {
@@ -214,7 +214,7 @@ M._defaults = {
api_version = "2024-06-01", api_version = "2024-06-01",
timeout = 30000, -- Timeout in milliseconds timeout = 30000, -- Timeout in milliseconds
temperature = 0, temperature = 0,
max_tokens = 8192, max_tokens = 10240,
}, },
---@type AvanteSupportedProvider ---@type AvanteSupportedProvider
claude = { claude = {
@@ -222,14 +222,14 @@ M._defaults = {
model = "claude-3-7-sonnet-20250219", model = "claude-3-7-sonnet-20250219",
timeout = 30000, -- Timeout in milliseconds timeout = 30000, -- Timeout in milliseconds
temperature = 0, temperature = 0,
max_tokens = 8192, max_tokens = 10240,
}, },
---@type AvanteSupportedProvider ---@type AvanteSupportedProvider
bedrock = { bedrock = {
model = "anthropic.claude-3-5-sonnet-20241022-v2:0", model = "anthropic.claude-3-5-sonnet-20241022-v2:0",
timeout = 30000, -- Timeout in milliseconds timeout = 30000, -- Timeout in milliseconds
temperature = 0, temperature = 0,
max_tokens = 8192, max_tokens = 10240,
}, },
---@type AvanteSupportedProvider ---@type AvanteSupportedProvider
gemini = { gemini = {
@@ -237,7 +237,7 @@ M._defaults = {
model = "gemini-1.5-flash-latest", model = "gemini-1.5-flash-latest",
timeout = 30000, -- Timeout in milliseconds timeout = 30000, -- Timeout in milliseconds
temperature = 0, temperature = 0,
max_tokens = 8192, max_tokens = 10240,
}, },
---@type AvanteSupportedProvider ---@type AvanteSupportedProvider
vertex = { vertex = {
@@ -245,7 +245,7 @@ M._defaults = {
model = "gemini-1.5-flash-002", model = "gemini-1.5-flash-002",
timeout = 30000, -- Timeout in milliseconds timeout = 30000, -- Timeout in milliseconds
temperature = 0, temperature = 0,
max_tokens = 8192, max_tokens = 10240,
}, },
---@type AvanteSupportedProvider ---@type AvanteSupportedProvider
cohere = { cohere = {
@@ -253,7 +253,7 @@ M._defaults = {
model = "command-r-plus-08-2024", model = "command-r-plus-08-2024",
timeout = 30000, -- Timeout in milliseconds timeout = 30000, -- Timeout in milliseconds
temperature = 0, temperature = 0,
max_tokens = 8192, max_tokens = 10240,
}, },
---@type AvanteSupportedProvider ---@type AvanteSupportedProvider
ollama = { ollama = {
@@ -261,7 +261,7 @@ M._defaults = {
timeout = 30000, -- Timeout in milliseconds timeout = 30000, -- Timeout in milliseconds
options = { options = {
temperature = 0, temperature = 0,
num_ctx = 8192, num_ctx = 10240,
}, },
}, },
---@type AvanteSupportedProvider ---@type AvanteSupportedProvider
@@ -270,7 +270,7 @@ M._defaults = {
model = "claude-3-5-sonnet-v2@20241022", model = "claude-3-5-sonnet-v2@20241022",
timeout = 30000, -- Timeout in milliseconds timeout = 30000, -- Timeout in milliseconds
temperature = 0, temperature = 0,
max_tokens = 8192, max_tokens = 10240,
}, },
---To add support for custom provider, follow the format below ---To add support for custom provider, follow the format below
---See https://github.com/yetone/avante.nvim/wiki#custom-providers for more details ---See https://github.com/yetone/avante.nvim/wiki#custom-providers for more details
@@ -282,7 +282,7 @@ M._defaults = {
model = "claude-3-5-haiku-20241022", model = "claude-3-5-haiku-20241022",
timeout = 30000, -- Timeout in milliseconds timeout = 30000, -- Timeout in milliseconds
temperature = 0, temperature = 0,
max_tokens = 8192, max_tokens = 10240,
}, },
---@type AvanteSupportedProvider ---@type AvanteSupportedProvider
["claude-opus"] = { ["claude-opus"] = {
@@ -290,7 +290,7 @@ M._defaults = {
model = "claude-3-opus-20240229", model = "claude-3-opus-20240229",
timeout = 30000, -- Timeout in milliseconds timeout = 30000, -- Timeout in milliseconds
temperature = 0, temperature = 0,
max_tokens = 8192, max_tokens = 10240,
}, },
["openai-gpt-4o-mini"] = { ["openai-gpt-4o-mini"] = {
__inherited_from = "openai", __inherited_from = "openai",