fix: revert max_completion_tokens to max_tokens (#1741)
This commit is contained in:
@@ -196,7 +196,7 @@ M._defaults = {
|
||||
model = "gpt-4o",
|
||||
timeout = 30000, -- Timeout in milliseconds, increase this for reasoning models
|
||||
temperature = 0,
|
||||
max_completion_tokens = 16384, -- Increase this to include reasoning tokens (for reasoning models)
|
||||
max_tokens = 16384, -- Increase this to include reasoning tokens (for reasoning models)
|
||||
reasoning_effort = "medium", -- low|medium|high, only used for reasoning models
|
||||
},
|
||||
---@type AvanteSupportedProvider
|
||||
@@ -216,7 +216,7 @@ M._defaults = {
|
||||
api_version = "2024-12-01-preview",
|
||||
timeout = 30000, -- Timeout in milliseconds, increase this for reasoning models
|
||||
temperature = 0,
|
||||
max_completion_tokens = 20480, -- Increase this to include reasoning tokens (for reasoning models)
|
||||
max_tokens = 20480, -- Increase this to include reasoning tokens (for reasoning models)
|
||||
reasoning_effort = "medium", -- low|medium|high, only used for reasoning models
|
||||
},
|
||||
---@type AvanteSupportedProvider
|
||||
|
||||
Reference in New Issue
Block a user