fix: update gemini max_tokens and use 2.0 model (#1767)
docs ref for max output tokens: https://ai.google.dev/gemini-api/docs/models#gemini-2.0-flash Token limits[*] Input token limit 1,048,576 Output token limit 8,192
This commit is contained in:
@@ -237,10 +237,10 @@ M._defaults = {
|
|||||||
---@type AvanteSupportedProvider
|
---@type AvanteSupportedProvider
|
||||||
gemini = {
|
gemini = {
|
||||||
endpoint = "https://generativelanguage.googleapis.com/v1beta/models",
|
endpoint = "https://generativelanguage.googleapis.com/v1beta/models",
|
||||||
model = "gemini-1.5-flash-latest",
|
model = "gemini-2.0-flash",
|
||||||
timeout = 30000, -- Timeout in milliseconds
|
timeout = 30000, -- Timeout in milliseconds
|
||||||
temperature = 0,
|
temperature = 0,
|
||||||
max_tokens = 20480,
|
max_tokens = 8192,
|
||||||
},
|
},
|
||||||
---@type AvanteSupportedProvider
|
---@type AvanteSupportedProvider
|
||||||
vertex = {
|
vertex = {
|
||||||
|
|||||||
Reference in New Issue
Block a user