chore(llm): cut-back support (#159)

Signed-off-by: Aaron Pham <contact@aarnphm.xyz>
This commit is contained in:
Aaron Pham
2024-08-23 00:21:30 -04:00
committed by GitHub
parent 6475407d0d
commit c6d5073945
5 changed files with 10 additions and 117 deletions

View File

@@ -7,9 +7,9 @@ local M = {}
---@class avante.Config
M.defaults = {
debug = false,
---Currently, default supported providers include "claude", "openai", "azure", "deepseek", "groq", "gemini"
---Currently, default supported providers include "claude", "openai", "azure", "gemini"
---For custom provider, see README.md
---@alias Provider "openai" | "claude" | "azure" | "deepseek" | "groq" | "copilot" | "gemini" | string
---@alias Provider "openai" | "claude" | "azure" | "copilot" | "gemini" | string
provider = "claude",
---@type AvanteSupportedProvider
openai = {
@@ -27,7 +27,7 @@ M.defaults = {
allow_insecure = false, -- Allow insecure server connections
timeout = 30000, -- Timeout in milliseconds
temperature = 0,
max_tokens = 8192,
max_tokens = 4096,
},
---@type AvanteAzureProvider
azure = {
@@ -46,22 +46,6 @@ M.defaults = {
temperature = 0,
max_tokens = 4096,
},
---@type AvanteSupportedProvider
deepseek = {
endpoint = "https://api.deepseek.com",
model = "deepseek-coder",
temperature = 0,
max_tokens = 4096,
["local"] = false,
},
---@type AvanteSupportedProvider
groq = {
endpoint = "https://api.groq.com",
model = "llama-3.1-70b-versatile",
temperature = 0,
max_tokens = 4096,
["local"] = false,
},
---@type AvanteGeminiProvider
gemini = {
endpoint = "https://generativelanguage.googleapis.com/v1beta/models",