updating config on codetyper
This commit is contained in:
@@ -5,6 +5,12 @@ return {
|
|||||||
name = "codetyper.nvim",
|
name = "codetyper.nvim",
|
||||||
lazy = false, -- Load on startup to create .coder folder
|
lazy = false, -- Load on startup to create .coder folder
|
||||||
priority = 100, -- Load early
|
priority = 100, -- Load early
|
||||||
|
dependencies = {
|
||||||
|
"nvim-lua/plenary.nvim", -- Required: async utilities
|
||||||
|
"nvim-treesitter/nvim-treesitter", -- Required: scope detection via Tree-sitter
|
||||||
|
"nvim-treesitter/nvim-treesitter-textobjects", -- Optional: better text object support
|
||||||
|
"MunifTanjim/nui.nvim", -- Optional: UI components
|
||||||
|
},
|
||||||
event = {
|
event = {
|
||||||
"BufReadPre *.coder.*",
|
"BufReadPre *.coder.*",
|
||||||
"BufNewFile *.coder.*",
|
"BufNewFile *.coder.*",
|
||||||
@@ -43,11 +49,45 @@ return {
|
|||||||
config = function()
|
config = function()
|
||||||
require("codetyper").setup({
|
require("codetyper").setup({
|
||||||
llm = {
|
llm = {
|
||||||
|
-- Available providers: "ollama", "claude", "openai", "gemini", "copilot"
|
||||||
provider = "ollama",
|
provider = "ollama",
|
||||||
|
|
||||||
|
-- Ollama (local LLM)
|
||||||
ollama = {
|
ollama = {
|
||||||
host = "http://localhost:11434",
|
host = "http://localhost:11434",
|
||||||
model = "deepseek-coder:6.7b",
|
model = "deepseek-coder:6.7b",
|
||||||
|
-- model = "codellama:7b",
|
||||||
|
-- model = "qwen2.5-coder:7b",
|
||||||
},
|
},
|
||||||
|
|
||||||
|
-- Claude (Anthropic) - https://console.anthropic.com/
|
||||||
|
-- claude = {
|
||||||
|
-- api_key = vim.env.ANTHROPIC_API_KEY, -- or hardcode: "sk-ant-..."
|
||||||
|
-- model = "claude-sonnet-4-20250514",
|
||||||
|
-- -- model = "claude-3-5-sonnet-20241022",
|
||||||
|
-- -- model = "claude-3-haiku-20240307",
|
||||||
|
-- },
|
||||||
|
|
||||||
|
-- OpenAI - https://platform.openai.com/api-keys
|
||||||
|
-- openai = {
|
||||||
|
-- api_key = vim.env.OPENAI_API_KEY, -- or hardcode: "sk-..."
|
||||||
|
-- model = "gpt-4o",
|
||||||
|
-- -- model = "gpt-4o-mini",
|
||||||
|
-- -- model = "gpt-4-turbo",
|
||||||
|
-- },
|
||||||
|
|
||||||
|
-- Google Gemini - https://aistudio.google.com/apikey
|
||||||
|
-- gemini = {
|
||||||
|
-- api_key = vim.env.GEMINI_API_KEY, -- or hardcode: "AI..."
|
||||||
|
-- model = "gemini-1.5-pro",
|
||||||
|
-- -- model = "gemini-1.5-flash",
|
||||||
|
-- -- model = "gemini-2.0-flash-exp",
|
||||||
|
-- },
|
||||||
|
|
||||||
|
-- GitHub Copilot (requires copilot.vim or copilot.lua plugin)
|
||||||
|
-- copilot = {
|
||||||
|
-- enabled = true,
|
||||||
|
-- },
|
||||||
},
|
},
|
||||||
window = {
|
window = {
|
||||||
width = 0.25, -- 1/4 of window
|
width = 0.25, -- 1/4 of window
|
||||||
@@ -62,6 +102,14 @@ return {
|
|||||||
auto_gitignore = true,
|
auto_gitignore = true,
|
||||||
auto_open_ask = false, -- Always open Ask panel on startup
|
auto_open_ask = false, -- Always open Ask panel on startup
|
||||||
agent_mode = false,
|
agent_mode = false,
|
||||||
|
scheduler = {
|
||||||
|
enabled = true,
|
||||||
|
ollama_scout = true, -- Use ollama first, escalate if low confidence
|
||||||
|
escalation_threshold = 0.7,
|
||||||
|
max_concurrent = 2,
|
||||||
|
completion_delay_ms = 100, -- Delay before checking completion visibility
|
||||||
|
apply_delay_ms = 2000, -- Wait 2 seconds before applying code
|
||||||
|
},
|
||||||
})
|
})
|
||||||
end,
|
end,
|
||||||
}
|
}
|
||||||
|
|||||||
Reference in New Issue
Block a user