soulfy
f65e72900f
fix(provider.openai) reasoning_content maybe empty string ( #2000 )
...
Co-authored-by: yetone <yetoneful@gmail.com >
Co-authored-by: pre-commit-ci-lite[bot] <117423508+pre-commit-ci-lite[bot]@users.noreply.github.com>
2025-05-08 12:42:14 +08:00
yetone
3ab2a5d68e
fix: check on_chunk if it is a nil ( #1998 )
2025-05-07 01:02:11 +08:00
yetone
60897ee9a6
fix: o1 model ( #1992 )
2025-05-06 19:32:55 +08:00
yetone
8f23cc85ab
fix: gemini openai compatible api ( #1976 )
2025-05-03 14:09:46 +08:00
yetone
d3094d6c9f
fix: help cmd ( #1963 )
2025-05-02 00:53:55 +08:00
yetone
f2330a0701
fix: dispatch agent ( #1953 )
2025-05-01 05:12:23 +08:00
Daniel
c2ec1a8e49
fix(openai): wrap single text block in array before appending images ( #1948 )
2025-05-01 03:33:53 +08:00
yetone
99a2e35ff3
fix: copilot gemini function calling ( #1947 )
2025-05-01 00:09:39 +08:00
yetone
f10b8383e3
refactor: history messages ( #1934 )
2025-04-30 03:07:18 +08:00
yetone
33f95aedde
feat: streaming json parser ( #1883 )
2025-04-16 23:41:38 +08:00
yetone
04336913b3
Revert "fix max_tokens for reasoning models ( #1819 )" ( #1839 )
...
* Revert "fix max_tokens for reasoning models (#1819 )"
This reverts commit 1e2e233ff5 .
* Revert "fix: revert max_completion_tokens to max_tokens (#1741 )"
This reverts commit cd13eeb7d9 .
* fix: nvim_version
2025-04-09 16:58:54 +08:00
PSeitz
1e2e233ff5
fix max_tokens for reasoning models ( #1819 )
2025-04-07 13:55:40 +08:00
yetone
cd13eeb7d9
fix: revert max_completion_tokens to max_tokens ( #1741 )
2025-03-27 16:53:55 +08:00
yetone
25e9b1c111
fix: copilot's openai tool description cannot use long text ( #1700 )
2025-03-24 19:54:06 +08:00
yetone
49ae3c84fd
feat: add view_range parameter and remove read_file llm tool ( #1690 )
2025-03-24 15:07:14 +08:00
kernitus
7d137d32cb
fix: do not allow max_tokens & max_completion_tokens at the same time ( #1633 )
...
* fix: do not allow max_tokens & max_completion_tokens at the same time
* chore: fix formatting
2025-03-19 01:44:41 +08:00
kernitus
10ce065d9e
feat: update openai/azure params ( #1604 )
...
* feat(openai): use max_completion_tokens & reasoning_effort params
* feat(openai): use developer prompt for reasoning models
* docs: update openai config in readme
* refactor: follow lua style quotes
* fix(azure): rename max_tokens to max_completion_tokens
* refactor(azure): remove duplicate field
* refactor: update types
* refactor(azure): update type
2025-03-18 19:40:20 +08:00
Florian Engelhardt
3b86549485
Add extra headers for OpenAI ( #1615 )
2025-03-17 17:51:08 +08:00
brook hong
e204ca1cf3
feat: support tools in bedrock ( #1598 )
2025-03-17 13:42:18 +08:00
yetone
eb5a2f6cca
fix: missing tool use ( #1610 )
2025-03-17 05:00:03 +08:00
yetone
dfdab1d41f
fix: openai tool histories ( #1606 )
2025-03-17 03:06:26 +08:00
yetone
6e77da83c1
fix: better sidebar ( #1603 )
...
* fix: better sidebar
* feat: better msg history
* fix: tests
2025-03-17 01:40:05 +08:00
yetone
f9f92dc9d4
Revert "fix: Always handle the extra response outside of stream, such as the exceptions from bedrock ( #1526 )" ( #1569 )
...
This reverts commit f9ab6934d2 .
2025-03-12 19:10:05 +08:00
brook hong
f9ab6934d2
fix: Always handle the extra response outside of stream, such as the exceptions from bedrock ( #1526 )
2025-03-12 16:29:32 +08:00
yetone
aa2c2b0054
fix: stricter static typing ( #1550 )
2025-03-11 00:41:21 +08:00
yetone
750ee80971
feat: add ollama as supported provider ( #1543 )
...
* feat: add ollama as supported provider
*This implementation is only working with `stream = true`*
- Uses the actual ollama api and allows for passing additional options
- Properly passes the system prompt to api
Use ollama as provider in opts like this:
opts = {
debug = true,
provider = "ollama",
ollama = {
api_key_name = "",
endpoint = "http://127.0.0.1:11434 ",
model = "qwen2.5-coder:latest",
options = {
num_ctx = 32768,
temperature = 0,
},
stream = true,
},
* fix: ollama types
---------
Co-authored-by: jtabke <25010496+jtabke@users.noreply.github.com >
2025-03-10 02:23:56 +08:00
yetone
8620ea3e12
refactor: summarize memory ( #1508 )
2025-03-07 00:12:57 +08:00
yetone
5aa55689ff
fix: bedrock claude do not support prompt caching ( #1507 )
2025-03-06 18:31:56 +08:00
yetone
0d592f440c
Revert "refactor: message content ( #1424 )" ( #1442 )
...
This reverts commit ae8497faf1 .
2025-03-01 13:25:51 +08:00
yetone
ae8497faf1
refactor: message content ( #1424 )
2025-02-28 14:43:39 +08:00
Peter Cardenas
afa674c6fd
chore: prefer not to use function assignment ( #1381 )
2025-02-25 12:08:03 +08:00
yetone
b04bffa441
fix: types ( #1344 )
2025-02-22 23:24:20 +08:00
zerogog
a17d4b00e2
fix : #1176 Gemini model in OpenRouter not finish their response ( #1292 )
2025-02-18 22:59:16 +08:00
8uff3r
f70eb1040c
fix(providers/openai): check for vim.NIl on tool_calls ( #1283 )
2025-02-17 12:54:58 +08:00
yetone
ce55d7ac9e
refactor: better value name ( #1261 )
2025-02-13 01:39:02 +08:00
yetone
5f6612fb8c
fix: support multiple tool use ( #1192 )
2025-02-06 16:41:28 +08:00
yetone
d1cc23fa54
fix: multiple tool use histories and disable tools ( #1185 )
2025-02-06 02:46:52 +08:00
yetone
3aaf7dad77
feat: tools support copilot ( #1183 )
2025-02-05 23:47:52 +08:00
yetone
1437f319d2
feat: tools ( #1180 )
...
* feat: tools
* feat: claude use tools
* feat: openai use tools
2025-02-05 22:39:54 +08:00
yetone
ef4b6077ec
feat: supports openrouter reasoning ( #1174 )
2025-02-04 01:38:18 +08:00
yetone
cd7390de21
fix: remove unnecessary think tag ( #1173 )
2025-02-03 22:32:41 +08:00
yetone
8536d102be
fix: reasoning content processing ( #1171 )
2025-02-03 21:55:12 +08:00
yetone
5ac934f228
chores: remove debug log ( #1160 )
2025-02-02 02:29:41 +08:00
yetone
b5ac768416
feat: supports reasoning_content ( #1159 )
2025-02-02 02:12:14 +08:00
yetone
3ec847e3cb
fix(ci): lua lint ( #1035 )
2025-01-05 17:11:15 +08:00
Larry Lv
ec5d1abf34
fix(openai): support all o series models ( #1031 )
...
Before this change, since `max_completion_tokens` was not set for `o` series models, the completion request will time out sometimes. This makes sure it converts the `max_tokens` parameter to `max_completion_tokens` for `o` series models.
I tested this change with `gpt-4o-mini`, `o1-mini` and `o3-mini`, and they all still work as expected.
2025-01-05 13:23:33 +08:00
yetone
01e05a538b
fix: more reasonable error reporting ( #965 )
2024-12-18 23:16:41 +08:00
Aaron Batilo
e60ccd2db4
feat: enable streaming for o1 models ( #896 )
...
As of a few days ago, o1 models support streaming responses. Please see:
https://community.openai.com/t/openai-o1-streaming-now-available-api-access-for-tiers-1-5/1025430
2024-11-24 17:28:27 +08:00
yetone
9891b03656
fix(openai): user and assistant roles should be alternating ( #859 )
2024-11-17 03:49:02 +08:00
yetone
ff85b9c1e2
refactor: remove redundant local field to facilitate provider configuration ( #858 )
2024-11-17 02:55:40 +08:00