mirror of
https://github.com/NotAShelf/nvf.git
synced 2025-09-06 10:21:31 +00:00
avante-nvim: Migrate provider options
Migrate the provider options according to the migration guide. https://github.com/yetone/avante.nvim/wiki/Provider-configuration-migration-guide
This commit is contained in:
parent
77a32f0961
commit
6ad7500a06
1 changed files with 34 additions and 18 deletions
|
@ -13,27 +13,43 @@ in {
|
||||||
description = "The provider used in Aider mode or in the planning phase of Cursor Planning Mode.";
|
description = "The provider used in Aider mode or in the planning phase of Cursor Planning Mode.";
|
||||||
};
|
};
|
||||||
|
|
||||||
vendors = mkOption {
|
providers = mkOption {
|
||||||
type = nullOr attrs;
|
type = nullOr attrs;
|
||||||
default = null;
|
default = null;
|
||||||
description = "Define Your Custom providers.";
|
description = "Define settings for builtin and custom providers.";
|
||||||
example = literalMD ''
|
example = literalMD ''
|
||||||
```nix
|
```nix
|
||||||
ollama = {
|
openai = {
|
||||||
__inherited_from = "openai";
|
endpoint = "https://api.openai.com/v1";
|
||||||
api_key_name = "";
|
model = "gpt-4o"; # your desired model (or use gpt-4o, etc.)
|
||||||
endpoint = "http://127.0.0.1:11434/v1";
|
timeout = 30000; # Timeout in milliseconds, increase this for reasoning models
|
||||||
model = "qwen2.5u-coder:7b";
|
extra_request_body = {
|
||||||
max_tokens = 4096;
|
temperature = 0;
|
||||||
disable_tools = true;
|
max_completion_tokens = 8192; # Increase this to include reasoning tokens (for reasoning models)
|
||||||
|
reasoning_effort = "medium"; # low|medium|high, only used for reasoning models
|
||||||
};
|
};
|
||||||
ollama_ds = {
|
};
|
||||||
|
ollama = {
|
||||||
|
endpoint = "http://127.0.0.1:11434";
|
||||||
|
timeout = 30000; # Timeout in milliseconds
|
||||||
|
extra_request_body = {
|
||||||
|
options = {
|
||||||
|
temperature = 0.75;
|
||||||
|
num_ctx = 20480;
|
||||||
|
keep_alive = "5m";
|
||||||
|
};
|
||||||
|
};
|
||||||
|
};
|
||||||
|
groq = {
|
||||||
__inherited_from = "openai";
|
__inherited_from = "openai";
|
||||||
api_key_name = "";
|
api_key_name = "GROQ_API_KEY";
|
||||||
endpoint = "http://127.0.0.1:11434/v1";
|
endpoint = "https://api.groq.com/openai/v1/";
|
||||||
model = "deepseek-r1u:7b";
|
model = "llama-3.3-70b-versatile";
|
||||||
max_tokens = 4096;
|
|
||||||
disable_tools = true;
|
disable_tools = true;
|
||||||
|
extra_request_body = {
|
||||||
|
temperature = 1;
|
||||||
|
max_tokens = 32768; # remember to increase this value, otherwise it will stop generating halfway
|
||||||
|
};
|
||||||
};
|
};
|
||||||
```
|
```
|
||||||
'';
|
'';
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue