Compare commits
10 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
| 3baa3102a3 | |||
|
|
2d4fad596c | ||
| 7259e59d2a | |||
| cec04c4597 | |||
|
|
a7f5677195 | ||
|
|
6075f0a190 | ||
| 15310a9e2c | |||
| f7df54f2f7 | |||
| 212d4bace4 | |||
| f4b3267c89 |
@@ -1,3 +1,9 @@
|
||||
## v0.1.3 (2025-12-13)
|
||||
|
||||
### Feat
|
||||
|
||||
- Improved MCP implementation to minimize the tokens needed to utilize it so it doesn't quickly overwhelm the token space for a given model
|
||||
|
||||
## v0.1.2 (2025-11-08)
|
||||
|
||||
### Refactor
|
||||
|
||||
Generated
+364
-279
File diff suppressed because it is too large
Load Diff
+1
-1
@@ -1,6 +1,6 @@
|
||||
[package]
|
||||
name = "loki-ai"
|
||||
version = "0.1.2"
|
||||
version = "0.1.3"
|
||||
edition = "2024"
|
||||
authors = ["Alex Clarke <alex.j.tusa@gmail.com>"]
|
||||
description = "An all-in-one, batteries included LLM CLI Tool"
|
||||
|
||||
@@ -84,7 +84,7 @@ You can also customize the location of full agent configurations using the follo
|
||||
|
||||
| Environment Variable | Description |
|
||||
|------------------------------|-------------------------------------------------------------------------------------------------------------------------------------|
|
||||
| `<AGENT_NAME>_CONFIG_FILE | Customize the location of the agent's configuration file; e.g. `SQL_CONFIG_FILE` |
|
||||
| `<AGENT_NAME>_CONFIG_FILE` | Customize the location of the agent's configuration file; e.g. `SQL_CONFIG_FILE` |
|
||||
| `<AGENT_NAME>_MODEL` | Customize the `model` used for the agent; e.g `SQL_MODEL` |
|
||||
| `<AGENT_NAME>_TEMPERATURE` | Customize the `temperature` used for the agent; e.g. `SQL_TEMPERATURE` |
|
||||
| `<AGENT_NAME>_TOP_P` | Customize the `top_p` used for the agent; e.g. `SQL_TOP_P` |
|
||||
|
||||
+1
-1
@@ -114,7 +114,7 @@ At the time of writing, the following files support Loki secret injection:
|
||||
|-------------------------|-----------------------------------|-------------------------------------------------------------------------------------------------------------------------------------------|
|
||||
| `config.yaml` | The main Loki configuration file | Cannot use secret injection on the `vault_password_file` field |
|
||||
| `functions/mcp.json` | The MCP server configuration file | |
|
||||
| `<agent>/tools.<py/sh>` | Tool files for agents | Specific configuration and only supported for Agents, not all global tools ([see below](#environment-variable-secret-injection-in-agents) |
|
||||
| `<agent>/tools.<py/sh>` | Tool files for agents | Specific configuration and only supported for Agents, not all global tools ([see below](#environment-variable-secret-injection-in-agents)) |
|
||||
|
||||
|
||||
Note that all paths are relative to the Loki configuration directory. The directory varies by system, so you can find yours by
|
||||
|
||||
@@ -83,9 +83,7 @@ enabled_mcp_servers: null # Which MCP servers to enable by default (e.g.
|
||||
```
|
||||
|
||||
A special note about `enabled_mcp_servers`: a user can set this to `all` to enable all configured MCP servers in the
|
||||
`functions/mcp.json` configuration. However, **this should be used with caution**. When there is a significant number
|
||||
of configured MCP servers, enabling all MCP servers may overwhelm the context length of a model, and quickly exceed
|
||||
token limits.
|
||||
`functions/mcp.json` configuration.
|
||||
|
||||
(See the [Configuration Example](../../config.example.yaml) file for an example global configuration with all options.)
|
||||
|
||||
|
||||
+210
-217
@@ -3,6 +3,20 @@
|
||||
# - https://platform.openai.com/docs/api-reference/chat
|
||||
- provider: openai
|
||||
models:
|
||||
- name: gpt-5.1
|
||||
max_input_tokens: 400000
|
||||
max_output_tokens: 128000
|
||||
input_price: 1.25
|
||||
output_price: 10
|
||||
supports_vision: true
|
||||
supports_function_calling: true
|
||||
- name: gpt-5.1-chat-latest
|
||||
max_input_tokens: 400000
|
||||
max_output_tokens: 128000
|
||||
input_price: 1.25
|
||||
output_price: 10
|
||||
supports_vision: true
|
||||
supports_function_calling: true
|
||||
- name: gpt-5
|
||||
max_input_tokens: 400000
|
||||
max_output_tokens: 128000
|
||||
@@ -31,13 +45,6 @@
|
||||
output_price: 0.4
|
||||
supports_vision: true
|
||||
supports_function_calling: true
|
||||
- name: gpt-5-codex
|
||||
max_input_tokens: 400000
|
||||
max_output_tokens: 128000
|
||||
input_price: 1.25
|
||||
output_price: 10
|
||||
supports_vision: true
|
||||
supports_function_calling: true
|
||||
- name: gpt-4.1
|
||||
max_input_tokens: 1047576
|
||||
max_output_tokens: 32768
|
||||
@@ -259,6 +266,30 @@
|
||||
thinking:
|
||||
type: enabled
|
||||
budget_tokens: 16000
|
||||
- name: claude-haiku-4-5-20251001
|
||||
max_input_tokens: 200000
|
||||
max_output_tokens: 8192
|
||||
require_max_tokens: true
|
||||
input_price: 1
|
||||
output_price: 5
|
||||
supports_vision: true
|
||||
supports_function_calling: true
|
||||
- name: claude-haiku-4-5-20251001:thinking
|
||||
real_name: claude-haiku-4-5-20251001
|
||||
max_input_tokens: 200000
|
||||
max_output_tokens: 24000
|
||||
require_max_tokens: true
|
||||
input_price: 1
|
||||
output_price: 5
|
||||
supports_vision: true
|
||||
supports_function_calling: true
|
||||
patch:
|
||||
body:
|
||||
temperature: null
|
||||
top_p: null
|
||||
thinking:
|
||||
type: enabled
|
||||
budget_tokens: 16000
|
||||
- name: claude-opus-4-1-20250805
|
||||
max_input_tokens: 200000
|
||||
max_output_tokens: 8192
|
||||
@@ -660,6 +691,29 @@
|
||||
thinking:
|
||||
type: enabled
|
||||
budget_tokens: 16000
|
||||
- name: claude-haiku-4-5@20251001
|
||||
max_input_tokens: 200000
|
||||
max_output_tokens: 8192
|
||||
require_max_tokens: true
|
||||
input_price: 1
|
||||
output_price: 5
|
||||
supports_vision: true
|
||||
supports_function_calling: true
|
||||
- name: claude-haiku-4-5@20251001:thinking
|
||||
real_name: claude-haiku-4-5@20251001
|
||||
max_input_tokens: 200000
|
||||
max_output_tokens: 24000
|
||||
require_max_tokens: true
|
||||
input_price: 1
|
||||
output_price: 5
|
||||
supports_vision: true
|
||||
patch:
|
||||
body:
|
||||
temperature: null
|
||||
top_p: null
|
||||
thinking:
|
||||
type: enabled
|
||||
budget_tokens: 16000
|
||||
- name: claude-opus-4-1@20250805
|
||||
max_input_tokens: 200000
|
||||
max_output_tokens: 8192
|
||||
@@ -817,6 +871,31 @@
|
||||
thinking:
|
||||
type: enabled
|
||||
budget_tokens: 16000
|
||||
- name: us.anthropic.claude-haiku-4-5-20251001-v1:0
|
||||
max_input_tokens: 200000
|
||||
max_output_tokens: 8192
|
||||
require_max_tokens: true
|
||||
input_price: 1
|
||||
output_price: 5
|
||||
supports_vision: true
|
||||
supports_function_calling: true
|
||||
- name: us.anthropic.claude-haiku-4-5-20251001-v1:0:thinking
|
||||
real_name: us.anthropic.claude-haiku-4-5-20251001-v1:0
|
||||
max_input_tokens: 200000
|
||||
max_output_tokens: 24000
|
||||
require_max_tokens: true
|
||||
input_price: 1
|
||||
output_price: 5
|
||||
supports_vision: true
|
||||
patch:
|
||||
body:
|
||||
inferenceConfig:
|
||||
temperature: null
|
||||
topP: null
|
||||
additionalModelRequestFields:
|
||||
thinking:
|
||||
type: enabled
|
||||
budget_tokens: 16000
|
||||
- name: us.anthropic.claude-opus-4-1-20250805-v1:0
|
||||
max_input_tokens: 200000
|
||||
max_output_tokens: 8192
|
||||
@@ -1004,6 +1083,12 @@
|
||||
require_max_tokens: true
|
||||
input_price: 0
|
||||
output_price: 0
|
||||
- name: '@cf/qwen/qwen3-30b-a3b-fp8'
|
||||
max_input_tokens: 131072
|
||||
max_output_tokens: 2048
|
||||
require_max_tokens: true
|
||||
input_price: 0
|
||||
output_price: 0
|
||||
- name: '@cf/qwen/qwen2.5-coder-32b-instruct'
|
||||
max_input_tokens: 131072
|
||||
max_output_tokens: 2048
|
||||
@@ -1030,8 +1115,8 @@
|
||||
max_batch_size: 100
|
||||
|
||||
# Links:
|
||||
# - https://cloud.baidu.com/doc/WENXINWORKSHOP/s/Wm9cvy6rl
|
||||
# - https://cloud.baidu.com/doc/WENXINWORKSHOP/s/Qm9cw2s7m
|
||||
# - https://cloud.baidu.com/doc/qianfan/s/rmh4stp0j
|
||||
# - https://cloud.baidu.com/doc/qianfan/s/wmh4sv6ya
|
||||
- provider: ernie
|
||||
models:
|
||||
- name: ernie-4.5-turbo-128k
|
||||
@@ -1043,8 +1128,12 @@
|
||||
input_price: 0.42
|
||||
output_price: 1.26
|
||||
supports_vision: true
|
||||
- name: ernie-x1-turbo-32k
|
||||
max_input_tokens: 32768
|
||||
- name: ernie-5.0-thinking-preview
|
||||
max_input_tokens: 131072
|
||||
input_price: 1.4
|
||||
output_price: 5.6
|
||||
- name: ernie-x1.1-preview
|
||||
max_input_tokens: 65536
|
||||
input_price: 0.14
|
||||
output_price: 0.56
|
||||
- name: bge-large-zh
|
||||
@@ -1064,75 +1153,31 @@
|
||||
max_input_tokens: 1024
|
||||
input_price: 0.07
|
||||
|
||||
|
||||
# Links:
|
||||
# - https://help.aliyun.com/zh/model-studio/getting-started/models
|
||||
# - https://help.aliyun.com/zh/model-studio/developer-reference/use-qwen-by-calling-api
|
||||
- provider: qianwen
|
||||
models:
|
||||
- name: qwen-max-latest
|
||||
max_input_tokens: 32678
|
||||
max_output_tokens: 8192
|
||||
input_price: 1.6
|
||||
output_price: 6.4
|
||||
supports_function_calling: true
|
||||
- name: qwen-plus-latest
|
||||
max_input_tokens: 131072
|
||||
max_output_tokens: 8192
|
||||
input_price: 0.112
|
||||
output_price: 0.28
|
||||
supports_function_calling: true
|
||||
- name: qwen-turbo-latest
|
||||
max_input_tokens: 1000000
|
||||
max_output_tokens: 8192
|
||||
input_price: 0.042
|
||||
output_price: 0.084
|
||||
supports_function_calling: true
|
||||
- name: qwen-long
|
||||
max_input_tokens: 1000000
|
||||
input_price: 0.07
|
||||
output_price: 0.28
|
||||
- name: qwen-omni-turbo-latest
|
||||
max_input_tokens: 32768
|
||||
max_output_tokens: 2048
|
||||
supports_vision: true
|
||||
- name: qwen-coder-plus-latest
|
||||
max_input_tokens: 131072
|
||||
max_output_tokens: 8192
|
||||
input_price: 0.49
|
||||
output_price: 0.98
|
||||
- name: qwen-coder-turbo-latest
|
||||
max_input_tokens: 131072
|
||||
max_output_tokens: 8192
|
||||
input_price: 0.28
|
||||
output_price: 0.84
|
||||
- name: qwen-vl-max-latest
|
||||
max_input_tokens: 30720
|
||||
max_output_tokens: 2048
|
||||
input_price: 0.42
|
||||
output_price: 1.26
|
||||
supports_vision: true
|
||||
- name: qwen-vl-plus-latest
|
||||
max_input_tokens: 30000
|
||||
max_output_tokens: 2048
|
||||
input_price: 0.21
|
||||
output_price: 0.63
|
||||
supports_vision: true
|
||||
- name: qwen3-max
|
||||
max_input_tokens: 262144
|
||||
input_price: 2.1
|
||||
output_price: 8.4
|
||||
supports_function_calling: true
|
||||
- name: qwen-plus
|
||||
max_input_tokens: 131072
|
||||
supports_function_calling: true
|
||||
- name: qwen-flash
|
||||
max_input_tokens: 1000000
|
||||
supports_function_calling: true
|
||||
- name: qwen3-vl-plus
|
||||
max_input_tokens: 262144
|
||||
input_price: 0.42
|
||||
output_price: 4.2
|
||||
supports_vision: true
|
||||
- name: qwen3-max-preview
|
||||
- name: qwen3-vl-flash
|
||||
max_input_tokens: 262144
|
||||
max_output_tokens: 32768
|
||||
input_price: 1.4
|
||||
output_price: 5.6
|
||||
supports_function_calling: true
|
||||
supports_vision: true
|
||||
- name: qwen-coder-plus
|
||||
max_input_tokens: 1000000
|
||||
- name: qwen-coder-flash
|
||||
max_input_tokens: 1000000
|
||||
- name: qwen3-next-80b-a3b-instruct
|
||||
max_input_tokens: 131072
|
||||
input_price: 0.14
|
||||
@@ -1160,6 +1205,16 @@
|
||||
max_input_tokens: 131072
|
||||
input_price: 0.105
|
||||
output_price: 1.05
|
||||
- name: qwen3-vl-32b-instruct
|
||||
max_input_tokens: 131072
|
||||
input_price: 0.28
|
||||
output_price: 1.12
|
||||
supports_vision: true
|
||||
- name: qwen3-vl-8b-instruct
|
||||
max_input_tokens: 131072
|
||||
input_price: 0.07
|
||||
output_price: 0.28
|
||||
supports_vision: true
|
||||
- name: qwen3-coder-480b-a35b-instruct
|
||||
max_input_tokens: 262144
|
||||
input_price: 1.26
|
||||
@@ -1168,32 +1223,10 @@
|
||||
max_input_tokens: 262144
|
||||
input_price: 0.315
|
||||
output_price: 1.26
|
||||
- name: qwen2.5-72b-instruct
|
||||
max_input_tokens: 129024
|
||||
max_output_tokens: 8192
|
||||
input_price: 0.56
|
||||
output_price: 1.68
|
||||
supports_function_calling: true
|
||||
- name: qwen2.5-vl-72b-instruct
|
||||
max_input_tokens: 129024
|
||||
max_output_tokens: 8192
|
||||
input_price: 2.24
|
||||
output_price: 6.72
|
||||
supports_vision: true
|
||||
- name: qwen2.5-coder-32b-instruct
|
||||
max_input_tokens: 129024
|
||||
max_output_tokens: 8192
|
||||
input_price: 0.49
|
||||
output_price: 0.98
|
||||
supports_function_calling: true
|
||||
- name: deepseek-v3.1
|
||||
- name: deepseek-v3.2-exp
|
||||
max_input_tokens: 131072
|
||||
input_price: 0.28
|
||||
output_price: 1.12
|
||||
- name: deepseek-r1-0528
|
||||
max_input_tokens: 65536
|
||||
input_price: 0.28
|
||||
output_price: 1.12
|
||||
output_price: 0.42
|
||||
- name: text-embedding-v4
|
||||
type: embedding
|
||||
input_price: 0.1
|
||||
@@ -1247,10 +1280,10 @@
|
||||
# - https://platform.moonshot.cn/docs/api/chat#%E5%85%AC%E5%BC%80%E7%9A%84%E6%9C%8D%E5%8A%A1%E5%9C%B0%E5%9D%80
|
||||
- provider: moonshot
|
||||
models:
|
||||
- name: kimi-latest
|
||||
max_input_tokens: 131072
|
||||
input_price: 1.4
|
||||
output_price: 4.2
|
||||
- name: kimi-k2-turbo-preview
|
||||
max_input_tokens: 262144
|
||||
input_price: 1.12
|
||||
output_price: 8.12
|
||||
supports_vision: true
|
||||
supports_function_calling: true
|
||||
- name: kimi-k2-0905-preview
|
||||
@@ -1259,16 +1292,15 @@
|
||||
output_price: 2.24
|
||||
supports_vision: true
|
||||
supports_function_calling: true
|
||||
- name: kimi-k2-turbo-preview
|
||||
max_input_tokens: 131072
|
||||
- name: kimi-k2-thinking-turbo
|
||||
max_input_tokens: 262144
|
||||
input_price: 1.12
|
||||
output_price: 4.48
|
||||
output_price: 8.12
|
||||
supports_vision: true
|
||||
supports_function_calling: true
|
||||
- name: kimi-thinking-preview
|
||||
max_input_tokens: 131072
|
||||
input_price: 28
|
||||
output_price: 28
|
||||
- name: kimi-k2-thinking
|
||||
max_input_tokens: 262144
|
||||
input_price: 0.56
|
||||
output_price: 2.24
|
||||
supports_vision: true
|
||||
|
||||
# Links:
|
||||
@@ -1293,7 +1325,7 @@
|
||||
# - https://open.bigmodel.cn/dev/api#glm-4
|
||||
- provider: zhipuai
|
||||
models:
|
||||
- name: glm-4.5
|
||||
- name: glm-4.6
|
||||
max_input_tokens: 202752
|
||||
input_price: 0.28
|
||||
output_price: 1.12
|
||||
@@ -1353,25 +1385,35 @@
|
||||
input_price: 0.112
|
||||
|
||||
# Links:
|
||||
# - https://platform.minimaxi.com/document/pricing
|
||||
# - https://platform.minimaxi.com/docs/guides/pricing
|
||||
# - https://platform.minimaxi.com/document/ChatCompletion%20v2
|
||||
- provider: minimax
|
||||
models:
|
||||
- name: minimax-text-01
|
||||
max_input_tokens: 1000192
|
||||
input_price: 0.14
|
||||
output_price: 1.12
|
||||
supports_vision: true
|
||||
- name: minimax-m1
|
||||
max_input_tokens: 131072
|
||||
input_price: 0.112
|
||||
output_price: 1.12
|
||||
- name: minimax-m2
|
||||
max_input_tokens: 204800
|
||||
input_price: 0.294
|
||||
output_price: 1.176
|
||||
supports_function_calling: true
|
||||
|
||||
# Links:
|
||||
# - https://openrouter.ai/models
|
||||
# - https://openrouter.ai/docs/api-reference/chat-completion
|
||||
- provider: openrouter
|
||||
models:
|
||||
- name: openai/gpt-5.1
|
||||
max_input_tokens: 400000
|
||||
max_output_tokens: 128000
|
||||
input_price: 1.25
|
||||
output_price: 10
|
||||
supports_vision: true
|
||||
supports_function_calling: true
|
||||
- name: openai/gpt-5.1-chat
|
||||
max_input_tokens: 400000
|
||||
max_output_tokens: 128000
|
||||
input_price: 1.25
|
||||
output_price: 10
|
||||
supports_vision: true
|
||||
supports_function_calling: true
|
||||
- name: openai/gpt-5
|
||||
max_input_tokens: 400000
|
||||
max_output_tokens: 128000
|
||||
@@ -1400,13 +1442,6 @@
|
||||
output_price: 0.4
|
||||
supports_vision: true
|
||||
supports_function_calling: true
|
||||
- name: openai/gpt-5-codex
|
||||
max_input_tokens: 400000
|
||||
max_output_tokens: 128000
|
||||
input_price: 1.25
|
||||
output_price: 10
|
||||
supports_vision: true
|
||||
supports_function_calling: true
|
||||
- name: openai/gpt-4.1
|
||||
max_input_tokens: 1047576
|
||||
max_output_tokens: 32768
|
||||
@@ -1563,6 +1598,14 @@
|
||||
output_price: 15
|
||||
supports_vision: true
|
||||
supports_function_calling: true
|
||||
- name: anthropic/claude-haiku-4.5
|
||||
max_input_tokens: 200000
|
||||
max_output_tokens: 8192
|
||||
require_max_tokens: true
|
||||
input_price: 1
|
||||
output_price: 5
|
||||
supports_vision: true
|
||||
supports_function_calling: true
|
||||
- name: anthropic/claude-opus-4.1
|
||||
max_input_tokens: 200000
|
||||
max_output_tokens: 8192
|
||||
@@ -1696,11 +1739,10 @@
|
||||
patch:
|
||||
body:
|
||||
include_reasoning: true
|
||||
- name: qwen/qwen-max
|
||||
max_input_tokens: 32768
|
||||
max_output_tokens: 8192
|
||||
input_price: 1.6
|
||||
output_price: 6.4
|
||||
- name: qwen/qwen3-max
|
||||
max_input_tokens: 262144
|
||||
input_price: 1.2
|
||||
output_price: 6
|
||||
supports_function_calling: true
|
||||
- name: qwen/qwen-plus
|
||||
max_input_tokens: 131072
|
||||
@@ -1708,22 +1750,6 @@
|
||||
input_price: 0.4
|
||||
output_price: 1.2
|
||||
supports_function_calling: true
|
||||
- name: qwen/qwen-turbo
|
||||
max_input_tokens: 1000000
|
||||
max_output_tokens: 8192
|
||||
input_price: 0.05
|
||||
output_price: 0.2
|
||||
supports_function_calling: true
|
||||
- name: qwen/qwen-vl-plus
|
||||
max_input_tokens: 7500
|
||||
input_price: 0.21
|
||||
output_price: 0.63
|
||||
supports_vision: true
|
||||
- name: qwen/qwen3-max
|
||||
max_input_tokens: 262144
|
||||
input_price: 1.2
|
||||
output_price: 6
|
||||
supports_function_calling: true
|
||||
- name: qwen/qwen3-next-80b-a3b-instruct
|
||||
max_input_tokens: 262144
|
||||
input_price: 0.1
|
||||
@@ -1733,7 +1759,7 @@
|
||||
max_input_tokens: 262144
|
||||
input_price: 0.1
|
||||
output_price: 0.8
|
||||
- name: qwen/qwen3-235b-a22b-2507
|
||||
- name: qwen/qwen5-235b-a22b-2507 # Qwen3 235B A22B Instruct 2507
|
||||
max_input_tokens: 262144
|
||||
input_price: 0.12
|
||||
output_price: 0.59
|
||||
@@ -1750,6 +1776,16 @@
|
||||
max_input_tokens: 262144
|
||||
input_price: 0.071
|
||||
output_price: 0.285
|
||||
- name: qwen/qwen3-vl-32b-instruct
|
||||
max_input_tokens: 262144
|
||||
input_price: 0.35
|
||||
output_price: 1.1
|
||||
supports_vision: true
|
||||
- name: qwen/qwen3-vl-8b-instruct
|
||||
max_input_tokens: 262144
|
||||
input_price: 0.08
|
||||
output_price: 0.50
|
||||
supports_vision: true
|
||||
- name: qwen/qwen3-coder-plus
|
||||
max_input_tokens: 128000
|
||||
input_price: 1
|
||||
@@ -1760,30 +1796,26 @@
|
||||
input_price: 0.3
|
||||
output_price: 1.5
|
||||
supports_function_calling: true
|
||||
- name: qwen/qwen3-coder # Qwen3 Coder 480B A35B
|
||||
max_input_tokens: 262144
|
||||
input_price: 0.22
|
||||
output_price: 0.95
|
||||
supports_function_calling: true
|
||||
- name: qwen/qwen3-coder-30b-a3b-instruct
|
||||
max_input_tokens: 262144
|
||||
input_price: 0.052
|
||||
output_price: 0.207
|
||||
supports_function_calling: true
|
||||
- name: qwen/qwen-2.5-72b-instruct
|
||||
max_input_tokens: 131072
|
||||
input_price: 0.35
|
||||
output_price: 0.4
|
||||
supports_function_calling: true
|
||||
- name: qwen/qwen2.5-vl-72b-instruct
|
||||
max_input_tokens: 32000
|
||||
input_price: 0.7
|
||||
output_price: 0.7
|
||||
supports_vision: true
|
||||
- name: qwen/qwen-2.5-coder-32b-instruct
|
||||
max_input_tokens: 32768
|
||||
input_price: 0.18
|
||||
output_price: 0.18
|
||||
- name: moonshotai/kimi-k2-0905
|
||||
max_input_tokens: 262144
|
||||
input_price: 0.296
|
||||
output_price: 1.185
|
||||
supports_function_calling: true
|
||||
- name: moonshotai/kimi-k2-thinking
|
||||
max_input_tokens: 262144
|
||||
input_price: 0.45
|
||||
output_price: 2.35
|
||||
supports_function_calling: true
|
||||
- name: moonshotai/kimi-dev-72b
|
||||
max_input_tokens: 131072
|
||||
input_price: 0.29
|
||||
@@ -1804,6 +1836,11 @@
|
||||
input_price: 0.2
|
||||
output_price: 1.5
|
||||
supports_function_calling: true
|
||||
- name: amazon/nova-premier-v1
|
||||
max_input_tokens: 1000000
|
||||
input_price: 2.5
|
||||
output_price: 12.5
|
||||
supports_vision: true
|
||||
- name: amazon/nova-pro-v1
|
||||
max_input_tokens: 300000
|
||||
max_output_tokens: 5120
|
||||
@@ -1850,29 +1887,15 @@
|
||||
patch:
|
||||
body:
|
||||
include_reasoning: true
|
||||
- name: minimax/minimax-01
|
||||
max_input_tokens: 1000192
|
||||
input_price: 0.2
|
||||
output_price: 1.1
|
||||
- name: minimax/minimax-m2
|
||||
max_input_tokens: 196608
|
||||
input_price: 0.15
|
||||
output_price: 0.45
|
||||
- name: z-ai/glm-4.6
|
||||
max_input_tokens: 202752
|
||||
input_price: 0.5
|
||||
output_price: 1.75
|
||||
supports_function_calling: true
|
||||
- name: z-ai/glm-4.5
|
||||
max_input_tokens: 131072
|
||||
input_price: 0.2
|
||||
output_price: 0.2
|
||||
supports_function_calling: true
|
||||
- name: z-ai/glm-4.5-air
|
||||
max_input_tokens: 131072
|
||||
input_price: 0.2
|
||||
output_price: 1.1
|
||||
- name: z-ai/glm-4.5v
|
||||
max_input_tokens: 65536
|
||||
input_price: 0.5
|
||||
output_price: 1.7
|
||||
supports_vision: true
|
||||
|
||||
# Links:
|
||||
# - https://github.com/marketplace?type=models
|
||||
@@ -2068,10 +2091,6 @@
|
||||
input_price: 0.08
|
||||
output_price: 0.3
|
||||
supports_vision: true
|
||||
- name: meta-llama/Llama-3.3-70B-Instruct
|
||||
max_input_tokens: 131072
|
||||
input_price: 0.23
|
||||
output_price: 0.40
|
||||
- name: Qwen/Qwen3-Next-80B-A3B-Instruct
|
||||
max_input_tokens: 262144
|
||||
input_price: 0.14
|
||||
@@ -2100,27 +2119,15 @@
|
||||
input_price: 0.07
|
||||
output_price: 0.27
|
||||
supports_function_calling: true
|
||||
- name: Qwen/Qwen3-235B-A22B
|
||||
max_input_tokens: 40960
|
||||
input_price: 0.15
|
||||
output_price: 0.6
|
||||
- name: Qwen/Qwen3-30B-A3B
|
||||
max_input_tokens: 40960
|
||||
input_price: 0.1
|
||||
output_price: 0.3
|
||||
- name: Qwen/Qwen3-32B
|
||||
max_input_tokens: 40960
|
||||
input_price: 0.1
|
||||
output_price: 0.3
|
||||
- name: Qwen/Qwen2.5-72B-Instruct
|
||||
max_input_tokens: 32768
|
||||
input_price: 0.23
|
||||
output_price: 0.40
|
||||
supports_function_calling: true
|
||||
- name: Qwen/Qwen2.5-Coder-32B-Instruct
|
||||
max_input_tokens: 32768
|
||||
input_price: 0.07
|
||||
output_price: 0.16
|
||||
- name: Qwen/Qwen3-VL-8B-Instruct
|
||||
max_input_tokens: 262144
|
||||
input_price: 0.18
|
||||
output_price: 0.69
|
||||
supports_vision: true
|
||||
- name: deepseek-ai/DeepSeek-V3.2-Exp
|
||||
max_input_tokens: 163840
|
||||
input_price: 0.27
|
||||
@@ -2145,35 +2152,21 @@
|
||||
max_input_tokens: 32768
|
||||
input_price: 0.06
|
||||
output_price: 0.12
|
||||
- name: mistralai/Devstral-Small-2507
|
||||
max_input_tokens: 131072
|
||||
input_price: 0.07
|
||||
output_price: 0.28
|
||||
- name: moonshotai/Kimi-K2-Instruct-0905
|
||||
max_input_tokens: 262144
|
||||
input_price: 0.5
|
||||
output_price: 2.0
|
||||
supports_function_calling: true
|
||||
- name: moonshotai/Kimi-K2-Thinking
|
||||
max_input_tokens: 262144
|
||||
input_price: 0.55
|
||||
output_price: 2.5
|
||||
supports_function_calling: true
|
||||
- name: zai-org/GLM-4.6
|
||||
max_input_tokens: 202752
|
||||
input_price: 0.6
|
||||
output_price: 1.9
|
||||
supports_function_calling: true
|
||||
- name: zai-org/GLM-4.5
|
||||
max_input_tokens: 131072
|
||||
input_price: 0.55
|
||||
output_price: 2.0
|
||||
supports_function_calling: true
|
||||
- name: zai-org/GLM-4.5-Air
|
||||
max_input_tokens: 131072
|
||||
input_price: 0.2
|
||||
output_price: 1.1
|
||||
supports_function_calling: true
|
||||
- name: zai-org/GLM-4.5V
|
||||
max_input_tokens: 65536
|
||||
input_price: 0.5
|
||||
output_price: 1.7
|
||||
supports_vision: true
|
||||
- name: BAAI/bge-large-en-v1.5
|
||||
type: embedding
|
||||
input_price: 0.01
|
||||
@@ -2271,4 +2264,4 @@
|
||||
- name: rerank-2-lite
|
||||
type: reranker
|
||||
max_input_tokens: 8000
|
||||
input_price: 0.02
|
||||
input_price: 0.02
|
||||
@@ -228,7 +228,7 @@ macro_rules! config_get_fn {
|
||||
std::env::var(&env_name)
|
||||
.ok()
|
||||
.or_else(|| self.config.$field_name.clone())
|
||||
.ok_or_else(|| anyhow::anyhow!("Miss '{}'", stringify!($field_name)))
|
||||
.ok_or_else(|| anyhow::anyhow!("Missing '{}'", stringify!($field_name)))
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
+1
-1
@@ -800,7 +800,7 @@ impl Config {
|
||||
|| s == "all"
|
||||
}) {
|
||||
bail!(
|
||||
"Some of the specified MCP servers in 'enabled_mcp_servers' are configured. Please check your MCP server configuration."
|
||||
"Some of the specified MCP servers in 'enabled_mcp_servers' are not fully configured. Please check your MCP server configuration."
|
||||
);
|
||||
}
|
||||
}
|
||||
|
||||
Reference in New Issue
Block a user