mirror of
https://github.com/MacRimi/ProxMenux.git
synced 2026-05-01 11:56:21 +00:00
Update notification service
This commit is contained in:
@@ -19,10 +19,11 @@
|
|||||||
"models": [
|
"models": [
|
||||||
"gemini-2.5-flash",
|
"gemini-2.5-flash",
|
||||||
"gemini-2.5-pro",
|
"gemini-2.5-pro",
|
||||||
"gemini-2.0-flash-lite"
|
"gemini-1.5-flash"
|
||||||
],
|
],
|
||||||
"recommended": "gemini-2.5-flash",
|
"recommended": "gemini-1.5-flash",
|
||||||
"_deprecated": ["gemini-2.0-flash", "gemini-1.0-pro", "gemini-pro"]
|
"_note": "gemini-2.5-* may have availability issues. gemini-1.5-flash is more stable.",
|
||||||
|
"_deprecated": ["gemini-2.0-flash", "gemini-2.0-flash-lite", "gemini-1.0-pro", "gemini-pro"]
|
||||||
},
|
},
|
||||||
|
|
||||||
"openai": {
|
"openai": {
|
||||||
|
|||||||
@@ -1270,10 +1270,11 @@ AI_LANGUAGES = {
|
|||||||
|
|
||||||
# Token limits for different detail levels
|
# Token limits for different detail levels
|
||||||
# max_tokens is a LIMIT, not fixed consumption - you only pay for tokens actually generated
|
# max_tokens is a LIMIT, not fixed consumption - you only pay for tokens actually generated
|
||||||
|
# Note: Some providers (especially Gemini) may have lower default limits, so we use generous values
|
||||||
AI_DETAIL_TOKENS = {
|
AI_DETAIL_TOKENS = {
|
||||||
'brief': 300, # Short messages, 2-3 lines
|
'brief': 500, # Short messages, 2-3 lines
|
||||||
'standard': 1000, # Standard messages, sufficient for 15-20 VMs
|
'standard': 1500, # Standard messages, sufficient for 15-20 VMs
|
||||||
'detailed': 2000, # Complete technical reports with all details
|
'detailed': 3000, # Complete technical reports with all details
|
||||||
}
|
}
|
||||||
|
|
||||||
# System prompt template - informative, no recommendations
|
# System prompt template - informative, no recommendations
|
||||||
|
|||||||
Reference in New Issue
Block a user