Files
ProxMenux/AppImage/config/verified_ai_models.json

78 lines
3.2 KiB
JSON
Raw Normal View History

2026-03-20 23:21:00 +01:00
{
"_description": "Verified AI models for ProxMenux notifications. Only models listed here will be shown to users. Models are tested to work with the chat/completions API format.",
2026-04-19 13:47:12 +02:00
"_updated": "2026-04-19",
"_verifier": "Refreshed with tools/ai-models-verifier (private). Re-run before each ProxMenux release to keep the list current. The verifier and ProxMenux share the same reasoning/thinking-model handlers so their verdicts stay aligned with runtime behaviour.",
2026-03-20 23:21:00 +01:00
"groq": {
"models": [
"llama-3.3-70b-versatile",
"llama-3.1-70b-versatile",
"llama-3.1-8b-instant",
"llama3-70b-8192",
"llama3-8b-8192",
"mixtral-8x7b-32768",
"gemma2-9b-it"
],
2026-04-19 13:47:12 +02:00
"recommended": "llama-3.3-70b-versatile",
"_note": "Not yet re-verified in 2026-04 refresh — kept from previous curation. Run the verifier with a Groq key to prune deprecated entries."
2026-03-20 23:21:00 +01:00
},
2026-04-19 13:47:12 +02:00
2026-03-20 23:21:00 +01:00
"gemini": {
"models": [
2026-03-22 21:14:19 +01:00
"gemini-2.5-flash-lite",
2026-04-19 13:47:12 +02:00
"gemini-2.5-flash",
"gemini-3-flash-preview"
2026-03-20 23:21:00 +01:00
],
2026-04-19 13:47:12 +02:00
"recommended": "gemini-2.5-flash-lite",
"_note": "flash-lite / flash pass the verifier consistently; pro variants reject thinkingBudget=0 and are overkill for notification translation anyway. 'latest' aliases (gemini-flash-latest, gemini-flash-lite-latest) are intentionally omitted because they resolved to different models across runs and produced timeouts in some regions.",
2026-03-22 21:14:19 +01:00
"_deprecated": ["gemini-2.0-flash", "gemini-2.0-flash-lite", "gemini-1.5-flash", "gemini-1.0-pro", "gemini-pro"]
2026-03-20 23:21:00 +01:00
},
2026-04-19 13:47:12 +02:00
2026-03-20 23:21:00 +01:00
"openai": {
"models": [
2026-04-19 13:47:12 +02:00
"gpt-4.1-nano",
2026-03-20 23:48:07 +01:00
"gpt-4.1-mini",
2026-04-19 13:47:12 +02:00
"gpt-4o-mini",
"gpt-4.1",
"gpt-4o",
"gpt-5-chat-latest",
"gpt-5.4-nano",
"gpt-5.4-mini"
2026-03-20 23:21:00 +01:00
],
2026-04-19 13:47:12 +02:00
"recommended": "gpt-4.1-nano",
"_note": "Reasoning models (o-series, gpt-5/5.1/5.2 non-chat variants) are supported by openai_provider.py via max_completion_tokens + reasoning_effort=minimal, but not listed here by default: their latency is higher than the chat models and they do not improve translation quality for notifications. Add specific reasoning IDs to this list only if a user explicitly wants them."
2026-03-20 23:21:00 +01:00
},
2026-04-19 13:47:12 +02:00
2026-03-20 23:21:00 +01:00
"anthropic": {
"models": [
"claude-3-5-haiku-latest",
"claude-3-5-sonnet-latest",
"claude-3-opus-latest"
],
2026-04-19 13:47:12 +02:00
"recommended": "claude-3-5-haiku-latest",
"_note": "Not re-verified in 2026-04 refresh — kept from previous curation. Add claude-4.x / claude-4.5 / claude-4.6 / claude-4.7 variants after running the verifier with an Anthropic key."
2026-03-20 23:21:00 +01:00
},
2026-04-19 13:47:12 +02:00
2026-03-20 23:21:00 +01:00
"openrouter": {
"models": [
"meta-llama/llama-3.3-70b-instruct",
"meta-llama/llama-3.1-70b-instruct",
"meta-llama/llama-3.1-8b-instruct",
"anthropic/claude-3.5-haiku",
"anthropic/claude-3.5-sonnet",
2026-04-19 13:47:12 +02:00
"google/gemini-flash-1.5",
2026-03-20 23:21:00 +01:00
"openai/gpt-4o-mini",
"mistralai/mistral-7b-instruct",
"mistralai/mixtral-8x7b-instruct"
],
2026-04-19 13:47:12 +02:00
"recommended": "meta-llama/llama-3.3-70b-instruct",
"_note": "Not re-verified in 2026-04 refresh. google/gemini-flash-2.5-flash-lite was malformed in the previous entry and has been replaced with google/gemini-flash-1.5."
2026-03-20 23:21:00 +01:00
},
2026-04-19 13:47:12 +02:00
2026-03-20 23:21:00 +01:00
"ollama": {
"_note": "Ollama models are local, we don't filter them. User manages their own models.",
"models": [],
"recommended": ""
}
}