mirror of
https://github.com/MacRimi/ProxMenux.git
synced 2026-04-25 08:56:21 +00:00
78 lines
3.2 KiB
JSON
78 lines
3.2 KiB
JSON
{
|
|
"_description": "Verified AI models for ProxMenux notifications. Only models listed here will be shown to users. Models are tested to work with the chat/completions API format.",
|
|
"_updated": "2026-04-19",
|
|
"_verifier": "Refreshed with tools/ai-models-verifier (private). Re-run before each ProxMenux release to keep the list current. The verifier and ProxMenux share the same reasoning/thinking-model handlers so their verdicts stay aligned with runtime behaviour.",
|
|
|
|
"groq": {
|
|
"models": [
|
|
"llama-3.3-70b-versatile",
|
|
"llama-3.1-70b-versatile",
|
|
"llama-3.1-8b-instant",
|
|
"llama3-70b-8192",
|
|
"llama3-8b-8192",
|
|
"mixtral-8x7b-32768",
|
|
"gemma2-9b-it"
|
|
],
|
|
"recommended": "llama-3.3-70b-versatile",
|
|
"_note": "Not yet re-verified in 2026-04 refresh — kept from previous curation. Run the verifier with a Groq key to prune deprecated entries."
|
|
},
|
|
|
|
"gemini": {
|
|
"models": [
|
|
"gemini-2.5-flash-lite",
|
|
"gemini-2.5-flash",
|
|
"gemini-3-flash-preview"
|
|
],
|
|
"recommended": "gemini-2.5-flash-lite",
|
|
"_note": "flash-lite / flash pass the verifier consistently; pro variants reject thinkingBudget=0 and are overkill for notification translation anyway. 'latest' aliases (gemini-flash-latest, gemini-flash-lite-latest) are intentionally omitted because they resolved to different models across runs and produced timeouts in some regions.",
|
|
"_deprecated": ["gemini-2.0-flash", "gemini-2.0-flash-lite", "gemini-1.5-flash", "gemini-1.0-pro", "gemini-pro"]
|
|
},
|
|
|
|
"openai": {
|
|
"models": [
|
|
"gpt-4.1-nano",
|
|
"gpt-4.1-mini",
|
|
"gpt-4o-mini",
|
|
"gpt-4.1",
|
|
"gpt-4o",
|
|
"gpt-5-chat-latest",
|
|
"gpt-5.4-nano",
|
|
"gpt-5.4-mini"
|
|
],
|
|
"recommended": "gpt-4.1-nano",
|
|
"_note": "Reasoning models (o-series, gpt-5/5.1/5.2 non-chat variants) are supported by openai_provider.py via max_completion_tokens + reasoning_effort=minimal, but not listed here by default: their latency is higher than the chat models and they do not improve translation quality for notifications. Add specific reasoning IDs to this list only if a user explicitly wants them."
|
|
},
|
|
|
|
"anthropic": {
|
|
"models": [
|
|
"claude-3-5-haiku-latest",
|
|
"claude-3-5-sonnet-latest",
|
|
"claude-3-opus-latest"
|
|
],
|
|
"recommended": "claude-3-5-haiku-latest",
|
|
"_note": "Not re-verified in 2026-04 refresh — kept from previous curation. Add claude-4.x / claude-4.5 / claude-4.6 / claude-4.7 variants after running the verifier with an Anthropic key."
|
|
},
|
|
|
|
"openrouter": {
|
|
"models": [
|
|
"meta-llama/llama-3.3-70b-instruct",
|
|
"meta-llama/llama-3.1-70b-instruct",
|
|
"meta-llama/llama-3.1-8b-instruct",
|
|
"anthropic/claude-3.5-haiku",
|
|
"anthropic/claude-3.5-sonnet",
|
|
"google/gemini-flash-1.5",
|
|
"openai/gpt-4o-mini",
|
|
"mistralai/mistral-7b-instruct",
|
|
"mistralai/mixtral-8x7b-instruct"
|
|
],
|
|
"recommended": "meta-llama/llama-3.3-70b-instruct",
|
|
"_note": "Not re-verified in 2026-04 refresh. google/gemini-flash-2.5-flash-lite was malformed in the previous entry and has been replaced with google/gemini-flash-1.5."
|
|
},
|
|
|
|
"ollama": {
|
|
"_note": "Ollama models are local, we don't filter them. User manages their own models.",
|
|
"models": [],
|
|
"recommended": ""
|
|
}
|
|
}
|