defc2c116de5a00caa1eac429409b24b1e0fe6af
- get_llm_provider() now accepts provider + model args so DB overrides propagate through to all provider constructors (was always reading env vars, ignoring the admin UI selection) - /test-llm replaced with lightweight ping (max_tokens=20) instead of running a full bill analysis; shows model name + reply, no truncation - /api/settings/llm-models endpoint fetches available models live from each provider's API (OpenAI, Anthropic REST, Gemini, Ollama) - Admin UI model picker dynamically populated from provider API; falls back to manual text input on error; Custom model name option kept - Default Gemini model updated: gemini-1.5-pro → gemini-2.0-flash Co-Authored-By: Jack Levy
Description
Self-hosted US Congress monitoring with AI bill analysis, personalized tracking, and multi-channel notifications.
Languages
Python
53.6%
TypeScript
45.8%
CSS
0.2%
Dockerfile
0.2%