v0.46.0: security, Docker UID/GID, model discovery, i18n, cancel fix
* fix: decode HTML entities before markdown processing + zh/zh-Hant translations (#239) Adds decode() helper in renderMd() to fix double-escaping of HTML entities from LLM output (e.g. <code> becoming &lt;code&gt; instead of rendering). XSS-safe: decode runs before esc(), only 5 entity patterns. Also adds 40+ missing zh (Simplified Chinese) translation keys and a new zh-Hant (Traditional Chinese) locale with 163 keys. Fix applied: removed duplicate settings_label_notifications key in both zh and zh-Hant locales. Fixes #240 * fix: restore custom model list discovery with config api key (#238) get_available_models() now reads api_key from config.yaml before env vars: 1. model.api_key 2. providers.<active>.api_key / providers.custom.api_key 3. env var fallbacks (HERMES_API_KEY, OPENAI_API_KEY, etc.) Also adds OpenAI/Python User-Agent header and a regression test covering authenticated /v1/models discovery. Fixes users with LM Studio / Ollama custom endpoints configured in config.yaml whose model picker silently collapsed to the default model. * feat: Docker UID/GID matching to avoid root-owned .hermes files (#237) Adds docker_init.bash with hermeswebuitoo/hermeswebui user pattern so container files match the host user UID/GID. Prevents .hermes volume mounts from being owned by root when using a non-root host user. Configure via WANTED_UID and WANTED_GID env vars (default 1000/1000). Readme updated with setup instructions. Fix applied: removed duplicate WANTED_GID=1000 line in docker-compose.yml that was overriding the ${GID:-1000} variable expansion. * security: redact credentials from API responses and fix credential file permissions (#243) Adds response-layer credential redaction to three endpoints: - GET /api/session — messages[], tool_calls[], and title - GET /api/session/export — download also redacted - SSE done event — session payload in stream - GET /api/memory — MEMORY.md and USER.md content Adds api/startup.py with fix_credential_permissions() at server startup. Adds 13 tests in tests/test_security_redaction.py. Merged with #237 container detection changes in server.py. * fix: cancel button now interrupts agent and cleans up UI state (#244) Wires agent.interrupt() into cancel_stream() so the backend actually stops tool execution when the user clicks Cancel, rather than only stopping the SSE stream while the agent keeps running. Changes: - api/config.py: adds AGENT_INSTANCES dict (stream_id -> AIAgent) - api/streaming.py: stores agent in AGENT_INSTANCES after creation, checks CANCEL_FLAGS immediately after store (race condition fix), calls agent.interrupt() in cancel_stream(), cleans up in finally block - static/boot.js: removes stale setStatus(cancelling) call - static/messages.js: setBusy(false)/setStatus('') unconditionally on cancel Race condition fix: after storing agent in AGENT_INSTANCES, immediately checks if CANCEL_FLAGS[stream_id] is already set (cancel arrived during agent init) and interrupts before starting. Check is inside the same STREAMS_LOCK acquisition, making it atomic. New test file: tests/test_cancel_interrupt.py with 6 unit tests. * docs: v0.46.0 release notes, bump version, update test counts --------- Co-authored-by: Nathan Esquenazi <nesquena@gmail.com>
This commit is contained in:
@@ -613,15 +613,33 @@ def get_available_models() -> dict:
|
||||
except ValueError:
|
||||
pass
|
||||
|
||||
# Resolve API key from environment (check profile .env keys too)
|
||||
# Resolve API key for the custom / OpenAI-compatible endpoint.
|
||||
# Priority:
|
||||
# 1. model.api_key in config.yaml
|
||||
# 2. provider-specific providers.<active>.api_key / providers.custom.api_key
|
||||
# 3. env/.env fallbacks
|
||||
headers = {}
|
||||
api_key_vars = ('HERMES_API_KEY', 'HERMES_OPENAI_API_KEY', 'OPENAI_API_KEY',
|
||||
'LOCAL_API_KEY', 'OPENROUTER_API_KEY', 'API_KEY')
|
||||
for key in api_key_vars:
|
||||
api_key = all_env.get(key) or os.getenv(key)
|
||||
if api_key:
|
||||
headers['Authorization'] = f'Bearer {api_key}'
|
||||
break
|
||||
api_key = ''
|
||||
if isinstance(model_cfg, dict):
|
||||
api_key = (model_cfg.get('api_key') or '').strip()
|
||||
if not api_key:
|
||||
providers_cfg = cfg.get('providers', {})
|
||||
if isinstance(providers_cfg, dict):
|
||||
for provider_key in filter(None, [active_provider, 'custom']):
|
||||
provider_cfg = providers_cfg.get(provider_key, {})
|
||||
if isinstance(provider_cfg, dict):
|
||||
api_key = (provider_cfg.get('api_key') or '').strip()
|
||||
if api_key:
|
||||
break
|
||||
if not api_key:
|
||||
api_key_vars = ('HERMES_API_KEY', 'HERMES_OPENAI_API_KEY', 'OPENAI_API_KEY',
|
||||
'LOCAL_API_KEY', 'OPENROUTER_API_KEY', 'API_KEY')
|
||||
for key in api_key_vars:
|
||||
api_key = (all_env.get(key) or os.getenv(key) or '').strip()
|
||||
if api_key:
|
||||
break
|
||||
if api_key:
|
||||
headers['Authorization'] = f'Bearer {api_key}'
|
||||
|
||||
# Fetch model list from endpoint (with SSRF protection)
|
||||
import socket
|
||||
@@ -641,6 +659,7 @@ def get_available_models() -> dict:
|
||||
except socket.gaierror:
|
||||
pass # DNS resolution failed -- let urllib handle it
|
||||
req = urllib.request.Request(endpoint_url, method='GET')
|
||||
req.add_header('User-Agent', 'OpenAI/Python 1.0')
|
||||
for k, v in headers.items():
|
||||
req.add_header(k, v)
|
||||
with urllib.request.urlopen(req, timeout=10) as response:
|
||||
@@ -789,6 +808,7 @@ CHAT_LOCK = threading.Lock()
|
||||
STREAMS: dict = {}
|
||||
STREAMS_LOCK = threading.Lock()
|
||||
CANCEL_FLAGS: dict = {}
|
||||
AGENT_INSTANCES: dict = {} # stream_id -> AIAgent instance for interrupt propagation
|
||||
SERVER_START_TIME = time.time()
|
||||
|
||||
# ── Thread-local env context ─────────────────────────────────────────────────
|
||||
|
||||
Reference in New Issue
Block a user