mirror of
https://github.com/NousResearch/hermes-agent.git
synced 2026-05-04 09:47:54 +08:00
PR #17660 landed a sweep of CI fixes but left three loose ends:
1. tests/cli/test_cli_loading_indicator.py::test_reload_mcp_sets_busy_state_
and_prints_status — /reload-mcp gained a prompt-cache-invalidation
confirmation (commit 4d7fc0f37) that was never wired into this test.
The test exercises the loading-indicator path, so pre-approve via
config and go straight into _reload_mcp().
2. tools/mcp_tool.py _make_tool_handler — the added
getattr(server, '_rpc_lock', None) + 'skip the lock if missing'
branch is inconsistent with four sibling call sites that still
direct-access server._rpc_lock. The lock is guaranteed by
MCPServerTask.__init__; falling through to an unlocked
session.call_tool would silently serialize-strip RPCs if the guard
ever triggered. Restore direct access.
3. tui_gateway/server.py _messages_as_conversation — the helper
existed only to catch 'TypeError: include_ancestors unexpected'
from mocked SessionDBs that don't actually exist. The real
SessionDB.get_messages_as_conversation has accepted
include_ancestors since introduction, and every test FakeDB in
the repo already declares the kwarg. Remove the shim, inline the
two call sites.
5728 lines
201 KiB
Python
5728 lines
201 KiB
Python
import atexit
|
|
import concurrent.futures
|
|
import contextvars
|
|
import copy
|
|
import json
|
|
import logging
|
|
import os
|
|
import queue
|
|
import subprocess
|
|
import sys
|
|
import threading
|
|
import time
|
|
import uuid
|
|
from datetime import datetime
|
|
from pathlib import Path
|
|
from typing import Any, Optional
|
|
|
|
from hermes_constants import get_hermes_home
|
|
from hermes_cli.env_loader import load_hermes_dotenv
|
|
from tui_gateway.transport import (
|
|
StdioTransport,
|
|
Transport,
|
|
bind_transport,
|
|
current_transport,
|
|
reset_transport,
|
|
)
|
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
_hermes_home = get_hermes_home()
|
|
load_hermes_dotenv(
|
|
hermes_home=_hermes_home, project_env=Path(__file__).parent.parent / ".env"
|
|
)
|
|
|
|
|
|
# ── Panic logger ─────────────────────────────────────────────────────
|
|
# Gateway crashes in a TUI session leave no forensics: stdout is the
|
|
# JSON-RPC pipe (TUI side parses it, doesn't log raw), the root logger
|
|
# only catches handled warnings, and the subprocess exits before stderr
|
|
# flushes through the stderr->gateway.stderr event pump. This hook
|
|
# appends every unhandled exception to ~/.hermes/logs/tui_gateway_crash.log
|
|
# AND re-emits a one-line summary to stderr so the TUI can surface it in
|
|
# Activity — exactly what was missing when the voice-mode turns started
|
|
# exiting the gateway mid-TTS.
|
|
_CRASH_LOG = os.path.join(_hermes_home, "logs", "tui_gateway_crash.log")
|
|
|
|
|
|
def _panic_hook(exc_type, exc_value, exc_tb):
|
|
import traceback
|
|
|
|
trace = "".join(traceback.format_exception(exc_type, exc_value, exc_tb))
|
|
try:
|
|
os.makedirs(os.path.dirname(_CRASH_LOG), exist_ok=True)
|
|
with open(_CRASH_LOG, "a", encoding="utf-8") as f:
|
|
f.write(
|
|
f"\n=== unhandled exception · {time.strftime('%Y-%m-%d %H:%M:%S')} ===\n"
|
|
)
|
|
f.write(trace)
|
|
except Exception:
|
|
pass
|
|
# Stderr goes through to the TUI as a gateway.stderr Activity line —
|
|
# the first line here is what the user will see without opening any
|
|
# log files. Rest of the stack is still in the log for full context.
|
|
first = (
|
|
str(exc_value).strip().splitlines()[0]
|
|
if str(exc_value).strip()
|
|
else exc_type.__name__
|
|
)
|
|
print(f"[gateway-crash] {exc_type.__name__}: {first}", file=sys.stderr, flush=True)
|
|
# Chain to the default hook so the process still terminates normally.
|
|
sys.__excepthook__(exc_type, exc_value, exc_tb)
|
|
|
|
|
|
sys.excepthook = _panic_hook
|
|
|
|
|
|
def _thread_panic_hook(args):
|
|
# threading.excepthook signature: SimpleNamespace(exc_type, exc_value, exc_traceback, thread)
|
|
import traceback
|
|
|
|
trace = "".join(
|
|
traceback.format_exception(args.exc_type, args.exc_value, args.exc_traceback)
|
|
)
|
|
try:
|
|
os.makedirs(os.path.dirname(_CRASH_LOG), exist_ok=True)
|
|
with open(_CRASH_LOG, "a", encoding="utf-8") as f:
|
|
f.write(
|
|
f"\n=== thread exception · {time.strftime('%Y-%m-%d %H:%M:%S')} "
|
|
f"· thread={args.thread.name} ===\n"
|
|
)
|
|
f.write(trace)
|
|
except Exception:
|
|
pass
|
|
first_line = (
|
|
str(args.exc_value).strip().splitlines()[0]
|
|
if str(args.exc_value).strip()
|
|
else args.exc_type.__name__
|
|
)
|
|
print(
|
|
f"[gateway-crash] thread {args.thread.name} raised {args.exc_type.__name__}: {first_line}",
|
|
file=sys.stderr,
|
|
flush=True,
|
|
)
|
|
|
|
|
|
threading.excepthook = _thread_panic_hook
|
|
|
|
try:
|
|
from hermes_cli.banner import prefetch_update_check
|
|
|
|
prefetch_update_check()
|
|
except Exception:
|
|
pass
|
|
|
|
from tui_gateway.render import make_stream_renderer, render_diff, render_message
|
|
|
|
_sessions: dict[str, dict] = {}
|
|
_methods: dict[str, callable] = {}
|
|
_pending: dict[str, tuple[str, threading.Event]] = {}
|
|
_answers: dict[str, str] = {}
|
|
_db = None
|
|
_db_error: str | None = None
|
|
_stdout_lock = threading.Lock()
|
|
_cfg_lock = threading.Lock()
|
|
_cfg_cache: dict | None = None
|
|
_cfg_mtime: float | None = None
|
|
_cfg_path = None
|
|
_SLASH_WORKER_TIMEOUT_S = max(
|
|
5.0, float(os.environ.get("HERMES_TUI_SLASH_TIMEOUT_S", "45") or 45)
|
|
)
|
|
_DETAIL_SECTION_NAMES = ("thinking", "tools", "subagents", "activity")
|
|
_DETAIL_MODES = frozenset({"hidden", "collapsed", "expanded"})
|
|
|
|
# ── Async RPC dispatch (#12546) ──────────────────────────────────────
|
|
# A handful of handlers block the dispatcher loop in entry.py for seconds
|
|
# to minutes (slash.exec, cli.exec, shell.exec, session.resume,
|
|
# session.branch, session.compress, skills.manage). While they're running, inbound RPCs —
|
|
# notably approval.respond and session.interrupt — sit unread in the
|
|
# stdin pipe. We route only those slow handlers onto a small thread pool;
|
|
# everything else stays on the main thread so ordering stays sane for the
|
|
# fast path. write_json is already _stdout_lock-guarded, so concurrent
|
|
# response writes are safe.
|
|
_LONG_HANDLERS = frozenset(
|
|
{
|
|
"browser.manage",
|
|
"cli.exec",
|
|
"session.branch",
|
|
"session.compress",
|
|
"session.resume",
|
|
"shell.exec",
|
|
"skills.manage",
|
|
"slash.exec",
|
|
}
|
|
)
|
|
|
|
_pool = concurrent.futures.ThreadPoolExecutor(
|
|
max_workers=max(2, int(os.environ.get("HERMES_TUI_RPC_POOL_WORKERS", "4") or 4)),
|
|
thread_name_prefix="tui-rpc",
|
|
)
|
|
atexit.register(lambda: _pool.shutdown(wait=False, cancel_futures=True))
|
|
|
|
# Reserve real stdout for JSON-RPC only; redirect Python's stdout to stderr
|
|
# so stray print() from libraries/tools becomes harmless gateway.stderr instead
|
|
# of corrupting the JSON protocol.
|
|
_real_stdout = sys.stdout
|
|
sys.stdout = sys.stderr
|
|
|
|
# Module-level stdio transport — fallback sink when no transport is bound via
|
|
# contextvar or session. Stream resolved through a lambda so runtime monkey-
|
|
# patches of `_real_stdout` (used extensively in tests) still land correctly.
|
|
_stdio_transport = StdioTransport(lambda: _real_stdout, _stdout_lock)
|
|
|
|
|
|
class _SlashWorker:
|
|
"""Persistent HermesCLI subprocess for slash commands."""
|
|
|
|
def __init__(self, session_key: str, model: str):
|
|
self._lock = threading.Lock()
|
|
self._seq = 0
|
|
self.stderr_tail: list[str] = []
|
|
self.stdout_queue: queue.Queue[dict | None] = queue.Queue()
|
|
|
|
argv = [
|
|
sys.executable,
|
|
"-m",
|
|
"tui_gateway.slash_worker",
|
|
"--session-key",
|
|
session_key,
|
|
]
|
|
if model:
|
|
argv += ["--model", model]
|
|
|
|
self.proc = subprocess.Popen(
|
|
argv,
|
|
stdin=subprocess.PIPE,
|
|
stdout=subprocess.PIPE,
|
|
stderr=subprocess.PIPE,
|
|
text=True,
|
|
bufsize=1,
|
|
cwd=os.getcwd(),
|
|
env=os.environ.copy(),
|
|
)
|
|
threading.Thread(target=self._drain_stdout, daemon=True).start()
|
|
threading.Thread(target=self._drain_stderr, daemon=True).start()
|
|
|
|
def _drain_stdout(self):
|
|
for line in self.proc.stdout or []:
|
|
try:
|
|
self.stdout_queue.put(json.loads(line))
|
|
except json.JSONDecodeError:
|
|
continue
|
|
self.stdout_queue.put(None)
|
|
|
|
def _drain_stderr(self):
|
|
for line in self.proc.stderr or []:
|
|
if text := line.rstrip("\n"):
|
|
self.stderr_tail = (self.stderr_tail + [text])[-80:]
|
|
|
|
def run(self, command: str) -> str:
|
|
if self.proc.poll() is not None:
|
|
raise RuntimeError("slash worker exited")
|
|
|
|
with self._lock:
|
|
self._seq += 1
|
|
rid = self._seq
|
|
self.proc.stdin.write(json.dumps({"id": rid, "command": command}) + "\n")
|
|
self.proc.stdin.flush()
|
|
|
|
while True:
|
|
try:
|
|
msg = self.stdout_queue.get(timeout=_SLASH_WORKER_TIMEOUT_S)
|
|
except queue.Empty:
|
|
raise RuntimeError("slash worker timed out")
|
|
if msg is None:
|
|
break
|
|
if msg.get("id") != rid:
|
|
continue
|
|
if not msg.get("ok"):
|
|
raise RuntimeError(msg.get("error", "slash worker failed"))
|
|
return str(msg.get("output", "")).rstrip()
|
|
|
|
raise RuntimeError(
|
|
f"slash worker closed pipe{': ' + chr(10).join(self.stderr_tail[-8:]) if self.stderr_tail else ''}"
|
|
)
|
|
|
|
def close(self):
|
|
try:
|
|
if self.proc.poll() is None:
|
|
self.proc.terminate()
|
|
self.proc.wait(timeout=1)
|
|
except Exception:
|
|
try:
|
|
self.proc.kill()
|
|
except Exception:
|
|
pass
|
|
|
|
|
|
def _load_busy_input_mode() -> str:
|
|
display = _load_cfg().get("display")
|
|
if not isinstance(display, dict):
|
|
display = {}
|
|
raw = str(display.get("busy_input_mode", "") or "").strip().lower()
|
|
return raw if raw in {"queue", "steer", "interrupt"} else "interrupt"
|
|
|
|
|
|
def _notify_session_boundary(event_type: str, session_id: str | None) -> None:
|
|
"""Fire session lifecycle hooks with CLI parity."""
|
|
try:
|
|
from hermes_cli.plugins import invoke_hook as _invoke_hook
|
|
|
|
_invoke_hook(event_type, session_id=session_id, platform="tui")
|
|
except Exception:
|
|
pass
|
|
|
|
|
|
def _finalize_session(session: dict | None) -> None:
|
|
"""Best-effort finalize hook + memory commit for a session."""
|
|
if not session or session.get("_finalized"):
|
|
return
|
|
session["_finalized"] = True
|
|
|
|
agent = session.get("agent")
|
|
lock = session.get("history_lock")
|
|
if lock is not None:
|
|
with lock:
|
|
history = list(session.get("history", []))
|
|
else:
|
|
history = list(session.get("history", []))
|
|
if agent is not None and history and hasattr(agent, "commit_memory_session"):
|
|
try:
|
|
agent.commit_memory_session(history)
|
|
except Exception:
|
|
pass
|
|
|
|
session_id = getattr(agent, "session_id", None) or session.get("session_key")
|
|
_notify_session_boundary("on_session_finalize", session_id)
|
|
|
|
|
|
def _shutdown_sessions() -> None:
|
|
for session in list(_sessions.values()):
|
|
_finalize_session(session)
|
|
try:
|
|
worker = session.get("slash_worker")
|
|
if worker:
|
|
worker.close()
|
|
except Exception:
|
|
pass
|
|
|
|
|
|
atexit.register(_shutdown_sessions)
|
|
|
|
|
|
# ── Plumbing ──────────────────────────────────────────────────────────
|
|
|
|
|
|
def _get_db():
|
|
global _db, _db_error
|
|
if _db is None:
|
|
from hermes_state import SessionDB
|
|
|
|
try:
|
|
_db = SessionDB()
|
|
_db_error = None
|
|
except Exception as exc:
|
|
_db_error = str(exc)
|
|
logger.warning(
|
|
"TUI session store unavailable — continuing without state.db features: %s",
|
|
exc,
|
|
)
|
|
return None
|
|
return _db
|
|
|
|
|
|
def _db_unavailable_error(rid, *, code: int):
|
|
detail = _db_error or "state.db unavailable"
|
|
return _err(rid, code, f"state.db unavailable: {detail}")
|
|
|
|
|
|
def write_json(obj: dict) -> bool:
|
|
"""Emit one JSON frame. Routes via the most-specific transport available.
|
|
|
|
Precedence:
|
|
|
|
1. Event frames with a session id → the transport stored on that session,
|
|
so async events land with the client that owns the session even if
|
|
the emitting thread has no contextvar binding.
|
|
2. Otherwise the transport bound on the current context (set by
|
|
:func:`dispatch` for the lifetime of a request).
|
|
3. Otherwise the module-level stdio transport, matching the historical
|
|
behaviour and keeping tests that monkey-patch ``_real_stdout`` green.
|
|
"""
|
|
if obj.get("method") == "event":
|
|
sid = ((obj.get("params") or {}).get("session_id")) or ""
|
|
if sid and (t := (_sessions.get(sid) or {}).get("transport")) is not None:
|
|
return t.write(obj)
|
|
|
|
return (current_transport() or _stdio_transport).write(obj)
|
|
|
|
|
|
def _emit(event: str, sid: str, payload: dict | None = None):
|
|
params = {"type": event, "session_id": sid}
|
|
if payload is not None:
|
|
params["payload"] = payload
|
|
write_json({"jsonrpc": "2.0", "method": "event", "params": params})
|
|
|
|
|
|
def _status_update(sid: str, kind: str, text: str | None = None):
|
|
body = (text if text is not None else kind).strip()
|
|
if not body:
|
|
return
|
|
_emit(
|
|
"status.update",
|
|
sid,
|
|
{"kind": kind if text is not None else "status", "text": body},
|
|
)
|
|
|
|
|
|
def _estimate_image_tokens(width: int, height: int) -> int:
|
|
"""Very rough UI estimate for image prompt cost.
|
|
|
|
Uses 512px tiles at ~85 tokens/tile as a lightweight cross-provider hint.
|
|
This is intentionally approximate and only used for attachment display.
|
|
"""
|
|
if width <= 0 or height <= 0:
|
|
return 0
|
|
return max(1, (width + 511) // 512) * max(1, (height + 511) // 512) * 85
|
|
|
|
|
|
def _image_meta(path: Path) -> dict:
|
|
meta = {"name": path.name}
|
|
try:
|
|
from PIL import Image
|
|
|
|
with Image.open(path) as img:
|
|
width, height = img.size
|
|
meta["width"] = int(width)
|
|
meta["height"] = int(height)
|
|
meta["token_estimate"] = _estimate_image_tokens(int(width), int(height))
|
|
except Exception:
|
|
pass
|
|
return meta
|
|
|
|
|
|
def _ok(rid, result: dict) -> dict:
|
|
return {"jsonrpc": "2.0", "id": rid, "result": result}
|
|
|
|
|
|
def _err(rid, code: int, msg: str) -> dict:
|
|
return {"jsonrpc": "2.0", "id": rid, "error": {"code": code, "message": msg}}
|
|
|
|
|
|
def method(name: str):
|
|
def dec(fn):
|
|
_methods[name] = fn
|
|
return fn
|
|
|
|
return dec
|
|
|
|
|
|
def handle_request(req: dict) -> dict | None:
|
|
fn = _methods.get(req.get("method", ""))
|
|
if not fn:
|
|
return _err(req.get("id"), -32601, f"unknown method: {req.get('method')}")
|
|
return fn(req.get("id"), req.get("params", {}))
|
|
|
|
|
|
def dispatch(req: dict, transport: Optional[Transport] = None) -> dict | None:
|
|
"""Route inbound RPCs — long handlers to the pool, everything else inline.
|
|
|
|
Returns a response dict when handled inline. Returns None when the
|
|
handler was scheduled on the pool; the worker writes its own response
|
|
via the bound transport when done.
|
|
|
|
*transport* (optional): pins every write produced by this request —
|
|
including any events emitted by the handler — to the given transport.
|
|
Omitting it falls back to the module-level stdio transport, preserving
|
|
the original behaviour for ``tui_gateway.entry``.
|
|
"""
|
|
t = transport or _stdio_transport
|
|
token = bind_transport(t)
|
|
try:
|
|
if req.get("method") not in _LONG_HANDLERS:
|
|
return handle_request(req)
|
|
|
|
# Snapshot the context so the pool worker sees the bound transport.
|
|
ctx = contextvars.copy_context()
|
|
|
|
def run():
|
|
try:
|
|
resp = handle_request(req)
|
|
except Exception as exc:
|
|
resp = _err(req.get("id"), -32000, f"handler error: {exc}")
|
|
if resp is not None:
|
|
t.write(resp)
|
|
|
|
_pool.submit(lambda: ctx.run(run))
|
|
|
|
return None
|
|
finally:
|
|
reset_transport(token)
|
|
|
|
|
|
def _wait_agent(session: dict, rid: str, timeout: float = 30.0) -> dict | None:
|
|
ready = session.get("agent_ready")
|
|
if ready is not None and not ready.wait(timeout=timeout):
|
|
return _err(rid, 5032, "agent initialization timed out")
|
|
err = session.get("agent_error")
|
|
return _err(rid, 5032, err) if err else None
|
|
|
|
|
|
def _start_agent_build(sid: str, session: dict) -> None:
|
|
"""Start building the real AIAgent for a TUI session, once.
|
|
|
|
Classic `hermes` shows the prompt before constructing AIAgent; the TUI used
|
|
to eagerly build it during session.create, making startup feel blocked on
|
|
tool discovery/model metadata even though the composer was visible. Keep
|
|
the shell responsive by deferring this work until the first prompt (or any
|
|
command that actually needs the agent), while retaining the same ready/error
|
|
event contract for the frontend.
|
|
"""
|
|
ready = session.get("agent_ready")
|
|
if ready is None:
|
|
return
|
|
lock = session.setdefault("agent_build_lock", threading.Lock())
|
|
with lock:
|
|
if ready.is_set() or session.get("agent_build_started"):
|
|
return
|
|
session["agent_build_started"] = True
|
|
key = session["session_key"]
|
|
|
|
def _build() -> None:
|
|
current = _sessions.get(sid)
|
|
if current is None:
|
|
ready.set()
|
|
return
|
|
|
|
worker = None
|
|
notify_registered = False
|
|
try:
|
|
tokens = _set_session_context(key)
|
|
try:
|
|
agent = _make_agent(sid, key)
|
|
finally:
|
|
_clear_session_context(tokens)
|
|
|
|
db = _get_db()
|
|
if db is not None:
|
|
db.create_session(key, source="tui", model=_resolve_model())
|
|
pending_title = (current.get("pending_title") or "").strip()
|
|
if pending_title:
|
|
try:
|
|
title_applied = db.set_session_title(key, pending_title)
|
|
if title_applied:
|
|
current["pending_title"] = None
|
|
else:
|
|
existing_row = db.get_session(key)
|
|
existing_title = ((existing_row or {}).get("title") or "").strip()
|
|
if existing_title == pending_title:
|
|
current["pending_title"] = None
|
|
else:
|
|
logger.info(
|
|
"Pending title still queued for session %s (wanted=%r, current=%r)",
|
|
sid,
|
|
pending_title,
|
|
existing_title,
|
|
)
|
|
except ValueError as e:
|
|
current["pending_title"] = None
|
|
logger.info("Dropping pending title for session %s: %s", sid, e)
|
|
except Exception:
|
|
logger.warning("Failed to apply pending title for session %s", sid, exc_info=True)
|
|
current["agent"] = agent
|
|
|
|
try:
|
|
worker = _SlashWorker(key, getattr(agent, "model", _resolve_model()))
|
|
current["slash_worker"] = worker
|
|
except Exception:
|
|
pass
|
|
|
|
try:
|
|
from tools.approval import (
|
|
register_gateway_notify,
|
|
load_permanent_allowlist,
|
|
)
|
|
register_gateway_notify(key, lambda data: _emit("approval.request", sid, data))
|
|
notify_registered = True
|
|
load_permanent_allowlist()
|
|
except Exception:
|
|
pass
|
|
|
|
_wire_callbacks(sid)
|
|
_notify_session_boundary("on_session_reset", key)
|
|
|
|
info = _session_info(agent)
|
|
warn = _probe_credentials(agent)
|
|
if warn:
|
|
info["credential_warning"] = warn
|
|
cfg_warn = _probe_config_health(_load_cfg())
|
|
if cfg_warn:
|
|
info["config_warning"] = cfg_warn
|
|
logger.warning(cfg_warn)
|
|
_emit("session.info", sid, info)
|
|
except Exception as e:
|
|
current["agent_error"] = str(e)
|
|
_emit("error", sid, {"message": f"agent init failed: {e}"})
|
|
finally:
|
|
if _sessions.get(sid) is not current:
|
|
if worker is not None:
|
|
try:
|
|
worker.close()
|
|
except Exception:
|
|
pass
|
|
if notify_registered:
|
|
try:
|
|
from tools.approval import unregister_gateway_notify
|
|
unregister_gateway_notify(key)
|
|
except Exception:
|
|
pass
|
|
ready.set()
|
|
|
|
threading.Thread(target=_build, daemon=True).start()
|
|
|
|
|
|
def _sess_nowait(params, rid):
|
|
s = _sessions.get(params.get("session_id") or "")
|
|
return (s, None) if s else (None, _err(rid, 4001, "session not found"))
|
|
|
|
|
|
def _sess(params, rid):
|
|
s, err = _sess_nowait(params, rid)
|
|
if err:
|
|
return (None, err)
|
|
_start_agent_build(params.get("session_id") or "", s)
|
|
return (s, _wait_agent(s, rid))
|
|
|
|
|
|
def _normalize_completion_path(path_part: str) -> str:
|
|
expanded = os.path.expanduser(path_part)
|
|
if os.name != "nt":
|
|
normalized = expanded.replace("\\", "/")
|
|
if (
|
|
len(normalized) >= 3
|
|
and normalized[1] == ":"
|
|
and normalized[2] == "/"
|
|
and normalized[0].isalpha()
|
|
):
|
|
return f"/mnt/{normalized[0].lower()}/{normalized[3:]}"
|
|
return expanded
|
|
|
|
|
|
# ── Config I/O ────────────────────────────────────────────────────────
|
|
|
|
|
|
# Keep aligned with `INDICATOR_STYLES` / `DEFAULT_INDICATOR_STYLE` in
|
|
# ``ui-tui/src/app/interfaces.ts`` — both ends validate against the
|
|
# same shape so `config.get indicator` and the live TUI render agree.
|
|
_INDICATOR_STYLES: tuple[str, ...] = ("ascii", "emoji", "kaomoji", "unicode")
|
|
_INDICATOR_DEFAULT = "kaomoji"
|
|
|
|
|
|
def _load_cfg() -> dict:
|
|
global _cfg_cache, _cfg_mtime, _cfg_path
|
|
try:
|
|
import yaml
|
|
|
|
p = _hermes_home / "config.yaml"
|
|
mtime = p.stat().st_mtime if p.exists() else None
|
|
with _cfg_lock:
|
|
if _cfg_cache is not None and _cfg_mtime == mtime and _cfg_path == p:
|
|
return copy.deepcopy(_cfg_cache)
|
|
if p.exists():
|
|
with open(p) as f:
|
|
data = yaml.safe_load(f) or {}
|
|
else:
|
|
data = {}
|
|
with _cfg_lock:
|
|
_cfg_cache = copy.deepcopy(data)
|
|
_cfg_mtime = mtime
|
|
_cfg_path = p
|
|
return data
|
|
except Exception:
|
|
pass
|
|
return {}
|
|
|
|
|
|
def _save_cfg(cfg: dict):
|
|
global _cfg_cache, _cfg_mtime, _cfg_path
|
|
import yaml
|
|
|
|
path = _hermes_home / "config.yaml"
|
|
with open(path, "w") as f:
|
|
yaml.safe_dump(cfg, f)
|
|
with _cfg_lock:
|
|
_cfg_cache = copy.deepcopy(cfg)
|
|
_cfg_path = path
|
|
try:
|
|
_cfg_mtime = path.stat().st_mtime
|
|
except Exception:
|
|
_cfg_mtime = None
|
|
|
|
|
|
def _set_session_context(session_key: str) -> list:
|
|
try:
|
|
from gateway.session_context import set_session_vars
|
|
|
|
return set_session_vars(session_key=session_key)
|
|
except Exception:
|
|
return []
|
|
|
|
|
|
def _clear_session_context(tokens: list) -> None:
|
|
if not tokens:
|
|
return
|
|
try:
|
|
from gateway.session_context import clear_session_vars
|
|
|
|
clear_session_vars(tokens)
|
|
except Exception:
|
|
pass
|
|
|
|
|
|
def _enable_gateway_prompts() -> None:
|
|
"""Route approvals through gateway callbacks instead of CLI input()."""
|
|
os.environ["HERMES_GATEWAY_SESSION"] = "1"
|
|
os.environ["HERMES_EXEC_ASK"] = "1"
|
|
os.environ["HERMES_INTERACTIVE"] = "1"
|
|
|
|
|
|
# ── Blocking prompt factory ──────────────────────────────────────────
|
|
|
|
|
|
def _block(event: str, sid: str, payload: dict, timeout: int = 300) -> str:
|
|
rid = uuid.uuid4().hex[:8]
|
|
ev = threading.Event()
|
|
_pending[rid] = (sid, ev)
|
|
payload["request_id"] = rid
|
|
_emit(event, sid, payload)
|
|
ev.wait(timeout=timeout)
|
|
_pending.pop(rid, None)
|
|
return _answers.pop(rid, "")
|
|
|
|
|
|
def _clear_pending(sid: str | None = None) -> None:
|
|
"""Release pending prompts with an empty answer.
|
|
|
|
When *sid* is provided, only prompts owned by that session are
|
|
released — critical for session.interrupt, which must not
|
|
collaterally cancel clarify/sudo/secret prompts on unrelated
|
|
sessions sharing the same tui_gateway process. When *sid* is
|
|
None, every pending prompt is released (used during shutdown).
|
|
"""
|
|
for rid, (owner_sid, ev) in list(_pending.items()):
|
|
if sid is None or owner_sid == sid:
|
|
_answers[rid] = ""
|
|
ev.set()
|
|
|
|
|
|
# ── Agent factory ────────────────────────────────────────────────────
|
|
|
|
|
|
def resolve_skin() -> dict:
|
|
try:
|
|
from hermes_cli.skin_engine import init_skin_from_config, get_active_skin
|
|
|
|
init_skin_from_config(_load_cfg())
|
|
skin = get_active_skin()
|
|
return {
|
|
"name": skin.name,
|
|
"colors": skin.colors,
|
|
"branding": skin.branding,
|
|
"banner_logo": skin.banner_logo,
|
|
"banner_hero": skin.banner_hero,
|
|
"tool_prefix": skin.tool_prefix,
|
|
"help_header": (skin.branding or {}).get("help_header", ""),
|
|
}
|
|
except Exception:
|
|
return {}
|
|
|
|
|
|
def _resolve_model() -> str:
|
|
env = (
|
|
os.environ.get("HERMES_MODEL", "")
|
|
or os.environ.get("HERMES_INFERENCE_MODEL", "")
|
|
).strip()
|
|
if env:
|
|
return env
|
|
m = _load_cfg().get("model", "")
|
|
if isinstance(m, dict):
|
|
return str(m.get("default", "") or "").strip()
|
|
if isinstance(m, str) and m:
|
|
return m.strip()
|
|
return "anthropic/claude-sonnet-4"
|
|
|
|
|
|
def _resolve_startup_runtime() -> tuple[str, str | None]:
|
|
model = _resolve_model()
|
|
explicit_provider = os.environ.get("HERMES_TUI_PROVIDER", "").strip()
|
|
if explicit_provider:
|
|
return model, explicit_provider
|
|
|
|
explicit_model = (
|
|
os.environ.get("HERMES_MODEL", "")
|
|
or os.environ.get("HERMES_INFERENCE_MODEL", "")
|
|
).strip()
|
|
if not explicit_model:
|
|
return model, None
|
|
|
|
try:
|
|
from hermes_cli.models import detect_static_provider_for_model
|
|
|
|
cfg = _load_cfg().get("model") or {}
|
|
current_provider = (
|
|
(
|
|
str(cfg.get("provider") or "").strip().lower()
|
|
if isinstance(cfg, dict)
|
|
else ""
|
|
)
|
|
or os.environ.get("HERMES_INFERENCE_PROVIDER", "").strip().lower()
|
|
or "auto"
|
|
)
|
|
detected = detect_static_provider_for_model(explicit_model, current_provider)
|
|
if detected:
|
|
provider, detected_model = detected
|
|
return detected_model, provider
|
|
except Exception:
|
|
pass
|
|
return model, None
|
|
|
|
|
|
def _write_config_key(key_path: str, value):
|
|
cfg = _load_cfg()
|
|
current = cfg
|
|
keys = key_path.split(".")
|
|
for key in keys[:-1]:
|
|
if key not in current or not isinstance(current.get(key), dict):
|
|
current[key] = {}
|
|
current = current[key]
|
|
current[keys[-1]] = value
|
|
_save_cfg(cfg)
|
|
|
|
|
|
_STATUSBAR_MODES = frozenset({"off", "top", "bottom"})
|
|
|
|
|
|
def _coerce_statusbar(raw) -> str:
|
|
if raw is False:
|
|
return "off"
|
|
if isinstance(raw, str) and (s := raw.strip().lower()) in _STATUSBAR_MODES:
|
|
return s
|
|
return "top"
|
|
|
|
|
|
def _display_mouse_tracking(display: dict) -> bool:
|
|
"""Return canonical display.mouse_tracking with legacy tui_mouse fallback."""
|
|
if not isinstance(display, dict):
|
|
return True
|
|
if "mouse_tracking" in display:
|
|
raw = display.get("mouse_tracking")
|
|
else:
|
|
raw = display.get("tui_mouse", True)
|
|
if raw is False or raw == 0:
|
|
return False
|
|
if isinstance(raw, str):
|
|
return raw.strip().lower() not in {"0", "false", "no", "off"}
|
|
return True
|
|
|
|
|
|
def _load_reasoning_config() -> dict | None:
|
|
from hermes_constants import parse_reasoning_effort
|
|
|
|
effort = str(
|
|
(_load_cfg().get("agent") or {}).get("reasoning_effort", "") or ""
|
|
).strip()
|
|
return parse_reasoning_effort(effort)
|
|
|
|
|
|
def _load_service_tier() -> str | None:
|
|
raw = (
|
|
str((_load_cfg().get("agent") or {}).get("service_tier", "") or "")
|
|
.strip()
|
|
.lower()
|
|
)
|
|
if not raw or raw in {"normal", "default", "standard", "off", "none"}:
|
|
return None
|
|
if raw in {"fast", "priority", "on"}:
|
|
return "priority"
|
|
return None
|
|
|
|
|
|
def _load_show_reasoning() -> bool:
|
|
return bool((_load_cfg().get("display") or {}).get("show_reasoning", False))
|
|
|
|
|
|
def _load_tool_progress_mode() -> str:
|
|
raw = (_load_cfg().get("display") or {}).get("tool_progress", "all")
|
|
if raw is False:
|
|
return "off"
|
|
if raw is True:
|
|
return "all"
|
|
mode = str(raw or "all").strip().lower()
|
|
return mode if mode in {"off", "new", "all", "verbose"} else "all"
|
|
|
|
|
|
def _load_enabled_toolsets() -> list[str] | None:
|
|
explicit = [
|
|
item.strip()
|
|
for item in os.environ.get("HERMES_TUI_TOOLSETS", "").split(",")
|
|
if item.strip()
|
|
]
|
|
cfg = None
|
|
fallback_notice = None
|
|
|
|
try:
|
|
from toolsets import validate_toolset
|
|
except Exception:
|
|
validate_toolset = None
|
|
|
|
if explicit and validate_toolset is not None:
|
|
built_in = [name for name in explicit if validate_toolset(name)]
|
|
unresolved = [name for name in explicit if name not in built_in]
|
|
|
|
if unresolved:
|
|
try:
|
|
from hermes_cli.plugins import discover_plugins
|
|
|
|
discover_plugins()
|
|
plugin_valid = [name for name in unresolved if validate_toolset(name)]
|
|
except Exception:
|
|
plugin_valid = []
|
|
|
|
if plugin_valid:
|
|
built_in.extend(plugin_valid)
|
|
unresolved = [name for name in unresolved if name not in plugin_valid]
|
|
|
|
if any(name in {"all", "*"} for name in built_in):
|
|
ignored = [name for name in explicit if name not in {"all", "*"}]
|
|
if ignored:
|
|
print(
|
|
"[tui] HERMES_TUI_TOOLSETS=all enables every toolset; "
|
|
f"ignoring additional entries: {', '.join(ignored)}",
|
|
file=sys.stderr,
|
|
flush=True,
|
|
)
|
|
return None
|
|
|
|
if not unresolved:
|
|
return built_in
|
|
|
|
mcp_names: set[str] = set()
|
|
mcp_disabled: set[str] = set()
|
|
try:
|
|
from hermes_cli.config import read_raw_config
|
|
from hermes_cli.tools_config import _parse_enabled_flag
|
|
|
|
raw_cfg = read_raw_config()
|
|
mcp_servers = raw_cfg.get("mcp_servers") if isinstance(raw_cfg.get("mcp_servers"), dict) else {}
|
|
for name, server_cfg in mcp_servers.items():
|
|
if not isinstance(server_cfg, dict):
|
|
continue
|
|
if _parse_enabled_flag(server_cfg.get("enabled", True), default=True):
|
|
mcp_names.add(str(name))
|
|
else:
|
|
mcp_disabled.add(str(name))
|
|
except Exception:
|
|
mcp_names = set()
|
|
mcp_disabled = set()
|
|
|
|
mcp_valid = [name for name in unresolved if name in mcp_names]
|
|
disabled = [name for name in unresolved if name in mcp_disabled]
|
|
unknown = [name for name in unresolved if name not in mcp_names and name not in mcp_disabled]
|
|
valid = built_in + mcp_valid
|
|
|
|
if unknown:
|
|
print(
|
|
f"[tui] ignoring unknown HERMES_TUI_TOOLSETS entries: {', '.join(unknown)}",
|
|
file=sys.stderr,
|
|
flush=True,
|
|
)
|
|
if disabled:
|
|
print(
|
|
"[tui] ignoring disabled MCP servers in HERMES_TUI_TOOLSETS "
|
|
"(set enabled: true in config.yaml to use): "
|
|
f"{', '.join(disabled)}",
|
|
file=sys.stderr,
|
|
flush=True,
|
|
)
|
|
|
|
if valid:
|
|
return valid
|
|
|
|
fallback_notice = "[tui] no valid HERMES_TUI_TOOLSETS entries; using configured CLI toolsets"
|
|
|
|
try:
|
|
from hermes_cli.config import load_config
|
|
from hermes_cli.tools_config import _get_platform_tools
|
|
|
|
cfg = cfg if cfg is not None else load_config()
|
|
|
|
# Runtime toolset resolution must include default MCP servers so the
|
|
# agent can actually call them. Passing ``False`` here is the
|
|
# config-editing variant — used when we need to persist a toolset
|
|
# list without baking in implicit MCP defaults. Using the wrong
|
|
# variant at agent creation time makes MCP tools silently missing
|
|
# from the TUI. See PR #3252 for the original design split.
|
|
enabled = sorted(
|
|
_get_platform_tools(cfg, "cli", include_default_mcp_servers=True)
|
|
)
|
|
if fallback_notice is not None:
|
|
print(fallback_notice, file=sys.stderr, flush=True)
|
|
return enabled or None
|
|
except Exception:
|
|
if fallback_notice is not None:
|
|
print(
|
|
"[tui] no valid HERMES_TUI_TOOLSETS entries and configured CLI toolsets could not be loaded; enabling all toolsets",
|
|
file=sys.stderr,
|
|
flush=True,
|
|
)
|
|
return None
|
|
|
|
|
|
def _session_tool_progress_mode(sid: str) -> str:
|
|
return str(_sessions.get(sid, {}).get("tool_progress_mode", "all") or "all")
|
|
|
|
|
|
def _tool_progress_enabled(sid: str) -> bool:
|
|
return _session_tool_progress_mode(sid) != "off"
|
|
|
|
|
|
def _restart_slash_worker(session: dict):
|
|
worker = session.get("slash_worker")
|
|
if worker:
|
|
try:
|
|
worker.close()
|
|
except Exception:
|
|
pass
|
|
try:
|
|
session["slash_worker"] = _SlashWorker(
|
|
session["session_key"],
|
|
getattr(session.get("agent"), "model", _resolve_model()),
|
|
)
|
|
except Exception:
|
|
session["slash_worker"] = None
|
|
|
|
|
|
def _persist_model_switch(result) -> None:
|
|
from hermes_cli.config import save_config
|
|
|
|
cfg = _load_cfg()
|
|
model_cfg = cfg.get("model")
|
|
if not isinstance(model_cfg, dict):
|
|
model_cfg = {}
|
|
cfg["model"] = model_cfg
|
|
|
|
model_cfg["default"] = result.new_model
|
|
model_cfg["provider"] = result.target_provider
|
|
if result.base_url:
|
|
model_cfg["base_url"] = result.base_url
|
|
else:
|
|
model_cfg.pop("base_url", None)
|
|
save_config(cfg)
|
|
|
|
|
|
def _apply_model_switch(sid: str, session: dict, raw_input: str) -> dict:
|
|
from hermes_cli.model_switch import parse_model_flags, switch_model
|
|
from hermes_cli.runtime_provider import resolve_runtime_provider
|
|
|
|
model_input, explicit_provider, persist_global = parse_model_flags(raw_input)
|
|
if not model_input:
|
|
raise ValueError("model value required")
|
|
|
|
agent = session.get("agent")
|
|
if agent:
|
|
current_provider = getattr(agent, "provider", "") or ""
|
|
current_model = getattr(agent, "model", "") or ""
|
|
current_base_url = getattr(agent, "base_url", "") or ""
|
|
current_api_key = getattr(agent, "api_key", "") or ""
|
|
else:
|
|
runtime = resolve_runtime_provider(requested=None)
|
|
current_provider = str(runtime.get("provider", "") or "")
|
|
current_model = _resolve_model()
|
|
current_base_url = str(runtime.get("base_url", "") or "")
|
|
current_api_key = str(runtime.get("api_key", "") or "")
|
|
|
|
# Load user-defined providers so switch_model can resolve named custom
|
|
# endpoints (e.g. "ollama-launch") and validate against saved model lists.
|
|
user_provs = None
|
|
custom_provs = None
|
|
try:
|
|
from hermes_cli.config import get_compatible_custom_providers, load_config
|
|
|
|
cfg = load_config()
|
|
user_provs = [
|
|
{"provider": k, **v} for k, v in (cfg.get("providers") or {}).items()
|
|
]
|
|
custom_provs = get_compatible_custom_providers(cfg)
|
|
except Exception:
|
|
pass
|
|
|
|
result = switch_model(
|
|
raw_input=model_input,
|
|
current_provider=current_provider,
|
|
current_model=current_model,
|
|
current_base_url=current_base_url,
|
|
current_api_key=current_api_key,
|
|
is_global=persist_global,
|
|
explicit_provider=explicit_provider,
|
|
user_providers=user_provs,
|
|
custom_providers=custom_provs,
|
|
)
|
|
if not result.success:
|
|
raise ValueError(result.error_message or "model switch failed")
|
|
|
|
if agent:
|
|
agent.switch_model(
|
|
new_model=result.new_model,
|
|
new_provider=result.target_provider,
|
|
api_key=result.api_key,
|
|
base_url=result.base_url,
|
|
api_mode=result.api_mode,
|
|
)
|
|
_restart_slash_worker(session)
|
|
_emit("session.info", sid, _session_info(agent))
|
|
|
|
os.environ["HERMES_MODEL"] = result.new_model
|
|
os.environ["HERMES_INFERENCE_MODEL"] = result.new_model
|
|
# Keep the process-level provider env vars in sync with the user's
|
|
# explicit choice so any ambient re-resolution (credential pool refresh,
|
|
# compressor rebuild, aux clients) and startup re-resolution on /new
|
|
# both pick up the new provider instead of the original one persisted
|
|
# in config or env.
|
|
#
|
|
# HERMES_TUI_PROVIDER is the canonical "explicit-this-process" carrier
|
|
# consumed by _resolve_startup_runtime() — set it unconditionally on
|
|
# /model so /new can't fall through to static-catalog detection and
|
|
# pick a coincidentally-matching native provider (fixes #16857).
|
|
if result.target_provider:
|
|
os.environ["HERMES_INFERENCE_PROVIDER"] = result.target_provider
|
|
os.environ["HERMES_TUI_PROVIDER"] = result.target_provider
|
|
if persist_global:
|
|
_persist_model_switch(result)
|
|
return {"value": result.new_model, "warning": result.warning_message or ""}
|
|
|
|
|
|
def _compress_session_history(
|
|
session: dict,
|
|
focus_topic: str | None = None,
|
|
approx_tokens: int | None = None,
|
|
before_messages: list | None = None,
|
|
history_version: int | None = None,
|
|
) -> tuple[int, dict]:
|
|
from agent.model_metadata import estimate_messages_tokens_rough
|
|
|
|
agent = session["agent"]
|
|
# Snapshot history under the lock so the LLM-bound compression call
|
|
# below does NOT hold history_lock for the duration of the request —
|
|
# otherwise other handlers acquiring the lock (prompt.submit etc.)
|
|
# block on the dispatcher loop while compaction runs.
|
|
if before_messages is None or history_version is None:
|
|
with session["history_lock"]:
|
|
before_messages = list(session.get("history", []))
|
|
history_version = int(session.get("history_version", 0))
|
|
history = before_messages
|
|
if len(history) < 4:
|
|
usage = _get_usage(agent)
|
|
return 0, usage
|
|
if approx_tokens is None:
|
|
approx_tokens = estimate_messages_tokens_rough(history)
|
|
# Pass system_message=None so AIAgent._compress_context rebuilds the
|
|
# system prompt cleanly via _build_system_prompt(None). Passing the
|
|
# cached prompt (which already contains the agent identity block)
|
|
# makes the rebuild append the identity a second time. Mirrors the
|
|
# CLI's _manual_compress fix for issue #15281.
|
|
compressed, _ = agent._compress_context(
|
|
history,
|
|
None,
|
|
approx_tokens=approx_tokens,
|
|
focus_topic=focus_topic or None,
|
|
)
|
|
with session["history_lock"]:
|
|
if int(session.get("history_version", 0)) != history_version:
|
|
# External mutation during compaction — drop the compressed
|
|
# result so we don't clobber concurrent edits.
|
|
usage = _get_usage(agent)
|
|
return 0, usage
|
|
session["history"] = compressed
|
|
session["history_version"] = history_version + 1
|
|
usage = _get_usage(agent)
|
|
return len(history) - len(compressed), usage
|
|
|
|
|
|
def _sync_session_key_after_compress(sid: str, session: dict) -> None:
|
|
"""Re-anchor session_key when AIAgent._compress_context rotates session_id.
|
|
|
|
AIAgent._compress_context ends the current SessionDB session and creates
|
|
a new continuation session, rotating ``agent.session_id``. The TUI
|
|
gateway keeps the gateway-side ``session_key`` separate (used for
|
|
approval routing, slash worker init, DB title/history lookups, yolo
|
|
state). Without this sync, those operations would target the ended
|
|
parent session while the agent writes to the new continuation session.
|
|
Mirrors HermesCLI._manual_compress's session_id sync.
|
|
"""
|
|
agent = session.get("agent")
|
|
new_session_id = getattr(agent, "session_id", None) or ""
|
|
old_key = session.get("session_key", "") or ""
|
|
if not new_session_id or new_session_id == old_key:
|
|
return
|
|
|
|
try:
|
|
from tools.approval import (
|
|
disable_session_yolo,
|
|
enable_session_yolo,
|
|
is_session_yolo_enabled,
|
|
register_gateway_notify,
|
|
unregister_gateway_notify,
|
|
)
|
|
|
|
try:
|
|
unregister_gateway_notify(old_key)
|
|
except Exception:
|
|
pass
|
|
session["session_key"] = new_session_id
|
|
try:
|
|
yolo_was_on = is_session_yolo_enabled(old_key)
|
|
except Exception:
|
|
yolo_was_on = False
|
|
if yolo_was_on:
|
|
try:
|
|
enable_session_yolo(new_session_id)
|
|
disable_session_yolo(old_key)
|
|
except Exception:
|
|
pass
|
|
try:
|
|
register_gateway_notify(
|
|
new_session_id,
|
|
lambda data: _emit("approval.request", sid, data),
|
|
)
|
|
except Exception:
|
|
pass
|
|
except Exception:
|
|
# Even if the approval module fails to import, still anchor the
|
|
# session_key on the new continuation id so downstream lookups
|
|
# don't keep targeting the ended row.
|
|
session["session_key"] = new_session_id
|
|
|
|
session["pending_title"] = None
|
|
try:
|
|
_restart_slash_worker(session)
|
|
except Exception:
|
|
pass
|
|
|
|
|
|
def _get_usage(agent) -> dict:
|
|
g = lambda k, fb=None: getattr(agent, k, 0) or (getattr(agent, fb, 0) if fb else 0)
|
|
usage = {
|
|
"model": getattr(agent, "model", "") or "",
|
|
"input": g("session_input_tokens", "session_prompt_tokens"),
|
|
"output": g("session_output_tokens", "session_completion_tokens"),
|
|
"cache_read": g("session_cache_read_tokens"),
|
|
"cache_write": g("session_cache_write_tokens"),
|
|
"prompt": g("session_prompt_tokens"),
|
|
"completion": g("session_completion_tokens"),
|
|
"total": g("session_total_tokens"),
|
|
"calls": g("session_api_calls"),
|
|
}
|
|
comp = getattr(agent, "context_compressor", None)
|
|
if comp:
|
|
ctx_used = getattr(comp, "last_prompt_tokens", 0) or usage["total"] or 0
|
|
ctx_max = getattr(comp, "context_length", 0) or 0
|
|
if ctx_max:
|
|
usage["context_used"] = ctx_used
|
|
usage["context_max"] = ctx_max
|
|
usage["context_percent"] = max(0, min(100, round(ctx_used / ctx_max * 100)))
|
|
usage["compressions"] = getattr(comp, "compression_count", 0) or 0
|
|
try:
|
|
from agent.usage_pricing import CanonicalUsage, estimate_usage_cost
|
|
|
|
cost = estimate_usage_cost(
|
|
usage["model"],
|
|
CanonicalUsage(
|
|
input_tokens=usage["input"],
|
|
output_tokens=usage["output"],
|
|
cache_read_tokens=usage["cache_read"],
|
|
cache_write_tokens=usage["cache_write"],
|
|
),
|
|
provider=getattr(agent, "provider", None),
|
|
base_url=getattr(agent, "base_url", None),
|
|
)
|
|
usage["cost_status"] = cost.status
|
|
if cost.amount_usd is not None:
|
|
usage["cost_usd"] = float(cost.amount_usd)
|
|
except Exception:
|
|
pass
|
|
return usage
|
|
|
|
|
|
def _probe_credentials(agent) -> str:
|
|
"""Light credential check at session creation — returns warning or ''."""
|
|
try:
|
|
key = getattr(agent, "api_key", "") or ""
|
|
provider = getattr(agent, "provider", "") or ""
|
|
if not key or key == "no-key-required":
|
|
return f"No API key configured for provider '{provider}'. First message will fail."
|
|
except Exception:
|
|
pass
|
|
return ""
|
|
|
|
|
|
def _probe_config_health(cfg: dict) -> str:
|
|
"""Flag bare YAML keys (`agent:` with no value → None) that silently
|
|
drop nested settings. Returns warning or ''."""
|
|
if not isinstance(cfg, dict):
|
|
return ""
|
|
warnings: list[str] = []
|
|
null_keys = sorted(k for k, v in cfg.items() if v is None)
|
|
if not null_keys:
|
|
pass
|
|
else:
|
|
keys = ", ".join(f"`{k}`" for k in null_keys)
|
|
warnings.append(
|
|
f"config.yaml has empty section(s): {keys}. "
|
|
f"Remove the line(s) or set them to `{{}}` — "
|
|
f"empty sections silently drop nested settings."
|
|
)
|
|
display_cfg = cfg.get("display")
|
|
agent_cfg = cfg.get("agent")
|
|
if isinstance(display_cfg, dict):
|
|
personality = str(display_cfg.get("personality", "") or "").strip().lower()
|
|
if (
|
|
personality
|
|
and personality not in {"default", "none", "neutral"}
|
|
and isinstance(agent_cfg, dict)
|
|
and agent_cfg.get("personalities") is None
|
|
):
|
|
warnings.append(
|
|
"`display.personality` is set but `agent.personalities` is empty/null; "
|
|
"personality overlay will be skipped."
|
|
)
|
|
return " ".join(warnings).strip()
|
|
|
|
|
|
def _session_info(agent) -> dict:
|
|
reasoning_config = getattr(agent, "reasoning_config", None)
|
|
reasoning_effort = ""
|
|
if (
|
|
isinstance(reasoning_config, dict)
|
|
and reasoning_config.get("enabled") is not False
|
|
):
|
|
reasoning_effort = str(reasoning_config.get("effort", "") or "")
|
|
service_tier = getattr(agent, "service_tier", None) or ""
|
|
info: dict = {
|
|
"model": getattr(agent, "model", ""),
|
|
"reasoning_effort": reasoning_effort,
|
|
"service_tier": service_tier,
|
|
"fast": service_tier == "priority",
|
|
"tools": {},
|
|
"skills": {},
|
|
"cwd": os.getcwd(),
|
|
"version": "",
|
|
"release_date": "",
|
|
"update_behind": None,
|
|
"update_command": "",
|
|
"usage": _get_usage(agent),
|
|
}
|
|
try:
|
|
from hermes_cli import __version__, __release_date__
|
|
|
|
info["version"] = __version__
|
|
info["release_date"] = __release_date__
|
|
except Exception:
|
|
pass
|
|
try:
|
|
from model_tools import get_toolset_for_tool
|
|
|
|
for t in getattr(agent, "tools", []) or []:
|
|
name = t["function"]["name"]
|
|
info["tools"].setdefault(get_toolset_for_tool(name) or "other", []).append(
|
|
name
|
|
)
|
|
except Exception:
|
|
pass
|
|
try:
|
|
from hermes_cli.banner import get_available_skills
|
|
|
|
info["skills"] = get_available_skills()
|
|
except Exception:
|
|
pass
|
|
try:
|
|
from tools.mcp_tool import get_mcp_status
|
|
|
|
info["mcp_servers"] = get_mcp_status()
|
|
except Exception:
|
|
info["mcp_servers"] = []
|
|
try:
|
|
from hermes_cli.banner import get_update_result
|
|
from hermes_cli.config import recommended_update_command
|
|
|
|
info["update_behind"] = get_update_result(timeout=0.5)
|
|
info["update_command"] = recommended_update_command()
|
|
except Exception:
|
|
pass
|
|
return info
|
|
|
|
|
|
def _tool_ctx(name: str, args: dict) -> str:
|
|
try:
|
|
from agent.display import build_tool_preview
|
|
|
|
return build_tool_preview(name, args, max_len=80) or ""
|
|
except Exception:
|
|
return ""
|
|
|
|
|
|
def _fmt_tool_duration(seconds: float | None) -> str:
|
|
if seconds is None:
|
|
return ""
|
|
if seconds < 10:
|
|
return f"{seconds:.1f}s"
|
|
if seconds < 60:
|
|
return f"{round(seconds)}s"
|
|
mins, secs = divmod(int(round(seconds)), 60)
|
|
return f"{mins}m {secs}s" if secs else f"{mins}m"
|
|
|
|
|
|
def _count_list(obj: object, *path: str) -> int | None:
|
|
cur = obj
|
|
for key in path:
|
|
if not isinstance(cur, dict):
|
|
return None
|
|
cur = cur.get(key)
|
|
return len(cur) if isinstance(cur, list) else None
|
|
|
|
|
|
def _tool_summary(name: str, result: str, duration_s: float | None) -> str | None:
|
|
try:
|
|
data = json.loads(result)
|
|
except Exception:
|
|
data = None
|
|
|
|
dur = _fmt_tool_duration(duration_s)
|
|
suffix = f" in {dur}" if dur else ""
|
|
text = None
|
|
|
|
if name == "web_search" and isinstance(data, dict):
|
|
n = _count_list(data, "data", "web")
|
|
if n is not None:
|
|
text = f"Did {n} {'search' if n == 1 else 'searches'}"
|
|
|
|
elif name == "web_extract" and isinstance(data, dict):
|
|
n = _count_list(data, "results") or _count_list(data, "data", "results")
|
|
if n is not None:
|
|
text = f"Extracted {n} {'page' if n == 1 else 'pages'}"
|
|
|
|
return f"{text}{suffix}" if text else None
|
|
|
|
|
|
def _on_tool_start(sid: str, tool_call_id: str, name: str, args: dict):
|
|
session = _sessions.get(sid)
|
|
if session is not None:
|
|
try:
|
|
from agent.display import capture_local_edit_snapshot
|
|
|
|
snapshot = capture_local_edit_snapshot(name, args)
|
|
if snapshot is not None:
|
|
session.setdefault("edit_snapshots", {})[tool_call_id] = snapshot
|
|
except Exception:
|
|
pass
|
|
session.setdefault("tool_started_at", {})[tool_call_id] = time.time()
|
|
if _tool_progress_enabled(sid):
|
|
# tool.complete is the source of truth for todos (full list from the
|
|
# tool result). args.todos here may be a partial merge update.
|
|
_emit(
|
|
"tool.start",
|
|
sid,
|
|
{"tool_id": tool_call_id, "name": name, "context": _tool_ctx(name, args)},
|
|
)
|
|
|
|
|
|
def _on_tool_complete(sid: str, tool_call_id: str, name: str, args: dict, result: str):
|
|
payload = {"tool_id": tool_call_id, "name": name}
|
|
session = _sessions.get(sid)
|
|
snapshot = None
|
|
started_at = None
|
|
if session is not None:
|
|
snapshot = session.setdefault("edit_snapshots", {}).pop(tool_call_id, None)
|
|
started_at = session.setdefault("tool_started_at", {}).pop(tool_call_id, None)
|
|
duration_s = time.time() - started_at if started_at else None
|
|
if duration_s is not None:
|
|
payload["duration_s"] = duration_s
|
|
summary = _tool_summary(name, result, duration_s)
|
|
if summary:
|
|
payload["summary"] = summary
|
|
if name == "todo":
|
|
try:
|
|
data = json.loads(result)
|
|
if isinstance(data, dict) and isinstance(data.get("todos"), list):
|
|
payload["todos"] = data.get("todos")
|
|
except Exception:
|
|
pass
|
|
try:
|
|
from agent.display import render_edit_diff_with_delta
|
|
|
|
rendered: list[str] = []
|
|
if render_edit_diff_with_delta(
|
|
name,
|
|
result,
|
|
function_args=args,
|
|
snapshot=snapshot,
|
|
print_fn=rendered.append,
|
|
):
|
|
payload["inline_diff"] = "\n".join(rendered)
|
|
except Exception:
|
|
pass
|
|
if _tool_progress_enabled(sid) or payload.get("inline_diff"):
|
|
_emit("tool.complete", sid, payload)
|
|
|
|
|
|
def _on_tool_progress(
|
|
sid: str,
|
|
event_type: str,
|
|
name: str | None = None,
|
|
preview: str | None = None,
|
|
_args: dict | None = None,
|
|
**_kwargs,
|
|
):
|
|
if not _tool_progress_enabled(sid):
|
|
return
|
|
if event_type == "tool.started" and name:
|
|
_emit("tool.progress", sid, {"name": name, "preview": preview or ""})
|
|
return
|
|
if event_type == "reasoning.available" and preview:
|
|
_emit("reasoning.available", sid, {"text": str(preview)})
|
|
return
|
|
if event_type.startswith("subagent."):
|
|
payload = {
|
|
"goal": str(_kwargs.get("goal") or ""),
|
|
"task_count": int(_kwargs.get("task_count") or 1),
|
|
"task_index": int(_kwargs.get("task_index") or 0),
|
|
}
|
|
# Identity fields for the TUI spawn tree. All optional — older
|
|
# emitters that omit them fall back to flat rendering client-side.
|
|
if _kwargs.get("subagent_id"):
|
|
payload["subagent_id"] = str(_kwargs["subagent_id"])
|
|
if _kwargs.get("parent_id"):
|
|
payload["parent_id"] = str(_kwargs["parent_id"])
|
|
if _kwargs.get("depth") is not None:
|
|
payload["depth"] = int(_kwargs["depth"])
|
|
if _kwargs.get("model"):
|
|
payload["model"] = str(_kwargs["model"])
|
|
if _kwargs.get("tool_count") is not None:
|
|
payload["tool_count"] = int(_kwargs["tool_count"])
|
|
if _kwargs.get("toolsets"):
|
|
payload["toolsets"] = [str(t) for t in _kwargs["toolsets"]]
|
|
# Per-branch rollups emitted on subagent.complete (features 1+2+4).
|
|
for int_key in (
|
|
"input_tokens",
|
|
"output_tokens",
|
|
"reasoning_tokens",
|
|
"api_calls",
|
|
):
|
|
val = _kwargs.get(int_key)
|
|
if val is not None:
|
|
try:
|
|
payload[int_key] = int(val)
|
|
except (TypeError, ValueError):
|
|
pass
|
|
if _kwargs.get("cost_usd") is not None:
|
|
try:
|
|
payload["cost_usd"] = float(_kwargs["cost_usd"])
|
|
except (TypeError, ValueError):
|
|
pass
|
|
if _kwargs.get("files_read"):
|
|
payload["files_read"] = [str(p) for p in _kwargs["files_read"]]
|
|
if _kwargs.get("files_written"):
|
|
payload["files_written"] = [str(p) for p in _kwargs["files_written"]]
|
|
if _kwargs.get("output_tail"):
|
|
payload["output_tail"] = list(_kwargs["output_tail"]) # list of dicts
|
|
if name:
|
|
payload["tool_name"] = str(name)
|
|
if preview:
|
|
payload["text"] = str(preview)
|
|
if _kwargs.get("status"):
|
|
payload["status"] = str(_kwargs["status"])
|
|
if _kwargs.get("summary"):
|
|
payload["summary"] = str(_kwargs["summary"])
|
|
if _kwargs.get("duration_seconds") is not None:
|
|
payload["duration_seconds"] = float(_kwargs["duration_seconds"])
|
|
if preview and event_type == "subagent.tool":
|
|
payload["tool_preview"] = str(preview)
|
|
payload["text"] = str(preview)
|
|
_emit(event_type, sid, payload)
|
|
|
|
|
|
def _agent_cbs(sid: str) -> dict:
|
|
return dict(
|
|
tool_start_callback=lambda tc_id, name, args: _on_tool_start(
|
|
sid, tc_id, name, args
|
|
),
|
|
tool_complete_callback=lambda tc_id, name, args, result: _on_tool_complete(
|
|
sid, tc_id, name, args, result
|
|
),
|
|
tool_progress_callback=lambda event_type, name=None, preview=None, args=None, **kwargs: _on_tool_progress(
|
|
sid, event_type, name, preview, args, **kwargs
|
|
),
|
|
tool_gen_callback=lambda name: _tool_progress_enabled(sid)
|
|
and _emit("tool.generating", sid, {"name": name}),
|
|
thinking_callback=lambda text: _emit("thinking.delta", sid, {"text": text}),
|
|
reasoning_callback=lambda text: _emit("reasoning.delta", sid, {"text": text}),
|
|
status_callback=lambda kind, text=None: _status_update(
|
|
sid, str(kind), None if text is None else str(text)
|
|
),
|
|
clarify_callback=lambda q, c: _block(
|
|
"clarify.request", sid, {"question": q, "choices": c}
|
|
),
|
|
)
|
|
|
|
|
|
def _wire_callbacks(sid: str):
|
|
from tools.terminal_tool import set_sudo_password_callback
|
|
from tools.skills_tool import set_secret_capture_callback
|
|
|
|
set_sudo_password_callback(lambda: _block("sudo.request", sid, {}, timeout=120))
|
|
|
|
def secret_cb(env_var, prompt, metadata=None):
|
|
pl = {"prompt": prompt, "env_var": env_var}
|
|
if metadata:
|
|
pl["metadata"] = metadata
|
|
val = _block("secret.request", sid, pl)
|
|
if not val:
|
|
return {
|
|
"success": True,
|
|
"stored_as": env_var,
|
|
"validated": False,
|
|
"skipped": True,
|
|
"message": "skipped",
|
|
}
|
|
from hermes_cli.config import save_env_value_secure
|
|
|
|
return {
|
|
**save_env_value_secure(env_var, val),
|
|
"skipped": False,
|
|
"message": "ok",
|
|
}
|
|
|
|
set_secret_capture_callback(secret_cb)
|
|
|
|
|
|
def _render_personality_prompt(value) -> str:
|
|
if isinstance(value, dict):
|
|
parts = [value.get("system_prompt", "")]
|
|
if value.get("tone"):
|
|
parts.append(f'Tone: {value["tone"]}')
|
|
if value.get("style"):
|
|
parts.append(f'Style: {value["style"]}')
|
|
return "\n".join(p for p in parts if p)
|
|
return str(value)
|
|
|
|
|
|
def _available_personalities(cfg: dict | None = None) -> dict:
|
|
try:
|
|
from cli import load_cli_config
|
|
|
|
return (load_cli_config().get("agent") or {}).get("personalities", {}) or {}
|
|
except Exception:
|
|
try:
|
|
from hermes_cli.config import load_config as _load_full_cfg
|
|
|
|
return (_load_full_cfg().get("agent") or {}).get("personalities", {}) or {}
|
|
except Exception:
|
|
cfg = cfg or _load_cfg()
|
|
return (cfg.get("agent") or {}).get("personalities", {}) or {}
|
|
|
|
|
|
def _validate_personality(value: str, cfg: dict | None = None) -> tuple[str, str]:
|
|
raw = str(value or "").strip()
|
|
name = raw.lower()
|
|
if not name or name in ("none", "default", "neutral"):
|
|
return "", ""
|
|
|
|
personalities = _available_personalities(cfg)
|
|
if name not in personalities:
|
|
names = sorted(personalities)
|
|
available = ", ".join(f"`{n}`" for n in names)
|
|
base = f"Unknown personality: `{raw}`."
|
|
if available:
|
|
base += f"\n\nAvailable: `none`, {available}"
|
|
else:
|
|
base += "\n\nNo personalities configured."
|
|
raise ValueError(base)
|
|
|
|
return name, _render_personality_prompt(personalities[name])
|
|
|
|
|
|
def _apply_personality_to_session(
|
|
sid: str, session: dict, new_prompt: str
|
|
) -> tuple[bool, dict | None]:
|
|
if not session:
|
|
return False, None
|
|
|
|
try:
|
|
info = _reset_session_agent(sid, session)
|
|
return True, info
|
|
except Exception:
|
|
if session.get("agent"):
|
|
agent = session["agent"]
|
|
agent.ephemeral_system_prompt = new_prompt or None
|
|
agent._cached_system_prompt = None
|
|
info = _session_info(agent)
|
|
_emit("session.info", sid, info)
|
|
return False, info
|
|
return False, None
|
|
|
|
|
|
def _background_agent_kwargs(agent, task_id: str) -> dict:
|
|
cfg = _load_cfg()
|
|
|
|
return {
|
|
"base_url": getattr(agent, "base_url", None) or None,
|
|
"api_key": getattr(agent, "api_key", None) or None,
|
|
"provider": getattr(agent, "provider", None) or None,
|
|
"api_mode": getattr(agent, "api_mode", None) or None,
|
|
"acp_command": getattr(agent, "acp_command", None) or None,
|
|
"acp_args": getattr(agent, "acp_args", None) or None,
|
|
"model": getattr(agent, "model", None) or _resolve_model(),
|
|
"max_iterations": int(cfg.get("max_turns", 25) or 25),
|
|
"enabled_toolsets": getattr(agent, "enabled_toolsets", None)
|
|
or _load_enabled_toolsets(),
|
|
"quiet_mode": True,
|
|
"verbose_logging": False,
|
|
"ephemeral_system_prompt": getattr(agent, "ephemeral_system_prompt", None)
|
|
or None,
|
|
"providers_allowed": getattr(agent, "providers_allowed", None),
|
|
"providers_ignored": getattr(agent, "providers_ignored", None),
|
|
"providers_order": getattr(agent, "providers_order", None),
|
|
"provider_sort": getattr(agent, "provider_sort", None),
|
|
"provider_require_parameters": getattr(
|
|
agent, "provider_require_parameters", False
|
|
),
|
|
"provider_data_collection": getattr(agent, "provider_data_collection", None),
|
|
"session_id": task_id,
|
|
"reasoning_config": getattr(agent, "reasoning_config", None)
|
|
or _load_reasoning_config(),
|
|
"service_tier": getattr(agent, "service_tier", None) or _load_service_tier(),
|
|
"request_overrides": dict(getattr(agent, "request_overrides", {}) or {}),
|
|
"platform": "tui",
|
|
"session_db": _get_db(),
|
|
"fallback_model": getattr(agent, "_fallback_model", None),
|
|
}
|
|
|
|
|
|
def _reset_session_agent(sid: str, session: dict) -> dict:
|
|
tokens = _set_session_context(session["session_key"])
|
|
try:
|
|
new_agent = _make_agent(
|
|
sid, session["session_key"], session_id=session["session_key"]
|
|
)
|
|
finally:
|
|
_clear_session_context(tokens)
|
|
session["agent"] = new_agent
|
|
session["attached_images"] = []
|
|
session["edit_snapshots"] = {}
|
|
session["image_counter"] = 0
|
|
session["running"] = False
|
|
session["show_reasoning"] = _load_show_reasoning()
|
|
session["tool_progress_mode"] = _load_tool_progress_mode()
|
|
session["tool_started_at"] = {}
|
|
with session["history_lock"]:
|
|
session["history"] = []
|
|
session["history_version"] = int(session.get("history_version", 0)) + 1
|
|
info = _session_info(new_agent)
|
|
_emit("session.info", sid, info)
|
|
_restart_slash_worker(session)
|
|
return info
|
|
|
|
|
|
def _make_agent(sid: str, key: str, session_id: str | None = None):
|
|
from run_agent import AIAgent
|
|
from hermes_cli.runtime_provider import resolve_runtime_provider
|
|
|
|
cfg = _load_cfg()
|
|
system_prompt = ((cfg.get("agent") or {}).get("system_prompt", "") or "").strip()
|
|
model, requested_provider = _resolve_startup_runtime()
|
|
runtime = resolve_runtime_provider(
|
|
requested=requested_provider,
|
|
target_model=model or None,
|
|
)
|
|
return AIAgent(
|
|
model=model,
|
|
provider=runtime.get("provider"),
|
|
base_url=runtime.get("base_url"),
|
|
api_key=runtime.get("api_key"),
|
|
api_mode=runtime.get("api_mode"),
|
|
acp_command=runtime.get("command"),
|
|
acp_args=runtime.get("args"),
|
|
credential_pool=runtime.get("credential_pool"),
|
|
quiet_mode=True,
|
|
verbose_logging=_load_tool_progress_mode() == "verbose",
|
|
reasoning_config=_load_reasoning_config(),
|
|
service_tier=_load_service_tier(),
|
|
enabled_toolsets=_load_enabled_toolsets(),
|
|
platform="tui",
|
|
session_id=session_id or key,
|
|
session_db=_get_db(),
|
|
ephemeral_system_prompt=system_prompt or None,
|
|
**_agent_cbs(sid),
|
|
)
|
|
|
|
|
|
def _init_session(sid: str, key: str, agent, history: list, cols: int = 80):
|
|
_sessions[sid] = {
|
|
"agent": agent,
|
|
"session_key": key,
|
|
"history": history,
|
|
"history_lock": threading.Lock(),
|
|
"history_version": 0,
|
|
"running": False,
|
|
"attached_images": [],
|
|
"image_counter": 0,
|
|
"cols": cols,
|
|
"slash_worker": None,
|
|
"show_reasoning": _load_show_reasoning(),
|
|
"tool_progress_mode": _load_tool_progress_mode(),
|
|
"edit_snapshots": {},
|
|
"tool_started_at": {},
|
|
# Pin async event emissions to whichever transport created the
|
|
# session (stdio for Ink, JSON-RPC WS for the dashboard sidebar).
|
|
"transport": current_transport() or _stdio_transport,
|
|
}
|
|
try:
|
|
_sessions[sid]["slash_worker"] = _SlashWorker(
|
|
key, getattr(agent, "model", _resolve_model())
|
|
)
|
|
except Exception:
|
|
# Defer hard-failure to slash.exec; chat still works without slash worker.
|
|
_sessions[sid]["slash_worker"] = None
|
|
try:
|
|
from tools.approval import register_gateway_notify, load_permanent_allowlist
|
|
|
|
register_gateway_notify(key, lambda data: _emit("approval.request", sid, data))
|
|
load_permanent_allowlist()
|
|
except Exception:
|
|
pass
|
|
_wire_callbacks(sid)
|
|
_notify_session_boundary("on_session_reset", key)
|
|
_emit("session.info", sid, _session_info(agent))
|
|
|
|
|
|
def _new_session_key() -> str:
|
|
return f"{datetime.now().strftime('%Y%m%d_%H%M%S')}_{uuid.uuid4().hex[:6]}"
|
|
|
|
|
|
def _with_checkpoints(session, fn):
|
|
return fn(session["agent"]._checkpoint_mgr, os.getenv("TERMINAL_CWD", os.getcwd()))
|
|
|
|
|
|
def _resolve_checkpoint_hash(mgr, cwd: str, ref: str) -> str:
|
|
try:
|
|
checkpoints = mgr.list_checkpoints(cwd)
|
|
idx = int(ref) - 1
|
|
except ValueError:
|
|
return ref
|
|
if 0 <= idx < len(checkpoints):
|
|
return checkpoints[idx].get("hash", ref)
|
|
raise ValueError(f"Invalid checkpoint number. Use 1-{len(checkpoints)}.")
|
|
|
|
|
|
def _enrich_with_attached_images(user_text: str, image_paths: list[str]) -> str:
|
|
"""Pre-analyze attached images via vision and prepend descriptions to user text."""
|
|
import asyncio, json as _json
|
|
from tools.vision_tools import vision_analyze_tool
|
|
|
|
prompt = (
|
|
"Describe everything visible in this image in thorough detail. "
|
|
"Include any text, code, data, objects, people, layout, colors, "
|
|
"and any other notable visual information."
|
|
)
|
|
|
|
parts: list[str] = []
|
|
for path in image_paths:
|
|
p = Path(path)
|
|
if not p.exists():
|
|
continue
|
|
hint = f"[You can examine it with vision_analyze using image_url: {p}]"
|
|
try:
|
|
r = _json.loads(
|
|
asyncio.run(vision_analyze_tool(image_url=str(p), user_prompt=prompt))
|
|
)
|
|
desc = r.get("analysis", "") if r.get("success") else None
|
|
parts.append(
|
|
f"[The user attached an image:\n{desc}]\n{hint}"
|
|
if desc
|
|
else f"[The user attached an image but analysis failed.]\n{hint}"
|
|
)
|
|
except Exception:
|
|
parts.append(f"[The user attached an image but analysis failed.]\n{hint}")
|
|
|
|
text = user_text or ""
|
|
prefix = "\n\n".join(parts)
|
|
if prefix:
|
|
return f"{prefix}\n\n{text}" if text else prefix
|
|
return text or "What do you see in this image?"
|
|
|
|
|
|
def _history_to_messages(history: list[dict]) -> list[dict]:
|
|
messages = []
|
|
tool_call_args = {}
|
|
|
|
for m in history:
|
|
if not isinstance(m, dict):
|
|
continue
|
|
role = m.get("role")
|
|
if role not in ("user", "assistant", "tool", "system"):
|
|
continue
|
|
if role == "assistant" and m.get("tool_calls"):
|
|
for tc in m["tool_calls"]:
|
|
fn = tc.get("function", {})
|
|
tc_id = tc.get("id", "")
|
|
if tc_id and fn.get("name"):
|
|
try:
|
|
args = json.loads(fn.get("arguments", "{}"))
|
|
except (json.JSONDecodeError, TypeError):
|
|
args = {}
|
|
tool_call_args[tc_id] = (fn["name"], args)
|
|
if not (m.get("content") or "").strip():
|
|
continue
|
|
if role == "tool":
|
|
tc_id = m.get("tool_call_id", "")
|
|
tc_info = tool_call_args.get(tc_id) if tc_id else None
|
|
name = (tc_info[0] if tc_info else None) or m.get("tool_name") or "tool"
|
|
args = (tc_info[1] if tc_info else None) or {}
|
|
messages.append(
|
|
{"role": "tool", "name": name, "context": _tool_ctx(name, args)}
|
|
)
|
|
continue
|
|
if not (m.get("content") or "").strip():
|
|
continue
|
|
messages.append({"role": role, "text": m.get("content") or ""})
|
|
|
|
return messages
|
|
|
|
|
|
# ── Methods: session ─────────────────────────────────────────────────
|
|
|
|
|
|
@method("session.create")
|
|
def _(rid, params: dict) -> dict:
|
|
sid = uuid.uuid4().hex[:8]
|
|
key = _new_session_key()
|
|
cols = int(params.get("cols", 80))
|
|
_enable_gateway_prompts()
|
|
|
|
ready = threading.Event()
|
|
|
|
_sessions[sid] = {
|
|
"agent": None,
|
|
"agent_error": None,
|
|
"agent_ready": ready,
|
|
"attached_images": [],
|
|
"cols": cols,
|
|
"edit_snapshots": {},
|
|
"history": [],
|
|
"history_lock": threading.Lock(),
|
|
"history_version": 0,
|
|
"image_counter": 0,
|
|
"pending_title": None,
|
|
"running": False,
|
|
"session_key": key,
|
|
"show_reasoning": _load_show_reasoning(),
|
|
"slash_worker": None,
|
|
"tool_progress_mode": _load_tool_progress_mode(),
|
|
"tool_started_at": {},
|
|
"transport": current_transport() or _stdio_transport,
|
|
}
|
|
|
|
# Return the lightweight session immediately so Ink can paint the composer
|
|
# + skeleton panel, then build the real AIAgent just after this response is
|
|
# flushed. This keeps startup responsive while still hydrating tools/skills
|
|
# without requiring the user to submit a first prompt.
|
|
def _deferred_build() -> None:
|
|
session = _sessions.get(sid)
|
|
if session is not None:
|
|
_start_agent_build(sid, session)
|
|
|
|
build_timer = threading.Timer(0.05, _deferred_build)
|
|
build_timer.daemon = True
|
|
build_timer.start()
|
|
|
|
return _ok(
|
|
rid,
|
|
{
|
|
"session_id": sid,
|
|
"info": {
|
|
"model": _resolve_model(),
|
|
"tools": {},
|
|
"skills": {},
|
|
"cwd": os.getenv("TERMINAL_CWD", os.getcwd()),
|
|
"lazy": True,
|
|
},
|
|
},
|
|
)
|
|
|
|
|
|
@method("session.list")
|
|
def _(rid, params: dict) -> dict:
|
|
db = _get_db()
|
|
if db is None:
|
|
return _db_unavailable_error(rid, code=5006)
|
|
try:
|
|
# Resume picker should surface human conversation sessions from every
|
|
# user-facing surface — CLI, TUI, all gateway platforms (including new
|
|
# ones not enumerated here), ACP adapter clients, webhook sessions,
|
|
# custom `HERMES_SESSION_SOURCE` values, and older installs with
|
|
# different source labels. We deny-list only the noisy internal
|
|
# sources (``tool`` sub-agent runs) rather than allow-listing a
|
|
# fixed set of platform names that goes stale whenever a new
|
|
# platform is added or a user names their own source.
|
|
deny = frozenset({"tool"})
|
|
|
|
limit = int(params.get("limit", 200) or 200)
|
|
# Over-fetch modestly so per-source filtering doesn't leave us
|
|
# short; the compression-tip projection in ``list_sessions_rich``
|
|
# can also merge rows.
|
|
fetch_limit = max(limit * 2, 200)
|
|
rows = [
|
|
s
|
|
for s in db.list_sessions_rich(source=None, limit=fetch_limit)
|
|
if (s.get("source") or "").strip().lower() not in deny
|
|
][:limit]
|
|
return _ok(
|
|
rid,
|
|
{
|
|
"sessions": [
|
|
{
|
|
"id": s["id"],
|
|
"title": s.get("title") or "",
|
|
"preview": s.get("preview") or "",
|
|
"started_at": s.get("started_at") or 0,
|
|
"message_count": s.get("message_count") or 0,
|
|
"source": s.get("source") or "",
|
|
}
|
|
for s in rows
|
|
]
|
|
},
|
|
)
|
|
except Exception as e:
|
|
return _err(rid, 5006, str(e))
|
|
|
|
|
|
@method("session.most_recent")
|
|
def _(rid, params: dict) -> dict:
|
|
"""Return the most recent human-facing session id, or ``None``.
|
|
|
|
Mirrors ``session.list``'s deny-list behaviour (drops ``tool``
|
|
sub-agent rows). Used by TUI auto-resume when
|
|
``display.tui_auto_resume_recent`` is on; the field is also handy
|
|
for any CLI tooling that wants "latest session" without paginating
|
|
the full list.
|
|
|
|
Contract: a ``{"session_id": null}`` result means "no eligible
|
|
session found right now". Errors are also folded into that
|
|
null-result shape (and logged) so callers don't have to special-
|
|
case JSON-RPC error envelopes for what is a normal "no answer".
|
|
"""
|
|
db = _get_db()
|
|
if db is None:
|
|
return _ok(rid, {"session_id": None})
|
|
try:
|
|
deny = frozenset({"tool"})
|
|
# Over-fetch by a generous bounded amount so heavy sub-agent
|
|
# users (lots of recent ``tool`` rows) don't get a false
|
|
# "no eligible session" answer. ``session.list`` uses a
|
|
# similar over-fetch strategy.
|
|
rows = db.list_sessions_rich(source=None, limit=200)
|
|
for row in rows:
|
|
src = (row.get("source") or "").strip().lower()
|
|
if src in deny:
|
|
continue
|
|
return _ok(
|
|
rid,
|
|
{
|
|
"session_id": row.get("id"),
|
|
"title": row.get("title") or "",
|
|
"started_at": row.get("started_at") or 0,
|
|
"source": row.get("source") or "",
|
|
},
|
|
)
|
|
return _ok(rid, {"session_id": None})
|
|
except Exception:
|
|
logger.exception("session.most_recent failed")
|
|
return _ok(rid, {"session_id": None})
|
|
|
|
|
|
@method("session.resume")
|
|
def _(rid, params: dict) -> dict:
|
|
target = params.get("session_id", "")
|
|
if not target:
|
|
return _err(rid, 4006, "session_id required")
|
|
db = _get_db()
|
|
if db is None:
|
|
return _db_unavailable_error(rid, code=5000)
|
|
found = db.get_session(target)
|
|
if not found:
|
|
found = db.get_session_by_title(target)
|
|
if found:
|
|
target = found["id"]
|
|
else:
|
|
return _err(rid, 4007, "session not found")
|
|
sid = uuid.uuid4().hex[:8]
|
|
_enable_gateway_prompts()
|
|
try:
|
|
db.reopen_session(target)
|
|
history = db.get_messages_as_conversation(target)
|
|
display_history = db.get_messages_as_conversation(
|
|
target, include_ancestors=True
|
|
)
|
|
messages = _history_to_messages(display_history)
|
|
tokens = _set_session_context(target)
|
|
try:
|
|
agent = _make_agent(sid, target, session_id=target)
|
|
finally:
|
|
_clear_session_context(tokens)
|
|
_init_session(sid, target, agent, history, cols=int(params.get("cols", 80)))
|
|
except Exception as e:
|
|
return _err(rid, 5000, f"resume failed: {e}")
|
|
return _ok(
|
|
rid,
|
|
{
|
|
"session_id": sid,
|
|
"resumed": target,
|
|
"message_count": len(messages),
|
|
"messages": messages,
|
|
"info": _session_info(agent),
|
|
},
|
|
)
|
|
|
|
|
|
@method("session.delete")
|
|
def _(rid, params: dict) -> dict:
|
|
"""Delete a stored session and its on-disk transcript files.
|
|
|
|
Used by the TUI resume picker (``d`` key) so users can prune old
|
|
sessions without dropping to the CLI. Refuses to delete a session
|
|
that is currently active in this gateway process — those rows are
|
|
still being written to and removing them out from under the live
|
|
agent corrupts message ordering and trips FK constraints when the
|
|
next message append flushes.
|
|
"""
|
|
target = params.get("session_id", "")
|
|
if not target:
|
|
return _err(rid, 4006, "session_id required")
|
|
db = _get_db()
|
|
if db is None:
|
|
return _db_unavailable_error(rid, code=5036)
|
|
# Block deletion of any session currently bound to a live TUI session
|
|
# in this process. The picker hides the active session anyway, but a
|
|
# racing caller could still target it. Snapshot via ``list(...)``
|
|
# because ``_sessions`` is mutated by concurrent RPCs on the thread
|
|
# pool — iterating the dict directly can raise ``RuntimeError:
|
|
# dictionary changed size during iteration``. If even the snapshot
|
|
# raises, fail closed (refuse the delete) rather than fail open.
|
|
try:
|
|
snapshot = list(_sessions.values())
|
|
except Exception as e:
|
|
return _err(rid, 5036, f"could not enumerate active sessions: {e}")
|
|
active = {s.get("session_key") for s in snapshot if s.get("session_key")}
|
|
if target in active:
|
|
return _err(rid, 4023, "cannot delete an active session")
|
|
sessions_dir = get_hermes_home() / "sessions"
|
|
try:
|
|
deleted = db.delete_session(target, sessions_dir=sessions_dir)
|
|
except Exception as e:
|
|
return _err(rid, 5036, f"delete failed: {e}")
|
|
if not deleted:
|
|
return _err(rid, 4007, "session not found")
|
|
return _ok(rid, {"deleted": target})
|
|
|
|
|
|
@method("session.title")
|
|
def _(rid, params: dict) -> dict:
|
|
session, err = _sess_nowait(params, rid)
|
|
if err:
|
|
return err
|
|
db = _get_db()
|
|
if db is None:
|
|
return _db_unavailable_error(rid, code=5007)
|
|
key = session["session_key"]
|
|
if "title" not in params:
|
|
fallback = session.get("pending_title") or ""
|
|
try:
|
|
resolved_title = db.get_session_title(key) or ""
|
|
if fallback:
|
|
if db.set_session_title(key, fallback):
|
|
session["pending_title"] = None
|
|
resolved_title = fallback
|
|
else:
|
|
existing_row = db.get_session(key)
|
|
existing_title = ((existing_row or {}).get("title") or "").strip()
|
|
if existing_title == fallback:
|
|
session["pending_title"] = None
|
|
resolved_title = fallback
|
|
elif not resolved_title:
|
|
resolved_title = fallback
|
|
elif resolved_title:
|
|
session["pending_title"] = None
|
|
except Exception:
|
|
resolved_title = fallback
|
|
return _ok(
|
|
rid,
|
|
{
|
|
"title": resolved_title,
|
|
"session_key": key,
|
|
},
|
|
)
|
|
title = (params.get("title", "") or "").strip()
|
|
if not title:
|
|
return _err(rid, 4021, "title required")
|
|
try:
|
|
if db.set_session_title(key, title):
|
|
session["pending_title"] = None
|
|
return _ok(rid, {"pending": False, "title": title})
|
|
# rowcount == 0 can mean "same value" as well as "missing row".
|
|
# Queue only when the session row truly does not exist yet.
|
|
existing_row = db.get_session(key)
|
|
if existing_row:
|
|
session["pending_title"] = None
|
|
return _ok(
|
|
rid,
|
|
{
|
|
"pending": False,
|
|
"title": (existing_row.get("title") or title),
|
|
},
|
|
)
|
|
session["pending_title"] = title
|
|
return _ok(rid, {"pending": True, "title": title})
|
|
except ValueError as e:
|
|
return _err(rid, 4022, str(e))
|
|
except Exception as e:
|
|
return _err(rid, 5007, str(e))
|
|
|
|
|
|
@method("session.usage")
|
|
def _(rid, params: dict) -> dict:
|
|
session, err = _sess_nowait(params, rid)
|
|
if err:
|
|
return err
|
|
agent = session.get("agent")
|
|
return _ok(rid, _get_usage(agent) if agent is not None else {"calls": 0, "input": 0, "output": 0, "total": 0})
|
|
|
|
|
|
@method("session.history")
|
|
def _(rid, params: dict) -> dict:
|
|
session, err = _sess_nowait(params, rid)
|
|
if err:
|
|
return err
|
|
history = list(session.get("history", []))
|
|
db = _get_db()
|
|
if db is not None and session.get("session_key"):
|
|
try:
|
|
history = db.get_messages_as_conversation(
|
|
session["session_key"], include_ancestors=True
|
|
)
|
|
except Exception:
|
|
pass
|
|
return _ok(
|
|
rid,
|
|
{
|
|
"count": len(history),
|
|
"messages": _history_to_messages(history),
|
|
},
|
|
)
|
|
|
|
|
|
@method("session.undo")
|
|
def _(rid, params: dict) -> dict:
|
|
session, err = _sess(params, rid)
|
|
if err:
|
|
return err
|
|
# Reject during an in-flight turn. If we mutated history while
|
|
# the agent thread is running, prompt.submit's post-run history
|
|
# write would either clobber the undo (version matches) or
|
|
# silently drop the agent's output (version mismatch, see below).
|
|
# Neither is what the user wants — make them /interrupt first.
|
|
if session.get("running"):
|
|
return _err(
|
|
rid, 4009, "session busy — /interrupt the current turn before /undo"
|
|
)
|
|
removed = 0
|
|
with session["history_lock"]:
|
|
history = session.get("history", [])
|
|
while history and history[-1].get("role") in ("assistant", "tool"):
|
|
history.pop()
|
|
removed += 1
|
|
if history and history[-1].get("role") == "user":
|
|
history.pop()
|
|
removed += 1
|
|
if removed:
|
|
session["history_version"] = int(session.get("history_version", 0)) + 1
|
|
return _ok(rid, {"removed": removed})
|
|
|
|
|
|
@method("session.compress")
|
|
def _(rid, params: dict) -> dict:
|
|
session, err = _sess(params, rid)
|
|
if err:
|
|
return err
|
|
if session.get("running"):
|
|
return _err(
|
|
rid, 4009, "session busy — /interrupt the current turn before /compress"
|
|
)
|
|
sid = params.get("session_id", "")
|
|
focus_topic = str(params.get("focus_topic", "") or "").strip()
|
|
try:
|
|
from agent.manual_compression_feedback import summarize_manual_compression
|
|
from agent.model_metadata import estimate_messages_tokens_rough
|
|
|
|
with session["history_lock"]:
|
|
before_messages = list(session.get("history", []))
|
|
history_version = int(session.get("history_version", 0))
|
|
before_count = len(before_messages)
|
|
before_tokens = (
|
|
estimate_messages_tokens_rough(before_messages) if before_count else 0
|
|
)
|
|
|
|
if before_count >= 4:
|
|
focus_suffix = f', focus: "{focus_topic}"' if focus_topic else ""
|
|
_status_update(
|
|
sid,
|
|
"compressing",
|
|
f"⠋ compressing {before_count} messages "
|
|
f"(~{before_tokens:,} tok){focus_suffix}…",
|
|
)
|
|
|
|
try:
|
|
removed, usage = _compress_session_history(
|
|
session,
|
|
focus_topic,
|
|
approx_tokens=before_tokens,
|
|
before_messages=before_messages,
|
|
history_version=history_version,
|
|
)
|
|
with session["history_lock"]:
|
|
messages = list(session.get("history", []))
|
|
after_count = len(messages)
|
|
after_tokens = (
|
|
estimate_messages_tokens_rough(messages) if after_count else 0
|
|
)
|
|
agent = session["agent"]
|
|
_sync_session_key_after_compress(sid, session)
|
|
summary = summarize_manual_compression(
|
|
before_messages, messages, before_tokens, after_tokens
|
|
)
|
|
info = _session_info(agent)
|
|
_emit("session.info", sid, info)
|
|
return _ok(
|
|
rid,
|
|
{
|
|
"status": "compressed",
|
|
"removed": removed,
|
|
"before_messages": before_count,
|
|
"after_messages": after_count,
|
|
"before_tokens": before_tokens,
|
|
"after_tokens": after_tokens,
|
|
"summary": summary,
|
|
"usage": usage,
|
|
"info": info,
|
|
"messages": messages,
|
|
},
|
|
)
|
|
finally:
|
|
# Always clear the pinned compressing status so the bar
|
|
# reverts to neutral whether compaction succeeded, was a
|
|
# no-op, or raised.
|
|
_status_update(sid, "ready")
|
|
except Exception as e:
|
|
return _err(rid, 5005, str(e))
|
|
|
|
|
|
@method("session.save")
|
|
def _(rid, params: dict) -> dict:
|
|
session, err = _sess(params, rid)
|
|
if err:
|
|
return err
|
|
import time as _time
|
|
|
|
filename = os.path.abspath(
|
|
f"hermes_conversation_{_time.strftime('%Y%m%d_%H%M%S')}.json"
|
|
)
|
|
try:
|
|
with open(filename, "w") as f:
|
|
json.dump(
|
|
{
|
|
"model": getattr(session["agent"], "model", ""),
|
|
"messages": session.get("history", []),
|
|
},
|
|
f,
|
|
indent=2,
|
|
ensure_ascii=False,
|
|
)
|
|
return _ok(rid, {"file": filename})
|
|
except Exception as e:
|
|
return _err(rid, 5011, str(e))
|
|
|
|
|
|
@method("session.close")
|
|
def _(rid, params: dict) -> dict:
|
|
sid = params.get("session_id", "")
|
|
session = _sessions.pop(sid, None)
|
|
if not session:
|
|
return _ok(rid, {"closed": False})
|
|
_finalize_session(session)
|
|
try:
|
|
from tools.approval import unregister_gateway_notify
|
|
|
|
unregister_gateway_notify(session["session_key"])
|
|
except Exception:
|
|
pass
|
|
try:
|
|
worker = session.get("slash_worker")
|
|
if worker:
|
|
worker.close()
|
|
except Exception:
|
|
pass
|
|
return _ok(rid, {"closed": True})
|
|
|
|
|
|
@method("session.branch")
|
|
def _(rid, params: dict) -> dict:
|
|
session, err = _sess(params, rid)
|
|
if err:
|
|
return err
|
|
db = _get_db()
|
|
if db is None:
|
|
return _db_unavailable_error(rid, code=5008)
|
|
old_key = session["session_key"]
|
|
with session["history_lock"]:
|
|
history = [dict(msg) for msg in session.get("history", [])]
|
|
if not history:
|
|
return _err(rid, 4008, "nothing to branch — send a message first")
|
|
new_key = _new_session_key()
|
|
branch_name = params.get("name", "")
|
|
try:
|
|
if branch_name:
|
|
title = branch_name
|
|
else:
|
|
current = db.get_session_title(old_key) or "branch"
|
|
title = (
|
|
db.get_next_title_in_lineage(current)
|
|
if hasattr(db, "get_next_title_in_lineage")
|
|
else f"{current} (branch)"
|
|
)
|
|
db.create_session(
|
|
new_key, source="tui", model=_resolve_model(), parent_session_id=old_key
|
|
)
|
|
for msg in history:
|
|
db.append_message(
|
|
session_id=new_key,
|
|
role=msg.get("role", "user"),
|
|
content=msg.get("content"),
|
|
)
|
|
db.set_session_title(new_key, title)
|
|
except Exception as e:
|
|
return _err(rid, 5008, f"branch failed: {e}")
|
|
new_sid = uuid.uuid4().hex[:8]
|
|
try:
|
|
tokens = _set_session_context(new_key)
|
|
try:
|
|
agent = _make_agent(new_sid, new_key, session_id=new_key)
|
|
finally:
|
|
_clear_session_context(tokens)
|
|
_init_session(
|
|
new_sid, new_key, agent, list(history), cols=session.get("cols", 80)
|
|
)
|
|
except Exception as e:
|
|
return _err(rid, 5000, f"agent init failed on branch: {e}")
|
|
return _ok(rid, {"session_id": new_sid, "title": title, "parent": old_key})
|
|
|
|
|
|
@method("session.interrupt")
|
|
def _(rid, params: dict) -> dict:
|
|
session, err = _sess(params, rid)
|
|
if err:
|
|
return err
|
|
if hasattr(session["agent"], "interrupt"):
|
|
session["agent"].interrupt()
|
|
# Scope the pending-prompt release to THIS session. A global
|
|
# _clear_pending() would collaterally cancel clarify/sudo/secret
|
|
# prompts on unrelated sessions sharing the same tui_gateway
|
|
# process, silently resolving them to empty strings.
|
|
_clear_pending(params.get("session_id", ""))
|
|
try:
|
|
from tools.approval import resolve_gateway_approval
|
|
|
|
resolve_gateway_approval(session["session_key"], "deny", resolve_all=True)
|
|
except Exception:
|
|
pass
|
|
return _ok(rid, {"status": "interrupted"})
|
|
|
|
|
|
# ── Delegation: subagent tree observability + controls ───────────────
|
|
# Powers the TUI's /agents overlay (see ui-tui/src/components/agentsOverlay).
|
|
# The registry lives in tools/delegate_tool — these handlers are thin
|
|
# translators between JSON-RPC and the Python API.
|
|
|
|
|
|
@method("delegation.status")
|
|
def _(rid, params: dict) -> dict:
|
|
from tools.delegate_tool import (
|
|
is_spawn_paused,
|
|
list_active_subagents,
|
|
_get_max_concurrent_children,
|
|
_get_max_spawn_depth,
|
|
)
|
|
|
|
return _ok(
|
|
rid,
|
|
{
|
|
"active": list_active_subagents(),
|
|
"paused": is_spawn_paused(),
|
|
"max_spawn_depth": _get_max_spawn_depth(),
|
|
"max_concurrent_children": _get_max_concurrent_children(),
|
|
},
|
|
)
|
|
|
|
|
|
@method("delegation.pause")
|
|
def _(rid, params: dict) -> dict:
|
|
from tools.delegate_tool import set_spawn_paused
|
|
|
|
paused = bool(params.get("paused", True))
|
|
return _ok(rid, {"paused": set_spawn_paused(paused)})
|
|
|
|
|
|
@method("subagent.interrupt")
|
|
def _(rid, params: dict) -> dict:
|
|
from tools.delegate_tool import interrupt_subagent
|
|
|
|
subagent_id = str(params.get("subagent_id") or "").strip()
|
|
if not subagent_id:
|
|
return _err(rid, 4000, "subagent_id required")
|
|
ok = interrupt_subagent(subagent_id)
|
|
return _ok(rid, {"found": ok, "subagent_id": subagent_id})
|
|
|
|
|
|
# ── Spawn-tree snapshots: TUI-written, disk-persisted ────────────────
|
|
# The TUI is the source of truth for subagent state (it assembles payloads
|
|
# from the event stream). On turn-complete it posts the final tree here;
|
|
# /replay and /replay-diff fetch past snapshots by session_id + filename.
|
|
#
|
|
# Layout: $HERMES_HOME/spawn-trees/<session_id>/<timestamp>.json
|
|
# Each file contains { session_id, started_at, finished_at, subagents: [...] }.
|
|
|
|
|
|
def _spawn_trees_root():
|
|
from pathlib import Path as _P
|
|
from hermes_constants import get_hermes_home
|
|
|
|
root = get_hermes_home() / "spawn-trees"
|
|
root.mkdir(parents=True, exist_ok=True)
|
|
return root
|
|
|
|
|
|
def _spawn_tree_session_dir(session_id: str):
|
|
safe = (
|
|
"".join(c if c.isalnum() or c in "-_" else "_" for c in session_id) or "unknown"
|
|
)
|
|
d = _spawn_trees_root() / safe
|
|
d.mkdir(parents=True, exist_ok=True)
|
|
return d
|
|
|
|
|
|
# Per-session append-only index of lightweight snapshot metadata. Read by
|
|
# `spawn_tree.list` so scanning doesn't require reading every full snapshot
|
|
# file (Copilot review on #14045). One JSON object per line.
|
|
_SPAWN_TREE_INDEX = "_index.jsonl"
|
|
|
|
|
|
def _append_spawn_tree_index(session_dir, entry: dict) -> None:
|
|
try:
|
|
with (session_dir / _SPAWN_TREE_INDEX).open("a", encoding="utf-8") as f:
|
|
f.write(json.dumps(entry, ensure_ascii=False) + "\n")
|
|
except OSError as exc:
|
|
# Index is a cache — losing a line just means list() falls back
|
|
# to a directory scan for that entry. Never block the save.
|
|
logger.debug("spawn_tree index append failed: %s", exc)
|
|
|
|
|
|
def _read_spawn_tree_index(session_dir) -> list[dict]:
|
|
index_path = session_dir / _SPAWN_TREE_INDEX
|
|
if not index_path.exists():
|
|
return []
|
|
out: list[dict] = []
|
|
try:
|
|
with index_path.open("r", encoding="utf-8") as f:
|
|
for line in f:
|
|
line = line.strip()
|
|
if not line:
|
|
continue
|
|
try:
|
|
out.append(json.loads(line))
|
|
except json.JSONDecodeError:
|
|
continue
|
|
except OSError:
|
|
return []
|
|
return out
|
|
|
|
|
|
@method("spawn_tree.save")
|
|
def _(rid, params: dict) -> dict:
|
|
session_id = str(params.get("session_id") or "").strip()
|
|
subagents = params.get("subagents") or []
|
|
if not isinstance(subagents, list) or not subagents:
|
|
return _err(rid, 4000, "subagents list required")
|
|
|
|
from datetime import datetime
|
|
|
|
started_at = params.get("started_at")
|
|
finished_at = params.get("finished_at") or time.time()
|
|
label = str(params.get("label") or "")
|
|
ts = datetime.utcfromtimestamp(float(finished_at)).strftime("%Y%m%dT%H%M%S")
|
|
fname = f"{ts}.json"
|
|
d = _spawn_tree_session_dir(session_id or "default")
|
|
path = d / fname
|
|
try:
|
|
payload = {
|
|
"session_id": session_id,
|
|
"started_at": float(started_at) if started_at else None,
|
|
"finished_at": float(finished_at),
|
|
"label": label,
|
|
"subagents": subagents,
|
|
}
|
|
path.write_text(json.dumps(payload, ensure_ascii=False), encoding="utf-8")
|
|
except OSError as exc:
|
|
return _err(rid, 5000, f"spawn_tree.save failed: {exc}")
|
|
|
|
_append_spawn_tree_index(
|
|
d,
|
|
{
|
|
"path": str(path),
|
|
"session_id": session_id,
|
|
"started_at": payload["started_at"],
|
|
"finished_at": payload["finished_at"],
|
|
"label": label,
|
|
"count": len(subagents),
|
|
},
|
|
)
|
|
|
|
return _ok(rid, {"path": str(path), "session_id": session_id})
|
|
|
|
|
|
@method("spawn_tree.list")
|
|
def _(rid, params: dict) -> dict:
|
|
session_id = str(params.get("session_id") or "").strip()
|
|
limit = int(params.get("limit") or 50)
|
|
cross_session = bool(params.get("cross_session"))
|
|
|
|
if cross_session:
|
|
root = _spawn_trees_root()
|
|
roots = [p for p in root.iterdir() if p.is_dir()]
|
|
else:
|
|
roots = [_spawn_tree_session_dir(session_id or "default")]
|
|
|
|
entries: list[dict] = []
|
|
for d in roots:
|
|
indexed = _read_spawn_tree_index(d)
|
|
if indexed:
|
|
# Skip index entries whose snapshot file was manually deleted.
|
|
entries.extend(
|
|
e for e in indexed if (p := e.get("path")) and Path(p).exists()
|
|
)
|
|
continue
|
|
|
|
# Fallback for legacy (pre-index) sessions: full scan. O(N) reads
|
|
# but only runs once per session until the next save writes the index.
|
|
for p in d.glob("*.json"):
|
|
if p.name == _SPAWN_TREE_INDEX:
|
|
continue
|
|
try:
|
|
stat = p.stat()
|
|
try:
|
|
raw = json.loads(p.read_text(encoding="utf-8"))
|
|
except Exception:
|
|
raw = {}
|
|
subagents = raw.get("subagents") or []
|
|
entries.append(
|
|
{
|
|
"path": str(p),
|
|
"session_id": raw.get("session_id") or d.name,
|
|
"finished_at": raw.get("finished_at") or stat.st_mtime,
|
|
"started_at": raw.get("started_at"),
|
|
"label": raw.get("label") or "",
|
|
"count": len(subagents) if isinstance(subagents, list) else 0,
|
|
}
|
|
)
|
|
except OSError:
|
|
continue
|
|
|
|
entries.sort(key=lambda e: e.get("finished_at") or 0, reverse=True)
|
|
return _ok(rid, {"entries": entries[:limit]})
|
|
|
|
|
|
@method("spawn_tree.load")
|
|
def _(rid, params: dict) -> dict:
|
|
from pathlib import Path
|
|
|
|
raw_path = str(params.get("path") or "").strip()
|
|
if not raw_path:
|
|
return _err(rid, 4000, "path required")
|
|
|
|
# Reject paths escaping the spawn-trees root.
|
|
root = _spawn_trees_root().resolve()
|
|
try:
|
|
resolved = Path(raw_path).resolve()
|
|
resolved.relative_to(root)
|
|
except (ValueError, OSError) as exc:
|
|
return _err(rid, 4030, f"path outside spawn-trees root: {exc}")
|
|
|
|
try:
|
|
payload = json.loads(resolved.read_text(encoding="utf-8"))
|
|
except (OSError, json.JSONDecodeError) as exc:
|
|
return _err(rid, 5000, f"spawn_tree.load failed: {exc}")
|
|
|
|
return _ok(rid, payload)
|
|
|
|
|
|
@method("session.steer")
|
|
def _(rid, params: dict) -> dict:
|
|
"""Inject a user message into the next tool result without interrupting.
|
|
|
|
Mirrors AIAgent.steer(). Safe to call while a turn is running — the text
|
|
lands on the last tool result of the next tool batch and the model sees
|
|
it on its next iteration. No interrupt, no new user turn, no role
|
|
alternation violation.
|
|
"""
|
|
text = (params.get("text") or "").strip()
|
|
if not text:
|
|
return _err(rid, 4002, "text is required")
|
|
session, err = _sess_nowait(params, rid)
|
|
if err:
|
|
return err
|
|
agent = session.get("agent")
|
|
if agent is None or not hasattr(agent, "steer"):
|
|
return _err(rid, 4010, "agent does not support steer")
|
|
try:
|
|
accepted = agent.steer(text)
|
|
except Exception as exc:
|
|
return _err(rid, 5000, f"steer failed: {exc}")
|
|
return _ok(rid, {"status": "queued" if accepted else "rejected", "text": text})
|
|
|
|
|
|
@method("terminal.resize")
|
|
def _(rid, params: dict) -> dict:
|
|
session, err = _sess_nowait(params, rid)
|
|
if err:
|
|
return err
|
|
session["cols"] = int(params.get("cols", 80))
|
|
return _ok(rid, {"cols": session["cols"]})
|
|
|
|
|
|
# ── Methods: prompt ──────────────────────────────────────────────────
|
|
|
|
|
|
@method("prompt.submit")
|
|
def _(rid, params: dict) -> dict:
|
|
sid, text = params.get("session_id", ""), params.get("text", "")
|
|
session, err = _sess_nowait(params, rid)
|
|
if err:
|
|
return err
|
|
with session["history_lock"]:
|
|
if session.get("running"):
|
|
return _err(rid, 4009, "session busy")
|
|
session["running"] = True
|
|
|
|
_start_agent_build(sid, session)
|
|
|
|
def run_after_agent_ready() -> None:
|
|
err = _wait_agent(session, rid)
|
|
if err:
|
|
_emit("error", sid, {"message": err.get("error", {}).get("message", "agent initialization failed")})
|
|
with session["history_lock"]:
|
|
session["running"] = False
|
|
return
|
|
_run_prompt_submit(rid, sid, session, text)
|
|
|
|
threading.Thread(target=run_after_agent_ready, daemon=True).start()
|
|
return _ok(rid, {"status": "streaming"})
|
|
|
|
|
|
def _run_prompt_submit(rid, sid: str, session: dict, text: Any) -> None:
|
|
with session["history_lock"]:
|
|
history = list(session["history"])
|
|
history_version = int(session.get("history_version", 0))
|
|
images = list(session.get("attached_images", []))
|
|
session["attached_images"] = []
|
|
agent = session["agent"]
|
|
_emit("message.start", sid)
|
|
|
|
def run():
|
|
approval_token = None
|
|
session_tokens = []
|
|
try:
|
|
from tools.approval import (
|
|
reset_current_session_key,
|
|
set_current_session_key,
|
|
)
|
|
|
|
approval_token = set_current_session_key(session["session_key"])
|
|
session_tokens = _set_session_context(session["session_key"])
|
|
cols = session.get("cols", 80)
|
|
streamer = make_stream_renderer(cols)
|
|
prompt = text
|
|
|
|
if isinstance(prompt, str) and "@" in prompt:
|
|
from agent.context_references import preprocess_context_references
|
|
from agent.model_metadata import get_model_context_length
|
|
|
|
ctx_len = get_model_context_length(
|
|
getattr(agent, "model", "") or _resolve_model(),
|
|
base_url=getattr(agent, "base_url", "") or "",
|
|
api_key=getattr(agent, "api_key", "") or "",
|
|
provider=getattr(agent, "provider", "") or "",
|
|
)
|
|
ctx = preprocess_context_references(
|
|
prompt,
|
|
cwd=os.environ.get("TERMINAL_CWD", os.getcwd()),
|
|
allowed_root=os.environ.get("TERMINAL_CWD", os.getcwd()),
|
|
context_length=ctx_len,
|
|
)
|
|
if ctx.blocked:
|
|
_emit(
|
|
"error",
|
|
sid,
|
|
{
|
|
"message": "\n".join(ctx.warnings)
|
|
or "Context injection refused."
|
|
},
|
|
)
|
|
return
|
|
prompt = ctx.message
|
|
|
|
# Decide image routing per-turn based on active provider/model.
|
|
# "native" → pass pixels to the main model as OpenAI-style content
|
|
# parts (adapters translate for Anthropic/Gemini/Bedrock/etc.).
|
|
# "text" → pre-analyze with vision_analyze and prepend the text.
|
|
# See agent/image_routing.py for the full decision table.
|
|
run_message: Any = prompt
|
|
if images:
|
|
try:
|
|
from agent.image_routing import (
|
|
decide_image_input_mode,
|
|
build_native_content_parts,
|
|
)
|
|
from agent.auxiliary_client import (
|
|
_read_main_model,
|
|
_read_main_provider,
|
|
)
|
|
from hermes_cli.config import load_config as _tui_load_config
|
|
|
|
_cfg = _tui_load_config()
|
|
_mode = decide_image_input_mode(
|
|
_read_main_provider(),
|
|
_read_main_model(),
|
|
_cfg,
|
|
)
|
|
except Exception as _img_exc:
|
|
print(
|
|
f"[tui_gateway] image_routing decision failed, defaulting to text: {_img_exc}",
|
|
file=sys.stderr,
|
|
)
|
|
_mode = "text"
|
|
|
|
if _mode == "native":
|
|
try:
|
|
_parts, _skipped = build_native_content_parts(
|
|
prompt,
|
|
images,
|
|
)
|
|
if _skipped:
|
|
print(
|
|
f"[tui_gateway] native image attachment skipped {len(_skipped)} unreadable path(s)",
|
|
file=sys.stderr,
|
|
)
|
|
if any(p.get("type") == "image_url" for p in _parts):
|
|
run_message = _parts
|
|
else:
|
|
run_message = _enrich_with_attached_images(prompt, images)
|
|
except Exception as _img_exc:
|
|
print(
|
|
f"[tui_gateway] native attach failed, falling back to text: {_img_exc}",
|
|
file=sys.stderr,
|
|
)
|
|
run_message = _enrich_with_attached_images(prompt, images)
|
|
else:
|
|
run_message = _enrich_with_attached_images(prompt, images)
|
|
|
|
def _stream(delta):
|
|
payload = {"text": delta}
|
|
if streamer and (r := streamer.feed(delta)) is not None:
|
|
payload["rendered"] = r
|
|
_emit("message.delta", sid, payload)
|
|
|
|
result = agent.run_conversation(
|
|
run_message,
|
|
conversation_history=list(history),
|
|
stream_callback=_stream,
|
|
)
|
|
|
|
last_reasoning = None
|
|
status_note = None
|
|
if isinstance(result, dict):
|
|
if isinstance(result.get("messages"), list):
|
|
with session["history_lock"]:
|
|
current_version = int(session.get("history_version", 0))
|
|
if current_version == history_version:
|
|
session["history"] = result["messages"]
|
|
session["history_version"] = history_version + 1
|
|
else:
|
|
# History mutated externally during the turn
|
|
# (undo/compress/retry/rollback now guard on
|
|
# session.running, but this is the defensive
|
|
# backstop for any path that slips past).
|
|
# Surface the desync rather than silently
|
|
# dropping the agent's output — the UI can
|
|
# show the response and warn that it was
|
|
# not persisted.
|
|
print(
|
|
f"[tui_gateway] prompt.submit: history_version mismatch "
|
|
f"(expected={history_version} current={current_version}) — "
|
|
f"agent output NOT written to session history",
|
|
file=sys.stderr,
|
|
)
|
|
status_note = (
|
|
"History changed during this turn — the response above is visible "
|
|
"but was not saved to session history."
|
|
)
|
|
raw = result.get("final_response", "")
|
|
status = (
|
|
"interrupted"
|
|
if result.get("interrupted")
|
|
else "error" if result.get("error") else "complete"
|
|
)
|
|
lr = result.get("last_reasoning")
|
|
if isinstance(lr, str) and lr.strip():
|
|
last_reasoning = lr.strip()
|
|
else:
|
|
raw = str(result)
|
|
status = "complete"
|
|
|
|
payload = {"text": raw, "usage": _get_usage(agent), "status": status}
|
|
if last_reasoning:
|
|
payload["reasoning"] = last_reasoning
|
|
if status_note:
|
|
payload["warning"] = status_note
|
|
rendered = render_message(raw, cols)
|
|
if rendered:
|
|
payload["rendered"] = rendered
|
|
_emit("message.complete", sid, payload)
|
|
|
|
if (
|
|
status == "complete"
|
|
and isinstance(raw, str)
|
|
and raw.strip()
|
|
and isinstance(text, str)
|
|
and text.strip()
|
|
):
|
|
try:
|
|
from agent.title_generator import maybe_auto_title
|
|
|
|
maybe_auto_title(
|
|
_get_db(),
|
|
session.get("session_key") or sid,
|
|
text,
|
|
raw,
|
|
session.get("history", []),
|
|
)
|
|
except Exception:
|
|
pass
|
|
|
|
# CLI parity: when voice-mode TTS is on, speak the agent reply
|
|
# (cli.py:_voice_speak_response). Only the final text — tool
|
|
# calls / reasoning already stream separately and would be
|
|
# noisy to read aloud.
|
|
if (
|
|
status == "complete"
|
|
and isinstance(raw, str)
|
|
and raw.strip()
|
|
and _voice_tts_enabled()
|
|
):
|
|
try:
|
|
from hermes_cli.voice import speak_text
|
|
|
|
spoken = raw
|
|
threading.Thread(
|
|
target=speak_text, args=(spoken,), daemon=True
|
|
).start()
|
|
except ImportError:
|
|
logger.warning("voice TTS skipped: hermes_cli.voice unavailable")
|
|
except Exception as e:
|
|
logger.warning("voice TTS dispatch failed: %s", e)
|
|
except Exception as e:
|
|
import traceback
|
|
|
|
trace = traceback.format_exc()
|
|
try:
|
|
os.makedirs(os.path.dirname(_CRASH_LOG), exist_ok=True)
|
|
with open(_CRASH_LOG, "a", encoding="utf-8") as f:
|
|
f.write(
|
|
f"\n=== turn-dispatcher exception · "
|
|
f"{time.strftime('%Y-%m-%d %H:%M:%S')} · sid={sid} ===\n"
|
|
)
|
|
f.write(trace)
|
|
except Exception:
|
|
pass
|
|
print(
|
|
f"[gateway-turn] {type(e).__name__}: {e}", file=sys.stderr, flush=True
|
|
)
|
|
_emit("error", sid, {"message": str(e)})
|
|
finally:
|
|
try:
|
|
if approval_token is not None:
|
|
reset_current_session_key(approval_token)
|
|
except Exception:
|
|
pass
|
|
_clear_session_context(session_tokens)
|
|
with session["history_lock"]:
|
|
session["running"] = False
|
|
|
|
threading.Thread(target=run, daemon=True).start()
|
|
|
|
|
|
@method("clipboard.paste")
|
|
def _(rid, params: dict) -> dict:
|
|
session, err = _sess(params, rid)
|
|
if err:
|
|
return err
|
|
try:
|
|
from hermes_cli.clipboard import has_clipboard_image, save_clipboard_image
|
|
except Exception as e:
|
|
return _err(rid, 5027, f"clipboard unavailable: {e}")
|
|
|
|
session["image_counter"] = session.get("image_counter", 0) + 1
|
|
img_dir = _hermes_home / "images"
|
|
img_dir.mkdir(parents=True, exist_ok=True)
|
|
img_path = (
|
|
img_dir
|
|
/ f"clip_{datetime.now().strftime('%Y%m%d_%H%M%S')}_{session['image_counter']}.png"
|
|
)
|
|
|
|
# Save-first: mirrors CLI keybinding path; more robust than has_image() precheck
|
|
if not save_clipboard_image(img_path):
|
|
session["image_counter"] = max(0, session["image_counter"] - 1)
|
|
msg = (
|
|
"Clipboard has image but extraction failed"
|
|
if has_clipboard_image()
|
|
else "No image found in clipboard"
|
|
)
|
|
return _ok(rid, {"attached": False, "message": msg})
|
|
|
|
session.setdefault("attached_images", []).append(str(img_path))
|
|
return _ok(
|
|
rid,
|
|
{
|
|
"attached": True,
|
|
"path": str(img_path),
|
|
"count": len(session["attached_images"]),
|
|
**_image_meta(img_path),
|
|
},
|
|
)
|
|
|
|
|
|
@method("image.attach")
|
|
def _(rid, params: dict) -> dict:
|
|
session, err = _sess(params, rid)
|
|
if err:
|
|
return err
|
|
raw = str(params.get("path", "") or "").strip()
|
|
if not raw:
|
|
return _err(rid, 4015, "path required")
|
|
try:
|
|
from cli import (
|
|
_IMAGE_EXTENSIONS,
|
|
_detect_file_drop,
|
|
_resolve_attachment_path,
|
|
_split_path_input,
|
|
)
|
|
|
|
dropped = _detect_file_drop(raw)
|
|
if dropped:
|
|
image_path = dropped["path"]
|
|
remainder = dropped["remainder"]
|
|
else:
|
|
path_token, remainder = _split_path_input(raw)
|
|
image_path = _resolve_attachment_path(path_token)
|
|
if image_path is None:
|
|
return _err(rid, 4016, f"image not found: {path_token}")
|
|
if image_path.suffix.lower() not in _IMAGE_EXTENSIONS:
|
|
return _err(rid, 4016, f"unsupported image: {image_path.name}")
|
|
session.setdefault("attached_images", []).append(str(image_path))
|
|
return _ok(
|
|
rid,
|
|
{
|
|
"attached": True,
|
|
"path": str(image_path),
|
|
"count": len(session["attached_images"]),
|
|
"remainder": remainder,
|
|
"text": remainder or f"[User attached image: {image_path.name}]",
|
|
**_image_meta(image_path),
|
|
},
|
|
)
|
|
except Exception as e:
|
|
return _err(rid, 5027, str(e))
|
|
|
|
|
|
@method("input.detect_drop")
|
|
def _(rid, params: dict) -> dict:
|
|
session, err = _sess_nowait(params, rid)
|
|
if err:
|
|
return err
|
|
try:
|
|
from cli import _detect_file_drop
|
|
|
|
raw = str(params.get("text", "") or "")
|
|
dropped = _detect_file_drop(raw)
|
|
if not dropped:
|
|
return _ok(rid, {"matched": False})
|
|
|
|
drop_path = dropped["path"]
|
|
remainder = dropped["remainder"]
|
|
if dropped["is_image"]:
|
|
session.setdefault("attached_images", []).append(str(drop_path))
|
|
text = remainder or f"[User attached image: {drop_path.name}]"
|
|
return _ok(
|
|
rid,
|
|
{
|
|
"matched": True,
|
|
"is_image": True,
|
|
"path": str(drop_path),
|
|
"count": len(session["attached_images"]),
|
|
"text": text,
|
|
**_image_meta(drop_path),
|
|
},
|
|
)
|
|
|
|
text = f"[User attached file: {drop_path}]" + (
|
|
f"\n{remainder}" if remainder else ""
|
|
)
|
|
return _ok(
|
|
rid,
|
|
{
|
|
"matched": True,
|
|
"is_image": False,
|
|
"path": str(drop_path),
|
|
"name": drop_path.name,
|
|
"text": text,
|
|
},
|
|
)
|
|
except Exception as e:
|
|
return _err(rid, 5027, str(e))
|
|
|
|
|
|
@method("prompt.background")
|
|
def _(rid, params: dict) -> dict:
|
|
session, err = _sess(params, rid)
|
|
if err:
|
|
return err
|
|
text, parent = params.get("text", ""), params.get("session_id", "")
|
|
if not text:
|
|
return _err(rid, 4012, "text required")
|
|
task_id = f"bg_{uuid.uuid4().hex[:6]}"
|
|
|
|
def run():
|
|
session_tokens = _set_session_context(task_id)
|
|
try:
|
|
from run_agent import AIAgent
|
|
|
|
result = AIAgent(
|
|
**_background_agent_kwargs(session["agent"], task_id)
|
|
).run_conversation(
|
|
user_message=text,
|
|
task_id=task_id,
|
|
)
|
|
_emit(
|
|
"background.complete",
|
|
parent,
|
|
{
|
|
"task_id": task_id,
|
|
"text": (
|
|
result.get("final_response", str(result))
|
|
if isinstance(result, dict)
|
|
else str(result)
|
|
),
|
|
},
|
|
)
|
|
except Exception as e:
|
|
_emit(
|
|
"background.complete",
|
|
parent,
|
|
{"task_id": task_id, "text": f"error: {e}"},
|
|
)
|
|
finally:
|
|
_clear_session_context(session_tokens)
|
|
|
|
threading.Thread(target=run, daemon=True).start()
|
|
return _ok(rid, {"task_id": task_id})
|
|
|
|
|
|
# ── Methods: respond ─────────────────────────────────────────────────
|
|
|
|
|
|
def _respond(rid, params, key):
|
|
r = params.get("request_id", "")
|
|
entry = _pending.get(r)
|
|
if not entry:
|
|
return _err(rid, 4009, f"no pending {key} request")
|
|
_, ev = entry
|
|
_answers[r] = params.get(key, "")
|
|
ev.set()
|
|
return _ok(rid, {"status": "ok"})
|
|
|
|
|
|
@method("clarify.respond")
|
|
def _(rid, params: dict) -> dict:
|
|
return _respond(rid, params, "answer")
|
|
|
|
|
|
@method("sudo.respond")
|
|
def _(rid, params: dict) -> dict:
|
|
return _respond(rid, params, "password")
|
|
|
|
|
|
@method("secret.respond")
|
|
def _(rid, params: dict) -> dict:
|
|
return _respond(rid, params, "value")
|
|
|
|
|
|
@method("approval.respond")
|
|
def _(rid, params: dict) -> dict:
|
|
session, err = _sess(params, rid)
|
|
if err:
|
|
return err
|
|
try:
|
|
from tools.approval import resolve_gateway_approval
|
|
|
|
return _ok(
|
|
rid,
|
|
{
|
|
"resolved": resolve_gateway_approval(
|
|
session["session_key"],
|
|
params.get("choice", "deny"),
|
|
resolve_all=params.get("all", False),
|
|
)
|
|
},
|
|
)
|
|
except Exception as e:
|
|
return _err(rid, 5004, str(e))
|
|
|
|
|
|
# ── Methods: config ──────────────────────────────────────────────────
|
|
|
|
|
|
@method("config.set")
|
|
def _(rid, params: dict) -> dict:
|
|
key, value = params.get("key", ""), params.get("value", "")
|
|
session = _sessions.get(params.get("session_id", ""))
|
|
|
|
if key == "model":
|
|
try:
|
|
if not value:
|
|
return _err(rid, 4002, "model value required")
|
|
if session:
|
|
# Reject during an in-flight turn. agent.switch_model()
|
|
# mutates self.model / self.provider / self.base_url /
|
|
# self.client in place; the worker thread running
|
|
# agent.run_conversation is reading those on every
|
|
# iteration. A mid-turn swap can send an HTTP request
|
|
# with the new base_url but old model (or vice versa),
|
|
# producing 400/404s the user never asked for. Parity
|
|
# with the gateway's running-agent /model guard.
|
|
if session.get("running"):
|
|
return _err(
|
|
rid,
|
|
4009,
|
|
"session busy — /interrupt the current turn before switching models",
|
|
)
|
|
result = _apply_model_switch(
|
|
params.get("session_id", ""), session, value
|
|
)
|
|
else:
|
|
result = _apply_model_switch("", {"agent": None}, value)
|
|
return _ok(
|
|
rid,
|
|
{"key": key, "value": result["value"], "warning": result["warning"]},
|
|
)
|
|
except Exception as e:
|
|
return _err(rid, 5001, str(e))
|
|
|
|
if key == "fast":
|
|
raw = str(value or "").strip().lower()
|
|
agent = session.get("agent") if session else None
|
|
if agent is not None:
|
|
current_fast = getattr(agent, "service_tier", None) == "priority"
|
|
else:
|
|
current_fast = _load_service_tier() == "priority"
|
|
|
|
if raw in {"status"}:
|
|
return _ok(
|
|
rid,
|
|
{"key": key, "value": "fast" if current_fast else "normal"},
|
|
)
|
|
|
|
if raw in ("", "toggle"):
|
|
nv = "normal" if current_fast else "fast"
|
|
elif raw in {"fast", "on"}:
|
|
nv = "fast"
|
|
elif raw in {"normal", "off"}:
|
|
nv = "normal"
|
|
else:
|
|
return _err(rid, 4002, f"unknown fast mode: {value}")
|
|
|
|
overrides = None
|
|
if nv == "fast":
|
|
from hermes_cli.models import resolve_fast_mode_overrides
|
|
|
|
target_model = (
|
|
getattr(agent, "model", None) if agent is not None else _resolve_model()
|
|
)
|
|
if not target_model:
|
|
return _err(
|
|
rid,
|
|
4002,
|
|
"fast mode is not available without a selected model",
|
|
)
|
|
overrides = resolve_fast_mode_overrides(target_model)
|
|
if overrides is None:
|
|
return _err(
|
|
rid,
|
|
4002,
|
|
"fast mode is not available for this model",
|
|
)
|
|
|
|
_write_config_key("agent.service_tier", nv)
|
|
if agent is not None:
|
|
agent.service_tier = "priority" if nv == "fast" else None
|
|
current_overrides = dict(getattr(agent, "request_overrides", {}) or {})
|
|
current_overrides.pop("service_tier", None)
|
|
current_overrides.pop("speed", None)
|
|
if nv == "fast":
|
|
current_overrides.update(overrides)
|
|
agent.request_overrides = current_overrides
|
|
_emit(
|
|
"session.info",
|
|
params.get("session_id", ""),
|
|
_session_info(agent),
|
|
)
|
|
return _ok(rid, {"key": key, "value": nv})
|
|
|
|
if key == "busy":
|
|
raw = str(value or "").strip().lower()
|
|
if raw in ("", "status"):
|
|
return _ok(rid, {"key": key, "value": _load_busy_input_mode()})
|
|
if raw not in {"queue", "steer", "interrupt"}:
|
|
return _err(rid, 4002, f"unknown busy mode: {value}")
|
|
_write_config_key("display.busy_input_mode", raw)
|
|
return _ok(rid, {"key": key, "value": raw})
|
|
|
|
if key == "verbose":
|
|
cycle = ["off", "new", "all", "verbose"]
|
|
cur = (
|
|
session.get("tool_progress_mode", _load_tool_progress_mode())
|
|
if session
|
|
else _load_tool_progress_mode()
|
|
)
|
|
if value and value != "cycle":
|
|
nv = str(value).strip().lower()
|
|
if nv not in cycle:
|
|
return _err(rid, 4002, f"unknown verbose mode: {value}")
|
|
else:
|
|
try:
|
|
idx = cycle.index(cur)
|
|
except ValueError:
|
|
idx = 2
|
|
nv = cycle[(idx + 1) % len(cycle)]
|
|
_write_config_key("display.tool_progress", nv)
|
|
if session:
|
|
session["tool_progress_mode"] = nv
|
|
agent = session.get("agent")
|
|
if agent is not None:
|
|
agent.verbose_logging = nv == "verbose"
|
|
return _ok(rid, {"key": key, "value": nv})
|
|
|
|
if key == "yolo":
|
|
try:
|
|
if session:
|
|
from tools.approval import (
|
|
disable_session_yolo,
|
|
enable_session_yolo,
|
|
is_session_yolo_enabled,
|
|
)
|
|
|
|
current = is_session_yolo_enabled(session["session_key"])
|
|
if current:
|
|
disable_session_yolo(session["session_key"])
|
|
nv = "0"
|
|
else:
|
|
enable_session_yolo(session["session_key"])
|
|
nv = "1"
|
|
else:
|
|
current = bool(os.environ.get("HERMES_YOLO_MODE"))
|
|
if current:
|
|
os.environ.pop("HERMES_YOLO_MODE", None)
|
|
nv = "0"
|
|
else:
|
|
os.environ["HERMES_YOLO_MODE"] = "1"
|
|
nv = "1"
|
|
return _ok(rid, {"key": key, "value": nv})
|
|
except Exception as e:
|
|
return _err(rid, 5001, str(e))
|
|
|
|
if key == "reasoning":
|
|
try:
|
|
from hermes_constants import parse_reasoning_effort
|
|
|
|
arg = str(value or "").strip().lower()
|
|
if arg in ("show", "on"):
|
|
cfg = _load_cfg()
|
|
display = cfg.get("display") if isinstance(cfg.get("display"), dict) else {}
|
|
sections = (
|
|
display.get("sections")
|
|
if isinstance(display.get("sections"), dict)
|
|
else {}
|
|
)
|
|
display["show_reasoning"] = True
|
|
sections["thinking"] = "expanded"
|
|
display["sections"] = sections
|
|
cfg["display"] = display
|
|
_save_cfg(cfg)
|
|
if session:
|
|
session["show_reasoning"] = True
|
|
return _ok(rid, {"key": key, "value": "show"})
|
|
if arg in ("hide", "off"):
|
|
cfg = _load_cfg()
|
|
display = cfg.get("display") if isinstance(cfg.get("display"), dict) else {}
|
|
sections = (
|
|
display.get("sections")
|
|
if isinstance(display.get("sections"), dict)
|
|
else {}
|
|
)
|
|
display["show_reasoning"] = False
|
|
sections["thinking"] = "hidden"
|
|
display["sections"] = sections
|
|
cfg["display"] = display
|
|
_save_cfg(cfg)
|
|
if session:
|
|
session["show_reasoning"] = False
|
|
return _ok(rid, {"key": key, "value": "hide"})
|
|
|
|
parsed = parse_reasoning_effort(arg)
|
|
if parsed is None:
|
|
return _err(rid, 4002, f"unknown reasoning value: {value}")
|
|
_write_config_key("agent.reasoning_effort", arg)
|
|
if session and session.get("agent") is not None:
|
|
session["agent"].reasoning_config = parsed
|
|
return _ok(rid, {"key": key, "value": arg})
|
|
except Exception as e:
|
|
return _err(rid, 5001, str(e))
|
|
|
|
if key == "details_mode":
|
|
nv = str(value or "").strip().lower()
|
|
if nv not in _DETAIL_MODES:
|
|
return _err(rid, 4002, f"unknown details_mode: {value}")
|
|
cfg = _load_cfg()
|
|
display = cfg.get("display") if isinstance(cfg.get("display"), dict) else {}
|
|
sections = display.get("sections") if isinstance(display.get("sections"), dict) else {}
|
|
display["details_mode"] = nv
|
|
for section in _DETAIL_SECTION_NAMES:
|
|
sections[section] = nv
|
|
display["sections"] = sections
|
|
cfg["display"] = display
|
|
_save_cfg(cfg)
|
|
return _ok(rid, {"key": key, "value": nv})
|
|
|
|
if key.startswith("details_mode."):
|
|
# Per-section override: `details_mode.<section>` writes to
|
|
# `display.sections.<section>`. Empty value clears the explicit
|
|
# override and lets frontend resolution apply built-in section defaults
|
|
# before the global details_mode.
|
|
section = key.split(".", 1)[1]
|
|
if section not in _DETAIL_SECTION_NAMES:
|
|
return _err(rid, 4002, f"unknown section: {section}")
|
|
|
|
cfg = _load_cfg()
|
|
display = cfg.get("display") if isinstance(cfg.get("display"), dict) else {}
|
|
sections_cfg = (
|
|
display.get("sections") if isinstance(display.get("sections"), dict) else {}
|
|
)
|
|
|
|
nv = str(value or "").strip().lower()
|
|
if not nv:
|
|
sections_cfg.pop(section, None)
|
|
display["sections"] = sections_cfg
|
|
cfg["display"] = display
|
|
_save_cfg(cfg)
|
|
return _ok(rid, {"key": key, "value": ""})
|
|
|
|
if nv not in _DETAIL_MODES:
|
|
return _err(rid, 4002, f"unknown details_mode: {value}")
|
|
|
|
sections_cfg[section] = nv
|
|
display["sections"] = sections_cfg
|
|
cfg["display"] = display
|
|
_save_cfg(cfg)
|
|
return _ok(rid, {"key": key, "value": nv})
|
|
|
|
if key == "thinking_mode":
|
|
nv = str(value or "").strip().lower()
|
|
allowed_tm = frozenset({"collapsed", "truncated", "full"})
|
|
if nv not in allowed_tm:
|
|
return _err(rid, 4002, f"unknown thinking_mode: {value}")
|
|
_write_config_key("display.thinking_mode", nv)
|
|
# Backward compatibility bridge: keep details_mode aligned.
|
|
_write_config_key(
|
|
"display.details_mode", "expanded" if nv == "full" else "collapsed"
|
|
)
|
|
return _ok(rid, {"key": key, "value": nv})
|
|
|
|
if key == "compact":
|
|
raw = str(value or "").strip().lower()
|
|
cfg0 = _load_cfg()
|
|
d0 = cfg0.get("display") if isinstance(cfg0.get("display"), dict) else {}
|
|
cur_b = bool(d0.get("tui_compact", False))
|
|
if raw in ("", "toggle"):
|
|
nv_b = not cur_b
|
|
elif raw == "on":
|
|
nv_b = True
|
|
elif raw == "off":
|
|
nv_b = False
|
|
else:
|
|
return _err(rid, 4002, f"unknown compact value: {value}")
|
|
_write_config_key("display.tui_compact", nv_b)
|
|
return _ok(rid, {"key": key, "value": "on" if nv_b else "off"})
|
|
|
|
if key == "statusbar":
|
|
raw = str(value or "").strip().lower()
|
|
display = _load_cfg().get("display")
|
|
d0 = display if isinstance(display, dict) else {}
|
|
current = _coerce_statusbar(d0.get("tui_statusbar", "top"))
|
|
|
|
if raw in ("", "toggle"):
|
|
nv = "top" if current == "off" else "off"
|
|
elif raw == "on":
|
|
nv = "top"
|
|
elif raw in _STATUSBAR_MODES:
|
|
nv = raw
|
|
else:
|
|
return _err(rid, 4002, f"unknown statusbar value: {value}")
|
|
|
|
_write_config_key("display.tui_statusbar", nv)
|
|
return _ok(rid, {"key": key, "value": nv})
|
|
|
|
if key == "mouse":
|
|
raw = str(value or "").strip().lower()
|
|
cfg = _load_cfg()
|
|
display = cfg.get("display") if isinstance(cfg.get("display"), dict) else {}
|
|
current = _display_mouse_tracking(display)
|
|
|
|
if raw in ("", "toggle"):
|
|
nv = not current
|
|
elif raw == "on":
|
|
nv = True
|
|
elif raw == "off":
|
|
nv = False
|
|
else:
|
|
return _err(rid, 4002, f"unknown mouse value: {value}")
|
|
|
|
_write_config_key("display.mouse_tracking", nv)
|
|
return _ok(rid, {"key": key, "value": "on" if nv else "off"})
|
|
|
|
if key == "indicator":
|
|
# Use an explicit None check rather than `value or ""` so falsy
|
|
# non-string inputs (0, False, []) still surface as themselves
|
|
# in the error message instead of looking like a blank value.
|
|
raw = ("" if value is None else str(value)).strip().lower()
|
|
if raw not in _INDICATOR_STYLES:
|
|
return _err(
|
|
rid,
|
|
4002,
|
|
f"unknown indicator: {raw!r}; pick one of {'|'.join(_INDICATOR_STYLES)}",
|
|
)
|
|
_write_config_key("display.tui_status_indicator", raw)
|
|
return _ok(rid, {"key": key, "value": raw})
|
|
|
|
if key in ("prompt", "personality", "skin"):
|
|
try:
|
|
cfg = _load_cfg()
|
|
if key == "prompt":
|
|
if value == "clear":
|
|
cfg.pop("custom_prompt", None)
|
|
nv = ""
|
|
else:
|
|
cfg["custom_prompt"] = value
|
|
nv = value
|
|
_save_cfg(cfg)
|
|
elif key == "personality":
|
|
sid_key = params.get("session_id", "")
|
|
pname, new_prompt = _validate_personality(str(value or ""), cfg)
|
|
_write_config_key("display.personality", pname)
|
|
_write_config_key("agent.system_prompt", new_prompt)
|
|
nv = str(value or "default")
|
|
history_reset, info = _apply_personality_to_session(
|
|
sid_key, session, new_prompt
|
|
)
|
|
else:
|
|
_write_config_key(f"display.{key}", value)
|
|
nv = value
|
|
if key == "skin":
|
|
_emit("skin.changed", "", resolve_skin())
|
|
resp = {"key": key, "value": nv}
|
|
if key == "personality":
|
|
resp["history_reset"] = history_reset
|
|
if info is not None:
|
|
resp["info"] = info
|
|
return _ok(rid, resp)
|
|
except Exception as e:
|
|
return _err(rid, 5001, str(e))
|
|
|
|
return _err(rid, 4002, f"unknown config key: {key}")
|
|
|
|
|
|
@method("config.get")
|
|
def _(rid, params: dict) -> dict:
|
|
key = params.get("key", "")
|
|
if key == "provider":
|
|
try:
|
|
from hermes_cli.models import list_available_providers, normalize_provider
|
|
|
|
model = _resolve_model()
|
|
parts = model.split("/", 1)
|
|
return _ok(
|
|
rid,
|
|
{
|
|
"model": model,
|
|
"provider": (
|
|
normalize_provider(parts[0]) if len(parts) > 1 else "unknown"
|
|
),
|
|
"providers": list_available_providers(),
|
|
},
|
|
)
|
|
except Exception as e:
|
|
return _err(rid, 5013, str(e))
|
|
if key == "profile":
|
|
from hermes_constants import display_hermes_home
|
|
|
|
return _ok(rid, {"home": str(_hermes_home), "display": display_hermes_home()})
|
|
if key == "full":
|
|
return _ok(rid, {"config": _load_cfg()})
|
|
if key == "prompt":
|
|
return _ok(rid, {"prompt": _load_cfg().get("custom_prompt", "")})
|
|
if key == "skin":
|
|
return _ok(
|
|
rid, {"value": (_load_cfg().get("display") or {}).get("skin", "default")}
|
|
)
|
|
if key == "indicator":
|
|
# Normalize so a hand-edited config.yaml with stray casing or
|
|
# an unknown value reads back the SAME value the TUI actually
|
|
# rendered (frontend's `normalizeIndicatorStyle` falls back to
|
|
# `_INDICATOR_DEFAULT` for the same inputs). Otherwise
|
|
# `/indicator` would print one thing while the UI shows another.
|
|
raw = (_load_cfg().get("display") or {}).get("tui_status_indicator", "")
|
|
norm = str(raw).strip().lower()
|
|
return _ok(
|
|
rid,
|
|
{"value": norm if norm in _INDICATOR_STYLES else _INDICATOR_DEFAULT},
|
|
)
|
|
if key == "personality":
|
|
return _ok(
|
|
rid,
|
|
{"value": (_load_cfg().get("display") or {}).get("personality", "default")},
|
|
)
|
|
if key == "reasoning":
|
|
cfg = _load_cfg()
|
|
effort = str(
|
|
(cfg.get("agent") or {}).get("reasoning_effort", "medium") or "medium"
|
|
)
|
|
display = (
|
|
"show"
|
|
if bool((cfg.get("display") or {}).get("show_reasoning", False))
|
|
else "hide"
|
|
)
|
|
return _ok(rid, {"value": effort, "display": display})
|
|
if key == "fast":
|
|
return _ok(
|
|
rid,
|
|
{
|
|
"value": (
|
|
"fast"
|
|
if (session := _sessions.get(params.get("session_id", "")))
|
|
and getattr(session.get("agent"), "service_tier", None)
|
|
== "priority"
|
|
else ("fast" if _load_service_tier() == "priority" else "normal")
|
|
),
|
|
},
|
|
)
|
|
if key == "busy":
|
|
return _ok(rid, {"value": _load_busy_input_mode()})
|
|
if key == "details_mode":
|
|
allowed_dm = frozenset({"hidden", "collapsed", "expanded"})
|
|
raw = (
|
|
str(
|
|
(_load_cfg().get("display") or {}).get("details_mode", "collapsed")
|
|
or "collapsed"
|
|
)
|
|
.strip()
|
|
.lower()
|
|
)
|
|
nv = raw if raw in allowed_dm else "collapsed"
|
|
return _ok(rid, {"value": nv})
|
|
if key == "thinking_mode":
|
|
allowed_tm = frozenset({"collapsed", "truncated", "full"})
|
|
cfg = _load_cfg()
|
|
raw = (
|
|
str((cfg.get("display") or {}).get("thinking_mode", "") or "")
|
|
.strip()
|
|
.lower()
|
|
)
|
|
if raw in allowed_tm:
|
|
nv = raw
|
|
else:
|
|
dm = (
|
|
str(
|
|
(cfg.get("display") or {}).get("details_mode", "collapsed")
|
|
or "collapsed"
|
|
)
|
|
.strip()
|
|
.lower()
|
|
)
|
|
nv = "full" if dm == "expanded" else "collapsed"
|
|
return _ok(rid, {"value": nv})
|
|
if key == "compact":
|
|
on = bool((_load_cfg().get("display") or {}).get("tui_compact", False))
|
|
return _ok(rid, {"value": "on" if on else "off"})
|
|
if key == "statusbar":
|
|
display = _load_cfg().get("display")
|
|
raw = (
|
|
display.get("tui_statusbar", "top") if isinstance(display, dict) else "top"
|
|
)
|
|
return _ok(rid, {"value": _coerce_statusbar(raw)})
|
|
if key == "mouse":
|
|
display = _load_cfg().get("display")
|
|
on = _display_mouse_tracking(display)
|
|
return _ok(rid, {"value": "on" if on else "off"})
|
|
if key == "mtime":
|
|
cfg_path = _hermes_home / "config.yaml"
|
|
try:
|
|
return _ok(
|
|
rid, {"mtime": cfg_path.stat().st_mtime if cfg_path.exists() else 0}
|
|
)
|
|
except Exception:
|
|
return _ok(rid, {"mtime": 0})
|
|
return _err(rid, 4002, f"unknown config key: {key}")
|
|
|
|
|
|
@method("setup.status")
|
|
def _(rid, params: dict) -> dict:
|
|
try:
|
|
from hermes_cli.main import _has_any_provider_configured
|
|
|
|
return _ok(rid, {"provider_configured": bool(_has_any_provider_configured())})
|
|
except Exception as e:
|
|
return _err(rid, 5016, str(e))
|
|
|
|
|
|
# ── Methods: tools & system ──────────────────────────────────────────
|
|
|
|
|
|
@method("process.stop")
|
|
def _(rid, params: dict) -> dict:
|
|
try:
|
|
from tools.process_registry import process_registry
|
|
|
|
return _ok(rid, {"killed": process_registry.kill_all()})
|
|
except Exception as e:
|
|
return _err(rid, 5010, str(e))
|
|
|
|
|
|
@method("reload.mcp")
|
|
def _(rid, params: dict) -> dict:
|
|
session = _sessions.get(params.get("session_id", ""))
|
|
try:
|
|
# Gate: /reload-mcp invalidates the prompt cache for this session.
|
|
# Respect the ``approvals.mcp_reload_confirm`` config toggle — if
|
|
# set (default true) AND the caller did not pass ``confirm=true``
|
|
# in params, surface a warning to the transcript instead of just
|
|
# reloading silently. Users pass confirm=true either by
|
|
# re-invoking after reading the warning, or by setting the
|
|
# config key to false permanently.
|
|
user_confirm = bool(params.get("confirm", False))
|
|
if not user_confirm:
|
|
try:
|
|
from hermes_cli.config import load_config as _load_config
|
|
_cfg = _load_config()
|
|
_approvals = _cfg.get("approvals") if isinstance(_cfg, dict) else None
|
|
_confirm_required = True
|
|
if isinstance(_approvals, dict):
|
|
_confirm_required = bool(_approvals.get("mcp_reload_confirm", True))
|
|
except Exception:
|
|
_confirm_required = True
|
|
if _confirm_required:
|
|
# Return a structured response the Ink client can surface
|
|
# as a warning/confirmation without actually reloading yet.
|
|
# Ink's ops.ts reads ``status`` and prints ``message`` to
|
|
# the transcript; a follow-up invocation with confirm=true
|
|
# (or an `always` choice that flips the config) proceeds.
|
|
return _ok(rid, {
|
|
"status": "confirm_required",
|
|
"message": (
|
|
"⚠️ /reload-mcp invalidates the prompt cache (next "
|
|
"message re-sends full input tokens). Reply `/reload-mcp "
|
|
"now` to proceed, or `/reload-mcp always` to proceed and "
|
|
"silence this prompt permanently."
|
|
),
|
|
})
|
|
|
|
from tools.mcp_tool import shutdown_mcp_servers, discover_mcp_tools
|
|
|
|
shutdown_mcp_servers()
|
|
discover_mcp_tools()
|
|
if session:
|
|
agent = session["agent"]
|
|
if hasattr(agent, "refresh_tools"):
|
|
agent.refresh_tools()
|
|
_emit("session.info", params.get("session_id", ""), _session_info(agent))
|
|
|
|
# Honor `always=true` by persisting the opt-out to config.
|
|
if bool(params.get("always", False)):
|
|
try:
|
|
from cli import save_config_value as _save_cfg
|
|
_save_cfg("approvals.mcp_reload_confirm", False)
|
|
except Exception as _exc:
|
|
logger.warning("Failed to persist mcp_reload_confirm=false: %s", _exc)
|
|
|
|
return _ok(rid, {"status": "reloaded"})
|
|
except Exception as e:
|
|
return _err(rid, 5015, str(e))
|
|
|
|
|
|
@method("reload.env")
|
|
def _(rid, params: dict) -> dict:
|
|
"""Re-read ``~/.hermes/.env`` into the gateway process via
|
|
``hermes_cli.config.reload_env``, matching classic CLI's ``/reload``
|
|
handler. Newly added API keys take effect on the next agent call
|
|
without restarting the TUI.
|
|
|
|
The credential pool / provider routing for any *already-constructed*
|
|
agent does not auto-rebuild — that's the same behaviour as classic
|
|
CLI's ``/reload``. Users who want a brand-new credential resolution
|
|
should follow with ``/new``.
|
|
"""
|
|
try:
|
|
from hermes_cli.config import reload_env
|
|
|
|
count = reload_env()
|
|
return _ok(rid, {"updated": int(count)})
|
|
except Exception as e:
|
|
return _err(rid, 5015, str(e))
|
|
|
|
|
|
_TUI_HIDDEN: frozenset[str] = frozenset(
|
|
{
|
|
"sethome",
|
|
"set-home",
|
|
"update",
|
|
"commands",
|
|
"status",
|
|
"approve",
|
|
"deny",
|
|
}
|
|
)
|
|
|
|
_TUI_EXTRA: list[tuple[str, str, str]] = [
|
|
("/compact", "Toggle compact display mode", "TUI"),
|
|
("/logs", "Show recent gateway log lines", "TUI"),
|
|
("/mouse", "Toggle mouse/wheel tracking [on|off|toggle]", "TUI"),
|
|
]
|
|
|
|
# Commands that queue messages onto _pending_input in the CLI.
|
|
# In the TUI the slash worker subprocess has no reader for that queue,
|
|
# so slash.exec rejects them → TUI falls through to command.dispatch.
|
|
_PENDING_INPUT_COMMANDS: frozenset[str] = frozenset(
|
|
{
|
|
"retry",
|
|
"queue",
|
|
"q",
|
|
"steer",
|
|
"plan",
|
|
}
|
|
)
|
|
|
|
|
|
@method("commands.catalog")
|
|
def _(rid, params: dict) -> dict:
|
|
"""Registry-backed slash metadata for the TUI — categorized, no aliases."""
|
|
try:
|
|
from hermes_cli.commands import (
|
|
COMMAND_REGISTRY,
|
|
SUBCOMMANDS,
|
|
_build_description,
|
|
)
|
|
|
|
all_pairs: list[list[str]] = []
|
|
canon: dict[str, str] = {}
|
|
categories: list[dict] = []
|
|
cat_map: dict[str, list[list[str]]] = {}
|
|
cat_order: list[str] = []
|
|
|
|
for cmd in COMMAND_REGISTRY:
|
|
c = f"/{cmd.name}"
|
|
canon[c.lower()] = c
|
|
for a in cmd.aliases:
|
|
canon[f"/{a}".lower()] = c
|
|
|
|
if cmd.name in _TUI_HIDDEN:
|
|
continue
|
|
|
|
desc = _build_description(cmd)
|
|
all_pairs.append([c, desc])
|
|
|
|
cat = cmd.category
|
|
if cat not in cat_map:
|
|
cat_map[cat] = []
|
|
cat_order.append(cat)
|
|
cat_map[cat].append([c, desc])
|
|
|
|
for name, desc, cat in _TUI_EXTRA:
|
|
all_pairs.append([name, desc])
|
|
if cat not in cat_map:
|
|
cat_map[cat] = []
|
|
cat_order.append(cat)
|
|
cat_map[cat].append([name, desc])
|
|
|
|
warning = ""
|
|
try:
|
|
qcmds = _load_cfg().get("quick_commands", {}) or {}
|
|
if isinstance(qcmds, dict) and qcmds:
|
|
bucket = "User commands"
|
|
if bucket not in cat_map:
|
|
cat_map[bucket] = []
|
|
cat_order.append(bucket)
|
|
for qname, qc in sorted(qcmds.items()):
|
|
if not isinstance(qc, dict):
|
|
continue
|
|
key = f"/{qname}"
|
|
canon[key.lower()] = key
|
|
qtype = qc.get("type", "")
|
|
if qtype == "exec":
|
|
default_desc = f"exec: {qc.get('command', '')}"
|
|
elif qtype == "alias":
|
|
default_desc = f"alias → {qc.get('target', '')}"
|
|
else:
|
|
default_desc = qtype or "quick command"
|
|
qdesc = str(qc.get("description") or default_desc)
|
|
qdesc = qdesc[:120] + ("…" if len(qdesc) > 120 else "")
|
|
all_pairs.append([key, qdesc])
|
|
cat_map[bucket].append([key, qdesc])
|
|
except Exception as e:
|
|
if not warning:
|
|
warning = f"quick_commands discovery unavailable: {e}"
|
|
|
|
skill_count = 0
|
|
try:
|
|
from agent.skill_commands import scan_skill_commands
|
|
|
|
for k, info in sorted(scan_skill_commands().items()):
|
|
d = str(info.get("description", "Skill"))
|
|
all_pairs.append([k, d[:120] + ("…" if len(d) > 120 else "")])
|
|
skill_count += 1
|
|
except Exception as e:
|
|
warning = f"skill discovery unavailable: {e}"
|
|
|
|
for cat in cat_order:
|
|
categories.append({"name": cat, "pairs": cat_map[cat]})
|
|
|
|
sub = {k: v[:] for k, v in SUBCOMMANDS.items()}
|
|
return _ok(
|
|
rid,
|
|
{
|
|
"pairs": all_pairs,
|
|
"sub": sub,
|
|
"canon": canon,
|
|
"categories": categories,
|
|
"skill_count": skill_count,
|
|
"warning": warning,
|
|
},
|
|
)
|
|
except Exception as e:
|
|
return _err(rid, 5020, str(e))
|
|
|
|
|
|
def _cli_exec_blocked(argv: list[str]) -> str | None:
|
|
"""Return user hint if this argv must not run headless in the gateway process."""
|
|
if not argv:
|
|
return "bare `hermes` is interactive — use `/hermes chat -q …` or run `hermes` in another terminal"
|
|
a0 = argv[0].lower()
|
|
if a0 == "setup":
|
|
return "`hermes setup` needs a full terminal — run it outside the TUI"
|
|
if a0 == "gateway":
|
|
return "`hermes gateway` is long-running — run it in another terminal"
|
|
if a0 == "sessions" and len(argv) > 1 and argv[1].lower() == "browse":
|
|
return "`hermes sessions browse` is interactive — use /resume here, or run browse in another terminal"
|
|
if a0 == "config" and len(argv) > 1 and argv[1].lower() == "edit":
|
|
return "`hermes config edit` needs $EDITOR in a real terminal"
|
|
return None
|
|
|
|
|
|
@method("cli.exec")
|
|
def _(rid, params: dict) -> dict:
|
|
"""Run `python -m hermes_cli.main` with argv; capture stdout/stderr (non-interactive only)."""
|
|
argv = params.get("argv", [])
|
|
if not isinstance(argv, list) or not all(isinstance(x, str) for x in argv):
|
|
return _err(rid, 4003, "argv must be list[str]")
|
|
hint = _cli_exec_blocked(argv)
|
|
if hint:
|
|
return _ok(rid, {"blocked": True, "hint": hint, "code": -1, "output": ""})
|
|
try:
|
|
r = subprocess.run(
|
|
[sys.executable, "-m", "hermes_cli.main", *argv],
|
|
capture_output=True,
|
|
text=True,
|
|
timeout=min(int(params.get("timeout", 240)), 600),
|
|
cwd=os.getcwd(),
|
|
env=os.environ.copy(),
|
|
)
|
|
parts = [r.stdout or "", r.stderr or ""]
|
|
out = "\n".join(p for p in parts if p).strip() or "(no output)"
|
|
return _ok(
|
|
rid, {"blocked": False, "code": r.returncode, "output": out[:48_000]}
|
|
)
|
|
except subprocess.TimeoutExpired:
|
|
return _err(rid, 5016, "cli.exec: timeout")
|
|
except Exception as e:
|
|
return _err(rid, 5017, str(e))
|
|
|
|
|
|
@method("command.resolve")
|
|
def _(rid, params: dict) -> dict:
|
|
try:
|
|
from hermes_cli.commands import resolve_command
|
|
|
|
r = resolve_command(params.get("name", ""))
|
|
if r:
|
|
return _ok(
|
|
rid,
|
|
{
|
|
"canonical": r.name,
|
|
"description": r.description,
|
|
"category": r.category,
|
|
},
|
|
)
|
|
return _err(rid, 4011, f"unknown command: {params.get('name')}")
|
|
except Exception as e:
|
|
return _err(rid, 5012, str(e))
|
|
|
|
|
|
def _resolve_name(name: str) -> str:
|
|
try:
|
|
from hermes_cli.commands import resolve_command
|
|
|
|
r = resolve_command(name)
|
|
return r.name if r else name
|
|
except Exception:
|
|
return name
|
|
|
|
|
|
@method("command.dispatch")
|
|
def _(rid, params: dict) -> dict:
|
|
name, arg = params.get("name", "").lstrip("/"), params.get("arg", "")
|
|
resolved = _resolve_name(name)
|
|
if resolved != name:
|
|
name = resolved
|
|
session = _sessions.get(params.get("session_id", ""))
|
|
|
|
qcmds = _load_cfg().get("quick_commands", {})
|
|
if name in qcmds:
|
|
qc = qcmds[name]
|
|
if qc.get("type") == "exec":
|
|
r = subprocess.run(
|
|
qc.get("command", ""),
|
|
shell=True,
|
|
capture_output=True,
|
|
text=True,
|
|
timeout=30,
|
|
)
|
|
output = (
|
|
(r.stdout or "")
|
|
+ ("\n" if r.stdout and r.stderr else "")
|
|
+ (r.stderr or "")
|
|
).strip()[:4000]
|
|
if r.returncode != 0:
|
|
return _err(
|
|
rid,
|
|
4018,
|
|
output or f"quick command failed with exit code {r.returncode}",
|
|
)
|
|
return _ok(rid, {"type": "exec", "output": output})
|
|
if qc.get("type") == "alias":
|
|
return _ok(rid, {"type": "alias", "target": qc.get("target", "")})
|
|
|
|
try:
|
|
from hermes_cli.plugins import get_plugin_command_handler
|
|
|
|
handler = get_plugin_command_handler(name)
|
|
if handler:
|
|
return _ok(rid, {"type": "plugin", "output": str(handler(arg) or "")})
|
|
except Exception:
|
|
pass
|
|
|
|
try:
|
|
from agent.skill_commands import (
|
|
scan_skill_commands,
|
|
build_skill_invocation_message,
|
|
)
|
|
|
|
cmds = scan_skill_commands()
|
|
key = f"/{name}"
|
|
if key in cmds:
|
|
msg = build_skill_invocation_message(
|
|
key, arg, task_id=session.get("session_key", "") if session else ""
|
|
)
|
|
if msg:
|
|
return _ok(
|
|
rid,
|
|
{
|
|
"type": "skill",
|
|
"message": msg,
|
|
"name": cmds[key].get("name", name),
|
|
},
|
|
)
|
|
except Exception:
|
|
pass
|
|
|
|
# ── Commands that queue messages onto _pending_input in the CLI ───
|
|
# In the TUI the slash worker subprocess has no reader for that queue,
|
|
# so we handle them here and return a structured payload.
|
|
|
|
if name in ("queue", "q"):
|
|
if not arg:
|
|
return _err(rid, 4004, "usage: /queue <prompt>")
|
|
return _ok(rid, {"type": "send", "message": arg})
|
|
|
|
if name == "retry":
|
|
if not session:
|
|
return _err(rid, 4001, "no active session to retry")
|
|
if session.get("running"):
|
|
return _err(
|
|
rid, 4009, "session busy — /interrupt the current turn before /retry"
|
|
)
|
|
history = session.get("history", [])
|
|
if not history:
|
|
return _err(rid, 4018, "no previous user message to retry")
|
|
# Walk backwards to find the last user message
|
|
last_user_idx = None
|
|
for i in range(len(history) - 1, -1, -1):
|
|
if history[i].get("role") == "user":
|
|
last_user_idx = i
|
|
break
|
|
if last_user_idx is None:
|
|
return _err(rid, 4018, "no previous user message to retry")
|
|
content = history[last_user_idx].get("content", "")
|
|
if isinstance(content, list):
|
|
content = " ".join(
|
|
p.get("text", "")
|
|
for p in content
|
|
if isinstance(p, dict) and p.get("type") == "text"
|
|
)
|
|
if not content:
|
|
return _err(rid, 4018, "last user message is empty")
|
|
# Truncate history: remove everything from the last user message onward
|
|
# (mirrors CLI retry_last() which strips the failed exchange)
|
|
with session["history_lock"]:
|
|
session["history"] = history[:last_user_idx]
|
|
session["history_version"] = int(session.get("history_version", 0)) + 1
|
|
return _ok(rid, {"type": "send", "message": content})
|
|
|
|
if name == "steer":
|
|
if not arg:
|
|
return _err(rid, 4004, "usage: /steer <prompt>")
|
|
agent = session.get("agent") if session else None
|
|
if agent and hasattr(agent, "steer"):
|
|
try:
|
|
accepted = agent.steer(arg)
|
|
if accepted:
|
|
return _ok(
|
|
rid,
|
|
{
|
|
"type": "exec",
|
|
"output": f"⏩ Steer queued — arrives after the next tool call: {arg[:80]}{'...' if len(arg) > 80 else ''}",
|
|
},
|
|
)
|
|
except Exception:
|
|
pass
|
|
# Fallback: no active run, treat as next-turn message
|
|
return _ok(rid, {"type": "send", "message": arg})
|
|
|
|
return _err(rid, 4018, f"not a quick/plugin/skill command: {name}")
|
|
|
|
|
|
# ── Methods: paste ────────────────────────────────────────────────────
|
|
|
|
_paste_counter = 0
|
|
|
|
|
|
@method("paste.collapse")
|
|
def _(rid, params: dict) -> dict:
|
|
global _paste_counter
|
|
text = params.get("text", "")
|
|
if not text:
|
|
return _err(rid, 4004, "empty paste")
|
|
|
|
_paste_counter += 1
|
|
line_count = text.count("\n") + 1
|
|
paste_dir = _hermes_home / "pastes"
|
|
paste_dir.mkdir(parents=True, exist_ok=True)
|
|
|
|
from datetime import datetime
|
|
|
|
paste_file = (
|
|
paste_dir / f"paste_{_paste_counter}_{datetime.now().strftime('%H%M%S')}.txt"
|
|
)
|
|
paste_file.write_text(text, encoding="utf-8")
|
|
|
|
placeholder = (
|
|
f"[Pasted text #{_paste_counter}: {line_count} lines \u2192 {paste_file}]"
|
|
)
|
|
return _ok(
|
|
rid, {"placeholder": placeholder, "path": str(paste_file), "lines": line_count}
|
|
)
|
|
|
|
|
|
# ── Methods: complete ─────────────────────────────────────────────────
|
|
|
|
_FUZZY_CACHE_TTL_S = 5.0
|
|
_FUZZY_CACHE_MAX_FILES = 20000
|
|
_FUZZY_FALLBACK_EXCLUDES = frozenset(
|
|
{
|
|
".git",
|
|
".hg",
|
|
".svn",
|
|
".next",
|
|
".cache",
|
|
".venv",
|
|
"venv",
|
|
"node_modules",
|
|
"__pycache__",
|
|
"dist",
|
|
"build",
|
|
"target",
|
|
".mypy_cache",
|
|
".pytest_cache",
|
|
".ruff_cache",
|
|
}
|
|
)
|
|
_fuzzy_cache_lock = threading.Lock()
|
|
_fuzzy_cache: dict[str, tuple[float, list[str]]] = {}
|
|
|
|
|
|
def _list_repo_files(root: str) -> list[str]:
|
|
"""Return file paths relative to ``root``.
|
|
|
|
Uses ``git ls-files`` from the repo top (resolved via
|
|
``rev-parse --show-toplevel``) so the listing covers tracked + untracked
|
|
files anywhere in the repo, then converts each path back to be relative
|
|
to ``root``. Files outside ``root`` (parent directories of cwd, sibling
|
|
subtrees) are excluded so the picker stays scoped to what's reachable
|
|
from the gateway's cwd. Falls back to a bounded ``os.walk(root)`` when
|
|
``root`` isn't inside a git repo. Result cached per-root for
|
|
``_FUZZY_CACHE_TTL_S`` so rapid keystrokes don't respawn git processes.
|
|
"""
|
|
now = time.monotonic()
|
|
with _fuzzy_cache_lock:
|
|
cached = _fuzzy_cache.get(root)
|
|
if cached and now - cached[0] < _FUZZY_CACHE_TTL_S:
|
|
return cached[1]
|
|
|
|
files: list[str] = []
|
|
try:
|
|
top_result = subprocess.run(
|
|
["git", "-C", root, "rev-parse", "--show-toplevel"],
|
|
capture_output=True,
|
|
timeout=2.0,
|
|
check=False,
|
|
)
|
|
if top_result.returncode == 0:
|
|
top = top_result.stdout.decode("utf-8", "replace").strip()
|
|
list_result = subprocess.run(
|
|
[
|
|
"git",
|
|
"-C",
|
|
top,
|
|
"ls-files",
|
|
"-z",
|
|
"--cached",
|
|
"--others",
|
|
"--exclude-standard",
|
|
],
|
|
capture_output=True,
|
|
timeout=2.0,
|
|
check=False,
|
|
)
|
|
if list_result.returncode == 0:
|
|
for p in list_result.stdout.decode("utf-8", "replace").split("\0"):
|
|
if not p:
|
|
continue
|
|
rel = os.path.relpath(os.path.join(top, p), root).replace(
|
|
os.sep, "/"
|
|
)
|
|
# Skip parents/siblings of cwd — keep the picker scoped
|
|
# to root-and-below, matching Cmd-P workspace semantics.
|
|
if rel.startswith("../"):
|
|
continue
|
|
files.append(rel)
|
|
if len(files) >= _FUZZY_CACHE_MAX_FILES:
|
|
break
|
|
except (OSError, subprocess.TimeoutExpired):
|
|
pass
|
|
|
|
if not files:
|
|
# Fallback walk: skip vendor/build dirs + dot-dirs so the walk stays
|
|
# tractable. Dotfiles themselves survive — the ranker decides based
|
|
# on whether the query starts with `.`.
|
|
try:
|
|
for dirpath, dirnames, filenames in os.walk(root, followlinks=False):
|
|
dirnames[:] = [
|
|
d
|
|
for d in dirnames
|
|
if d not in _FUZZY_FALLBACK_EXCLUDES and not d.startswith(".")
|
|
]
|
|
rel_dir = os.path.relpath(dirpath, root)
|
|
for f in filenames:
|
|
rel = f if rel_dir == "." else f"{rel_dir}/{f}"
|
|
files.append(rel.replace(os.sep, "/"))
|
|
if len(files) >= _FUZZY_CACHE_MAX_FILES:
|
|
break
|
|
if len(files) >= _FUZZY_CACHE_MAX_FILES:
|
|
break
|
|
except OSError:
|
|
pass
|
|
|
|
with _fuzzy_cache_lock:
|
|
_fuzzy_cache[root] = (now, files)
|
|
|
|
return files
|
|
|
|
|
|
def _fuzzy_basename_rank(name: str, query: str) -> tuple[int, int] | None:
|
|
"""Rank ``name`` against ``query``; lower is better. Returns None to reject.
|
|
|
|
Tiers (kind):
|
|
0 — exact basename
|
|
1 — basename prefix (e.g. `app` → `appChrome.tsx`)
|
|
2 — word-boundary / camelCase hit (e.g. `chrome` → `appChrome.tsx`)
|
|
3 — substring anywhere in basename
|
|
4 — subsequence match (every query char appears in order)
|
|
|
|
Secondary key is `len(name)` so shorter names win ties.
|
|
"""
|
|
if not query:
|
|
return (3, len(name))
|
|
|
|
nl = name.lower()
|
|
ql = query.lower()
|
|
|
|
if nl == ql:
|
|
return (0, len(name))
|
|
|
|
if nl.startswith(ql):
|
|
return (1, len(name))
|
|
|
|
# Word-boundary split: `foo-bar_baz.qux` → ["foo","bar","baz","qux"].
|
|
# camelCase split: `appChrome` → ["app","Chrome"]. Cheap approximation;
|
|
# falls through to substring/subsequence if it misses.
|
|
parts: list[str] = []
|
|
buf = ""
|
|
for ch in name:
|
|
if ch in "-_." or (ch.isupper() and buf and not buf[-1].isupper()):
|
|
if buf:
|
|
parts.append(buf)
|
|
buf = ch if ch not in "-_." else ""
|
|
else:
|
|
buf += ch
|
|
if buf:
|
|
parts.append(buf)
|
|
for p in parts:
|
|
if p.lower().startswith(ql):
|
|
return (2, len(name))
|
|
|
|
if ql in nl:
|
|
return (3, len(name))
|
|
|
|
i = 0
|
|
for ch in nl:
|
|
if ch == ql[i]:
|
|
i += 1
|
|
if i == len(ql):
|
|
return (4, len(name))
|
|
|
|
return None
|
|
|
|
|
|
@method("complete.path")
|
|
def _(rid, params: dict) -> dict:
|
|
word = params.get("word", "")
|
|
if not word:
|
|
return _ok(rid, {"items": []})
|
|
|
|
items: list[dict] = []
|
|
try:
|
|
is_context = word.startswith("@")
|
|
query = word[1:] if is_context else word
|
|
|
|
if is_context and not query:
|
|
items = [
|
|
{"text": "@diff", "display": "@diff", "meta": "git diff"},
|
|
{"text": "@staged", "display": "@staged", "meta": "staged diff"},
|
|
{"text": "@file:", "display": "@file:", "meta": "attach file"},
|
|
{"text": "@folder:", "display": "@folder:", "meta": "attach folder"},
|
|
{"text": "@url:", "display": "@url:", "meta": "fetch url"},
|
|
{"text": "@git:", "display": "@git:", "meta": "git log"},
|
|
]
|
|
return _ok(rid, {"items": items})
|
|
|
|
# Accept both `@folder:path` and the bare `@folder` form so the user
|
|
# sees directory listings as soon as they finish typing the keyword,
|
|
# without first accepting the static `@folder:` hint.
|
|
if is_context and query in ("file", "folder"):
|
|
prefix_tag, path_part = query, ""
|
|
elif is_context and query.startswith(("file:", "folder:")):
|
|
prefix_tag, _, tail = query.partition(":")
|
|
path_part = tail
|
|
else:
|
|
prefix_tag = ""
|
|
path_part = query if is_context else query
|
|
|
|
# Fuzzy basename search across the repo when the user types a bare
|
|
# name with no path separator — `@appChrome` surfaces every file
|
|
# whose basename matches, regardless of directory depth. Matches what
|
|
# editors like Cursor / VS Code do for Cmd-P. Path-ish queries (with
|
|
# `/`, `./`, `~/`, `/abs`) fall through to the directory-listing
|
|
# path so explicit navigation intent is preserved.
|
|
if is_context and path_part and "/" not in path_part and prefix_tag != "folder":
|
|
root = os.getcwd()
|
|
ranked: list[tuple[tuple[int, int], str, str]] = []
|
|
for rel in _list_repo_files(root):
|
|
basename = os.path.basename(rel)
|
|
if basename.startswith(".") and not path_part.startswith("."):
|
|
continue
|
|
rank = _fuzzy_basename_rank(basename, path_part)
|
|
if rank is None:
|
|
continue
|
|
ranked.append((rank, rel, basename))
|
|
|
|
ranked.sort(key=lambda r: (r[0], len(r[1]), r[1]))
|
|
tag = prefix_tag or "file"
|
|
for _, rel, basename in ranked[:30]:
|
|
items.append(
|
|
{
|
|
"text": f"@{tag}:{rel}",
|
|
"display": basename,
|
|
"meta": os.path.dirname(rel),
|
|
}
|
|
)
|
|
|
|
return _ok(rid, {"items": items})
|
|
|
|
expanded = _normalize_completion_path(path_part) if path_part else "."
|
|
if expanded == "." or not expanded:
|
|
search_dir, match = ".", ""
|
|
elif expanded.endswith("/"):
|
|
search_dir, match = expanded, ""
|
|
else:
|
|
search_dir = os.path.dirname(expanded) or "."
|
|
match = os.path.basename(expanded)
|
|
|
|
if not os.path.isdir(search_dir):
|
|
return _ok(rid, {"items": []})
|
|
|
|
want_dir = prefix_tag == "folder"
|
|
match_lower = match.lower()
|
|
for entry in sorted(os.listdir(search_dir)):
|
|
if match and not entry.lower().startswith(match_lower):
|
|
continue
|
|
if is_context and not prefix_tag and entry.startswith("."):
|
|
continue
|
|
full = os.path.join(search_dir, entry)
|
|
is_dir = os.path.isdir(full)
|
|
# Explicit `@folder:` / `@file:` — honour the user's filter. Skip
|
|
# the opposite kind instead of auto-rewriting the completion tag,
|
|
# which used to defeat the prefix and let `@folder:` list files.
|
|
if prefix_tag and want_dir != is_dir:
|
|
continue
|
|
rel = os.path.relpath(full)
|
|
suffix = "/" if is_dir else ""
|
|
|
|
if is_context and prefix_tag:
|
|
text = f"@{prefix_tag}:{rel}{suffix}"
|
|
elif is_context:
|
|
kind = "folder" if is_dir else "file"
|
|
text = f"@{kind}:{rel}{suffix}"
|
|
elif word.startswith("~"):
|
|
text = "~/" + os.path.relpath(full, os.path.expanduser("~")) + suffix
|
|
elif word.startswith("./"):
|
|
text = "./" + rel + suffix
|
|
else:
|
|
text = rel + suffix
|
|
|
|
items.append(
|
|
{
|
|
"text": text,
|
|
"display": entry + suffix,
|
|
"meta": "dir" if is_dir else "",
|
|
}
|
|
)
|
|
if len(items) >= 30:
|
|
break
|
|
except Exception as e:
|
|
return _err(rid, 5021, str(e))
|
|
|
|
return _ok(rid, {"items": items})
|
|
|
|
|
|
def _details_completion_item(value: str, meta: str = "") -> dict:
|
|
return {"text": value, "display": value, "meta": meta}
|
|
|
|
|
|
def _details_root_completion_item(
|
|
value: str, meta: str, needs_leading_space: bool
|
|
) -> dict:
|
|
return _details_completion_item(
|
|
f" {value}" if needs_leading_space else value,
|
|
meta,
|
|
)
|
|
|
|
|
|
def _details_completions(text: str) -> list[dict] | None:
|
|
if not text.lower().startswith("/details"):
|
|
return None
|
|
|
|
stripped = text.strip()
|
|
if stripped and not "/details".startswith(stripped.lower().split()[0]):
|
|
return None
|
|
|
|
body = text[len("/details") :]
|
|
if body.startswith(" "):
|
|
body = body[1:]
|
|
parts = body.split()
|
|
has_trailing_space = text.endswith(" ")
|
|
sections = ("thinking", "tools", "subagents", "activity")
|
|
modes = ("hidden", "collapsed", "expanded")
|
|
|
|
if not body or (len(parts) == 0 and has_trailing_space):
|
|
return [
|
|
*[
|
|
_details_root_completion_item(
|
|
mode, "global mode", not has_trailing_space
|
|
)
|
|
for mode in modes
|
|
],
|
|
_details_root_completion_item(
|
|
"cycle", "cycle global mode", not has_trailing_space
|
|
),
|
|
*[
|
|
_details_root_completion_item(
|
|
section, "section override", not has_trailing_space
|
|
)
|
|
for section in sections
|
|
],
|
|
]
|
|
|
|
if len(parts) == 1 and not has_trailing_space:
|
|
prefix = parts[0].lower()
|
|
candidates = [*modes, "cycle", *sections]
|
|
return [
|
|
_details_completion_item(
|
|
candidate,
|
|
(
|
|
"section override"
|
|
if candidate in sections
|
|
else "cycle global mode" if candidate == "cycle" else "global mode"
|
|
),
|
|
)
|
|
for candidate in candidates
|
|
if candidate.startswith(prefix) and candidate != prefix
|
|
]
|
|
|
|
if len(parts) == 1 and has_trailing_space and parts[0].lower() in sections:
|
|
return [
|
|
*[
|
|
_details_completion_item(mode, f"set {parts[0].lower()}")
|
|
for mode in modes
|
|
],
|
|
_details_completion_item("reset", f"clear {parts[0].lower()} override"),
|
|
]
|
|
|
|
if len(parts) == 2 and not has_trailing_space and parts[0].lower() in sections:
|
|
prefix = parts[1].lower()
|
|
return [
|
|
_details_completion_item(
|
|
candidate,
|
|
(
|
|
f"clear {parts[0].lower()} override"
|
|
if candidate == "reset"
|
|
else f"set {parts[0].lower()}"
|
|
),
|
|
)
|
|
for candidate in (*modes, "reset")
|
|
if candidate.startswith(prefix) and candidate != prefix
|
|
]
|
|
|
|
return []
|
|
|
|
|
|
@method("complete.slash")
|
|
def _(rid, params: dict) -> dict:
|
|
text = params.get("text", "")
|
|
if not text.startswith("/"):
|
|
return _ok(rid, {"items": []})
|
|
|
|
try:
|
|
from hermes_cli.commands import SlashCommandCompleter
|
|
from prompt_toolkit.document import Document
|
|
from prompt_toolkit.formatted_text import to_plain_text
|
|
|
|
from agent.skill_commands import get_skill_commands
|
|
|
|
completer = SlashCommandCompleter(
|
|
skill_commands_provider=lambda: get_skill_commands()
|
|
)
|
|
doc = Document(text, len(text))
|
|
items = [
|
|
{
|
|
"text": c.text,
|
|
"display": c.display or c.text,
|
|
"meta": to_plain_text(c.display_meta) if c.display_meta else "",
|
|
}
|
|
for c in completer.get_completions(doc, None)
|
|
][:30]
|
|
text_lower = text.lower()
|
|
extras = [
|
|
{
|
|
"text": "/compact",
|
|
"display": "/compact",
|
|
"meta": "Toggle compact display mode",
|
|
},
|
|
{
|
|
"text": "/details",
|
|
"display": "/details",
|
|
"meta": "Control agent detail visibility",
|
|
},
|
|
{
|
|
"text": "/logs",
|
|
"display": "/logs",
|
|
"meta": "Show recent gateway log lines",
|
|
},
|
|
{
|
|
"text": "/mouse",
|
|
"display": "/mouse",
|
|
"meta": "Toggle mouse/wheel tracking [on|off|toggle]",
|
|
},
|
|
]
|
|
for extra in extras:
|
|
if extra["text"].startswith(text_lower) and not any(
|
|
item["text"] == extra["text"] for item in items
|
|
):
|
|
items.append(extra)
|
|
|
|
details_items = _details_completions(text)
|
|
if details_items is not None:
|
|
return _ok(
|
|
rid,
|
|
{
|
|
"items": details_items,
|
|
"replace_from": text.rfind(" ") + 1 if " " in text else len(text),
|
|
},
|
|
)
|
|
|
|
return _ok(
|
|
rid,
|
|
{"items": items, "replace_from": text.rfind(" ") + 1 if " " in text else 1},
|
|
)
|
|
except Exception as e:
|
|
return _err(rid, 5020, str(e))
|
|
|
|
|
|
@method("model.options")
|
|
def _(rid, params: dict) -> dict:
|
|
try:
|
|
from hermes_cli.model_switch import list_authenticated_providers
|
|
|
|
session = _sessions.get(params.get("session_id", ""))
|
|
agent = session.get("agent") if session else None
|
|
cfg = _load_cfg()
|
|
current_provider = getattr(agent, "provider", "") or ""
|
|
current_model = getattr(agent, "model", "") or _resolve_model()
|
|
current_base_url = getattr(agent, "base_url", "") or ""
|
|
# list_authenticated_providers already populates each provider's
|
|
# "models" with the curated list (same source as `hermes model` and
|
|
# classic CLI's /model picker). Do NOT overwrite with live
|
|
# provider_model_ids() — that bypasses curation and pulls in
|
|
# non-agentic models (e.g. Nous /models returns ~400 IDs including
|
|
# TTS, embeddings, rerankers, image/video generators).
|
|
providers = list_authenticated_providers(
|
|
current_provider=current_provider,
|
|
current_base_url=current_base_url,
|
|
current_model=current_model,
|
|
user_providers=(
|
|
cfg.get("providers") if isinstance(cfg.get("providers"), dict) else {}
|
|
),
|
|
custom_providers=(
|
|
cfg.get("custom_providers")
|
|
if isinstance(cfg.get("custom_providers"), list)
|
|
else []
|
|
),
|
|
max_models=50,
|
|
)
|
|
return _ok(
|
|
rid,
|
|
{
|
|
"providers": providers,
|
|
"model": current_model,
|
|
"provider": current_provider,
|
|
},
|
|
)
|
|
except Exception as e:
|
|
return _err(rid, 5033, str(e))
|
|
|
|
|
|
# ── Methods: slash.exec ──────────────────────────────────────────────
|
|
|
|
|
|
def _mirror_slash_side_effects(sid: str, session: dict, command: str) -> str:
|
|
"""Apply side effects that must also hit the gateway's live agent."""
|
|
parts = command.lstrip("/").split(None, 1)
|
|
if not parts:
|
|
return ""
|
|
name, arg, agent = (
|
|
parts[0],
|
|
(parts[1].strip() if len(parts) > 1 else ""),
|
|
session.get("agent"),
|
|
)
|
|
|
|
# Reject agent-mutating commands during an in-flight turn. These
|
|
# all do read-then-mutate on live agent/session state that the
|
|
# worker thread running agent.run_conversation is using. Parity
|
|
# with the session.compress / session.undo guards and the gateway
|
|
# runner's running-agent /model guard.
|
|
_MUTATES_WHILE_RUNNING = {"model", "personality", "prompt", "compress"}
|
|
if name in _MUTATES_WHILE_RUNNING and session.get("running"):
|
|
return f"session busy — /interrupt the current turn before running /{name}"
|
|
|
|
try:
|
|
if name == "model" and arg and agent:
|
|
result = _apply_model_switch(sid, session, arg)
|
|
return result.get("warning", "")
|
|
elif name == "personality" and arg and agent:
|
|
_, new_prompt = _validate_personality(arg, _load_cfg())
|
|
_apply_personality_to_session(sid, session, new_prompt)
|
|
elif name == "prompt" and agent:
|
|
cfg = _load_cfg()
|
|
new_prompt = (cfg.get("agent") or {}).get("system_prompt", "") or ""
|
|
agent.ephemeral_system_prompt = new_prompt or None
|
|
agent._cached_system_prompt = None
|
|
elif name == "compress" and agent:
|
|
_compress_session_history(session, arg)
|
|
_sync_session_key_after_compress(sid, session)
|
|
_emit("session.info", sid, _session_info(agent))
|
|
elif name == "fast" and agent:
|
|
mode = arg.lower()
|
|
if mode in {"fast", "on"}:
|
|
agent.service_tier = "priority"
|
|
elif mode in {"normal", "off"}:
|
|
agent.service_tier = None
|
|
_emit("session.info", sid, _session_info(agent))
|
|
elif name == "reload-mcp" and agent and hasattr(agent, "reload_mcp_tools"):
|
|
agent.reload_mcp_tools()
|
|
elif name == "stop":
|
|
from tools.process_registry import process_registry
|
|
|
|
process_registry.kill_all()
|
|
except Exception as e:
|
|
return f"live session sync failed: {e}"
|
|
return ""
|
|
|
|
|
|
@method("slash.exec")
|
|
def _(rid, params: dict) -> dict:
|
|
session, err = _sess(params, rid)
|
|
if err:
|
|
return err
|
|
|
|
cmd = params.get("command", "").strip()
|
|
if not cmd:
|
|
return _err(rid, 4004, "empty command")
|
|
|
|
# Skill slash commands and _pending_input commands must NOT go through the
|
|
# slash worker — see _PENDING_INPUT_COMMANDS definition above.
|
|
_cmd_parts = cmd.split() if not cmd.startswith("/") else cmd.lstrip("/").split()
|
|
_cmd_base = _cmd_parts[0] if _cmd_parts else ""
|
|
|
|
if _cmd_base in _PENDING_INPUT_COMMANDS:
|
|
return _err(
|
|
rid, 4018, f"pending-input command: use command.dispatch for /{_cmd_base}"
|
|
)
|
|
|
|
try:
|
|
from agent.skill_commands import get_skill_commands
|
|
|
|
_cmd_key = f"/{_cmd_base}"
|
|
if _cmd_key in get_skill_commands():
|
|
return _err(
|
|
rid, 4018, f"skill command: use command.dispatch for {_cmd_key}"
|
|
)
|
|
except Exception:
|
|
pass
|
|
|
|
worker = session.get("slash_worker")
|
|
if not worker:
|
|
try:
|
|
worker = _SlashWorker(
|
|
session["session_key"],
|
|
getattr(session.get("agent"), "model", _resolve_model()),
|
|
)
|
|
session["slash_worker"] = worker
|
|
except Exception as e:
|
|
return _err(rid, 5030, f"slash worker start failed: {e}")
|
|
|
|
try:
|
|
output = worker.run(cmd)
|
|
warning = _mirror_slash_side_effects(params.get("session_id", ""), session, cmd)
|
|
payload = {"output": output or "(no output)"}
|
|
if warning:
|
|
payload["warning"] = warning
|
|
return _ok(rid, payload)
|
|
except Exception as e:
|
|
try:
|
|
worker.close()
|
|
except Exception:
|
|
pass
|
|
session["slash_worker"] = None
|
|
return _err(rid, 5030, str(e))
|
|
|
|
|
|
# ── Methods: voice ───────────────────────────────────────────────────
|
|
|
|
|
|
_voice_sid_lock = threading.Lock()
|
|
_voice_event_sid: str = ""
|
|
|
|
|
|
def _voice_emit(event: str, payload: dict | None = None) -> None:
|
|
"""Emit a voice event toward the session that most recently turned the
|
|
mode on. Voice is process-global (one microphone), so there's only ever
|
|
one sid to target; the TUI handler treats an empty sid as "active
|
|
session". Kept separate from _emit to make the lack of per-call sid
|
|
argument explicit."""
|
|
with _voice_sid_lock:
|
|
sid = _voice_event_sid
|
|
_emit(event, sid, payload)
|
|
|
|
|
|
def _voice_mode_enabled() -> bool:
|
|
"""Current voice-mode flag (runtime-only, CLI parity).
|
|
|
|
cli.py initialises ``_voice_mode = False`` at startup and only flips
|
|
it via ``/voice on``; it never reads a persisted enable bit from
|
|
config.yaml. We match that: no config lookup, env var only. This
|
|
avoids the TUI auto-starting in REC the next time the user opens it
|
|
just because they happened to enable voice in a prior session.
|
|
"""
|
|
return os.environ.get("HERMES_VOICE", "").strip() == "1"
|
|
|
|
|
|
def _voice_tts_enabled() -> bool:
|
|
"""Whether agent replies should be spoken back via TTS (runtime only)."""
|
|
return os.environ.get("HERMES_VOICE_TTS", "").strip() == "1"
|
|
|
|
|
|
@method("voice.toggle")
|
|
def _(rid, params: dict) -> dict:
|
|
"""CLI parity for the ``/voice`` slash command.
|
|
|
|
Subcommands:
|
|
|
|
* ``status`` — report mode + TTS flags (default when action is unknown).
|
|
* ``on`` / ``off`` — flip voice *mode* (the umbrella bit). Turning it
|
|
off also tears down any active continuous recording loop. Does NOT
|
|
start recording on its own; recording is driven by ``voice.record``
|
|
(Ctrl+B) after mode is on, matching cli.py's enable/Ctrl+B split.
|
|
* ``tts`` — toggle speech-output of agent replies. Requires mode on
|
|
(mirrors CLI's _toggle_voice_tts guard).
|
|
"""
|
|
action = params.get("action", "status")
|
|
|
|
if action == "status":
|
|
# Mirror CLI's _show_voice_status: include STT/TTS provider
|
|
# availability so the user can tell at a glance *why* voice mode
|
|
# isn't working ("STT provider: MISSING ..." is the common case).
|
|
payload: dict = {
|
|
"enabled": _voice_mode_enabled(),
|
|
"tts": _voice_tts_enabled(),
|
|
}
|
|
try:
|
|
from tools.voice_mode import check_voice_requirements
|
|
|
|
reqs = check_voice_requirements()
|
|
payload["available"] = bool(reqs.get("available"))
|
|
payload["audio_available"] = bool(reqs.get("audio_available"))
|
|
payload["stt_available"] = bool(reqs.get("stt_available"))
|
|
payload["details"] = reqs.get("details") or ""
|
|
except Exception as e:
|
|
# check_voice_requirements pulls optional transcription deps —
|
|
# swallow so /voice status always returns something useful.
|
|
logger.warning("voice.toggle status: requirements probe failed: %s", e)
|
|
|
|
return _ok(rid, payload)
|
|
|
|
if action in ("on", "off"):
|
|
enabled = action == "on"
|
|
# Runtime-only flag (CLI parity) — no _write_config_key, so the
|
|
# next TUI launch starts with voice OFF instead of auto-REC from a
|
|
# persisted stale toggle.
|
|
os.environ["HERMES_VOICE"] = "1" if enabled else "0"
|
|
|
|
if not enabled:
|
|
# Disabling the mode must tear the continuous loop down; the
|
|
# loop holds the microphone and would otherwise keep running.
|
|
try:
|
|
from hermes_cli.voice import stop_continuous
|
|
|
|
stop_continuous()
|
|
except ImportError:
|
|
pass
|
|
except Exception as e:
|
|
logger.warning("voice: stop_continuous failed during toggle off: %s", e)
|
|
|
|
return _ok(rid, {"enabled": enabled, "tts": _voice_tts_enabled()})
|
|
|
|
if action == "tts":
|
|
if not _voice_mode_enabled():
|
|
return _err(rid, 4014, "enable voice mode first: /voice on")
|
|
new_value = not _voice_tts_enabled()
|
|
# Runtime-only flag (CLI parity) — see voice.toggle on/off above.
|
|
os.environ["HERMES_VOICE_TTS"] = "1" if new_value else "0"
|
|
return _ok(rid, {"enabled": True, "tts": new_value})
|
|
|
|
return _err(rid, 4013, f"unknown voice action: {action}")
|
|
|
|
|
|
@method("voice.record")
|
|
def _(rid, params: dict) -> dict:
|
|
"""VAD-driven continuous record loop, CLI-parity.
|
|
|
|
``start`` turns on a VAD loop that emits ``voice.transcript`` events
|
|
for each detected utterance and auto-restarts for the next turn.
|
|
``stop`` halts the loop (manual stop; matches cli.py's Ctrl+B-while-
|
|
recording branch clearing ``_voice_continuous``). Three consecutive
|
|
silent cycles stop the loop automatically and emit a
|
|
``voice.transcript`` with ``no_speech_limit=True``.
|
|
"""
|
|
action = params.get("action", "start")
|
|
|
|
if action not in {"start", "stop"}:
|
|
return _err(rid, 4019, f"unknown voice action: {action}")
|
|
|
|
try:
|
|
if action == "start":
|
|
if not _voice_mode_enabled():
|
|
return _err(rid, 4015, "voice mode is off — enable with /voice on")
|
|
|
|
with _voice_sid_lock:
|
|
global _voice_event_sid
|
|
_voice_event_sid = params.get("session_id") or _voice_event_sid
|
|
|
|
from hermes_cli.voice import start_continuous
|
|
|
|
voice_cfg = _load_cfg().get("voice", {})
|
|
start_continuous(
|
|
on_transcript=lambda t: _voice_emit("voice.transcript", {"text": t}),
|
|
on_status=lambda s: _voice_emit("voice.status", {"state": s}),
|
|
on_silent_limit=lambda: _voice_emit(
|
|
"voice.transcript", {"no_speech_limit": True}
|
|
),
|
|
silence_threshold=voice_cfg.get("silence_threshold", 200),
|
|
silence_duration=voice_cfg.get("silence_duration", 3.0),
|
|
)
|
|
return _ok(rid, {"status": "recording"})
|
|
|
|
# action == "stop"
|
|
from hermes_cli.voice import stop_continuous
|
|
|
|
stop_continuous()
|
|
return _ok(rid, {"status": "stopped"})
|
|
except ImportError:
|
|
return _err(
|
|
rid, 5025, "voice module not available — install audio dependencies"
|
|
)
|
|
except Exception as e:
|
|
return _err(rid, 5025, str(e))
|
|
|
|
|
|
@method("voice.tts")
|
|
def _(rid, params: dict) -> dict:
|
|
text = params.get("text", "")
|
|
if not text:
|
|
return _err(rid, 4020, "text required")
|
|
try:
|
|
from hermes_cli.voice import speak_text
|
|
|
|
threading.Thread(target=speak_text, args=(text,), daemon=True).start()
|
|
return _ok(rid, {"status": "speaking"})
|
|
except ImportError:
|
|
return _err(rid, 5026, "voice module not available")
|
|
except Exception as e:
|
|
return _err(rid, 5026, str(e))
|
|
|
|
|
|
# ── Methods: insights ────────────────────────────────────────────────
|
|
|
|
|
|
@method("insights.get")
|
|
def _(rid, params: dict) -> dict:
|
|
days = params.get("days", 30)
|
|
db = _get_db()
|
|
if db is None:
|
|
return _db_unavailable_error(rid, code=5017)
|
|
try:
|
|
cutoff = time.time() - days * 86400
|
|
rows = [
|
|
s
|
|
for s in db.list_sessions_rich(limit=500)
|
|
if (s.get("started_at") or 0) >= cutoff
|
|
]
|
|
return _ok(
|
|
rid,
|
|
{
|
|
"days": days,
|
|
"sessions": len(rows),
|
|
"messages": sum(s.get("message_count", 0) for s in rows),
|
|
},
|
|
)
|
|
except Exception as e:
|
|
return _err(rid, 5017, str(e))
|
|
|
|
|
|
# ── Methods: rollback ────────────────────────────────────────────────
|
|
|
|
|
|
@method("rollback.list")
|
|
def _(rid, params: dict) -> dict:
|
|
session, err = _sess(params, rid)
|
|
if err:
|
|
return err
|
|
try:
|
|
|
|
def go(mgr, cwd):
|
|
if not mgr.enabled:
|
|
return _ok(rid, {"enabled": False, "checkpoints": []})
|
|
return _ok(
|
|
rid,
|
|
{
|
|
"enabled": True,
|
|
"checkpoints": [
|
|
{
|
|
"hash": c.get("hash", ""),
|
|
"timestamp": c.get("timestamp", ""),
|
|
"message": c.get("message", ""),
|
|
}
|
|
for c in mgr.list_checkpoints(cwd)
|
|
],
|
|
},
|
|
)
|
|
|
|
return _with_checkpoints(session, go)
|
|
except Exception as e:
|
|
return _err(rid, 5020, str(e))
|
|
|
|
|
|
@method("rollback.restore")
|
|
def _(rid, params: dict) -> dict:
|
|
session, err = _sess(params, rid)
|
|
if err:
|
|
return err
|
|
target = params.get("hash", "")
|
|
file_path = params.get("file_path", "")
|
|
if not target:
|
|
return _err(rid, 4014, "hash required")
|
|
# Full-history rollback mutates session history. Rejecting during
|
|
# an in-flight turn prevents prompt.submit from silently dropping
|
|
# the agent's output (version mismatch path) or clobbering the
|
|
# rollback (version-matches path). A file-scoped rollback only
|
|
# touches disk, so we allow it.
|
|
if not file_path and session.get("running"):
|
|
return _err(
|
|
rid,
|
|
4009,
|
|
"session busy — /interrupt the current turn before full rollback.restore",
|
|
)
|
|
try:
|
|
|
|
def go(mgr, cwd):
|
|
resolved = _resolve_checkpoint_hash(mgr, cwd, target)
|
|
result = mgr.restore(cwd, resolved, file_path=file_path or None)
|
|
if result.get("success") and not file_path:
|
|
removed = 0
|
|
with session["history_lock"]:
|
|
history = session.get("history", [])
|
|
while history and history[-1].get("role") in ("assistant", "tool"):
|
|
history.pop()
|
|
removed += 1
|
|
if history and history[-1].get("role") == "user":
|
|
history.pop()
|
|
removed += 1
|
|
if removed:
|
|
session["history_version"] = (
|
|
int(session.get("history_version", 0)) + 1
|
|
)
|
|
result["history_removed"] = removed
|
|
return result
|
|
|
|
return _ok(rid, _with_checkpoints(session, go))
|
|
except Exception as e:
|
|
return _err(rid, 5021, str(e))
|
|
|
|
|
|
@method("rollback.diff")
|
|
def _(rid, params: dict) -> dict:
|
|
session, err = _sess(params, rid)
|
|
if err:
|
|
return err
|
|
target = params.get("hash", "")
|
|
if not target:
|
|
return _err(rid, 4014, "hash required")
|
|
try:
|
|
r = _with_checkpoints(
|
|
session,
|
|
lambda mgr, cwd: mgr.diff(cwd, _resolve_checkpoint_hash(mgr, cwd, target)),
|
|
)
|
|
raw = r.get("diff", "")[:4000]
|
|
payload = {"stat": r.get("stat", ""), "diff": raw}
|
|
rendered = render_diff(raw, session.get("cols", 80))
|
|
if rendered:
|
|
payload["rendered"] = rendered
|
|
return _ok(rid, payload)
|
|
except Exception as e:
|
|
return _err(rid, 5022, str(e))
|
|
|
|
|
|
# ── Methods: browser / plugins / cron / skills ───────────────────────
|
|
|
|
|
|
def _resolve_browser_cdp_url() -> str:
|
|
"""Return the configured browser CDP override without network I/O.
|
|
|
|
``/browser status`` must be fast — calling
|
|
``tools.browser_tool._get_cdp_override`` would invoke
|
|
``_resolve_cdp_override``, which performs an HTTP probe to
|
|
``.../json/version`` for discovery-style URLs. That probe has
|
|
a multi-second timeout and would block the TUI on a slow or
|
|
unreachable host even though status only needs to report whether
|
|
an override is set.
|
|
|
|
Mirrors the env/config precedence of ``_get_cdp_override`` (env
|
|
var first, then ``browser.cdp_url`` from config.yaml) without the
|
|
websocket-resolution step, so the answer reflects user intent
|
|
even when the configured host is not currently reachable. The
|
|
actual WS normalization happens in ``browser_navigate`` on the
|
|
next tool call.
|
|
"""
|
|
env_url = os.environ.get("BROWSER_CDP_URL", "").strip()
|
|
if env_url:
|
|
return env_url
|
|
try:
|
|
from hermes_cli.config import read_raw_config
|
|
|
|
cfg = read_raw_config()
|
|
browser_cfg = cfg.get("browser", {}) if isinstance(cfg, dict) else {}
|
|
if isinstance(browser_cfg, dict):
|
|
return str(browser_cfg.get("cdp_url", "") or "").strip()
|
|
except Exception:
|
|
pass
|
|
return ""
|
|
|
|
|
|
def _is_default_local_cdp(parsed) -> bool:
|
|
"""Match the discovery-style local default; never the concrete WS form.
|
|
|
|
A user-supplied ``ws://127.0.0.1:9222/devtools/browser/<id>`` is a
|
|
real, connectable endpoint — collapsing it to bare ``http://...:9222``
|
|
would strip the path and break the connect.
|
|
"""
|
|
try:
|
|
port = parsed.port or 80
|
|
except ValueError:
|
|
return False
|
|
|
|
discovery_path = parsed.path in {"", "/", "/json", "/json/version"}
|
|
return (
|
|
parsed.scheme in {"http", "ws"}
|
|
and parsed.hostname in {"127.0.0.1", "localhost"}
|
|
and port == 9222
|
|
and discovery_path
|
|
)
|
|
|
|
|
|
def _http_ok(url: str, timeout: float) -> bool:
|
|
import urllib.request
|
|
|
|
try:
|
|
with urllib.request.urlopen(url, timeout=timeout) as resp:
|
|
return 200 <= getattr(resp, "status", 200) < 300
|
|
except Exception:
|
|
return False
|
|
|
|
|
|
def _probe_urls(parsed) -> list[str]:
|
|
scheme = {"ws": "http", "wss": "https"}.get(parsed.scheme, parsed.scheme)
|
|
root = f"{scheme}://{parsed.netloc}".rstrip("/")
|
|
return [f"{root}/json/version", f"{root}/json"]
|
|
|
|
|
|
def _normalize_cdp_url(parsed) -> str:
|
|
# Concrete ``/devtools/browser/<id>`` endpoints (Browserbase et al.)
|
|
# are connectable as-is. Discovery-style inputs collapse to bare
|
|
# ``scheme://host:port`` so ``_resolve_cdp_override`` can append
|
|
# ``/json/version`` later without doubling the path.
|
|
if parsed.path.startswith("/devtools/browser/"):
|
|
return parsed.geturl()
|
|
return parsed._replace(path="", params="", query="", fragment="").geturl()
|
|
|
|
|
|
def _failure_messages(url: str, port: int, system: str) -> list[str]:
|
|
from hermes_cli.browser_connect import manual_chrome_debug_command
|
|
|
|
command = manual_chrome_debug_command(port, system)
|
|
hint = (
|
|
["Start Chrome with remote debugging, then retry /browser connect:", command]
|
|
if command
|
|
else [
|
|
"No Chrome/Chromium executable was found in this environment.",
|
|
f"Install one or start Chrome with --remote-debugging-port={port}, then retry /browser connect.",
|
|
]
|
|
)
|
|
return [
|
|
f"Chrome is not reachable at {url}.",
|
|
*hint,
|
|
"Browser not connected — start Chrome with remote debugging and retry /browser connect",
|
|
]
|
|
|
|
|
|
@method("browser.manage")
|
|
def _(rid, params: dict) -> dict:
|
|
action = params.get("action", "status")
|
|
|
|
if action == "status":
|
|
url = _resolve_browser_cdp_url()
|
|
return _ok(rid, {"connected": bool(url), "url": url})
|
|
|
|
if action == "disconnect":
|
|
return _browser_disconnect(rid)
|
|
|
|
if action != "connect":
|
|
return _err(rid, 4015, f"unknown action: {action}")
|
|
|
|
return _browser_connect(rid, params)
|
|
|
|
|
|
def _browser_connect(rid, params: dict) -> dict:
|
|
import platform
|
|
|
|
from hermes_cli.browser_connect import DEFAULT_BROWSER_CDP_URL
|
|
from tools.browser_tool import cleanup_all_browsers
|
|
from urllib.parse import urlparse
|
|
|
|
raw_url = params.get("url")
|
|
if raw_url is not None and not isinstance(raw_url, str):
|
|
return _err(rid, 4015, f"browser url must be a string, got {type(raw_url).__name__}")
|
|
url = (raw_url or "").strip() or DEFAULT_BROWSER_CDP_URL
|
|
|
|
sid = params.get("session_id") or ""
|
|
system = platform.system()
|
|
messages: list[str] = []
|
|
|
|
def announce(message: str, *, level: str = "info") -> None:
|
|
messages.append(message)
|
|
# Without a session id the TUI prints `messages` from the
|
|
# response; emitting an event would double-render. Only stream
|
|
# progress when there's a real session to scope it to.
|
|
if sid:
|
|
_emit("browser.progress", sid, {"message": message, "level": level})
|
|
|
|
parsed = urlparse(url if "://" in url else f"http://{url}")
|
|
if parsed.scheme not in {"http", "https", "ws", "wss"}:
|
|
return _err(rid, 4015, f"unsupported browser url: {url}")
|
|
if not parsed.hostname:
|
|
return _err(rid, 4015, f"missing host in browser url: {url}")
|
|
try:
|
|
port = parsed.port or (443 if parsed.scheme in {"https", "wss"} else 80)
|
|
except ValueError:
|
|
return _err(rid, 4015, f"invalid port in browser url: {url}")
|
|
|
|
# Always normalize default-local to 127.0.0.1:9222 so downstream
|
|
# comparisons + messaging match what we'll actually persist.
|
|
if _is_default_local_cdp(parsed):
|
|
url = DEFAULT_BROWSER_CDP_URL
|
|
parsed = urlparse(url)
|
|
port = parsed.port or 9222
|
|
|
|
try:
|
|
# ws[s]://.../devtools/browser/<id> endpoints (hosted CDP
|
|
# providers) don't serve the HTTP discovery path; just check
|
|
# TCP-level reachability and let browser_navigate handshake.
|
|
if parsed.scheme in {"ws", "wss"} and parsed.path.startswith(
|
|
"/devtools/browser/"
|
|
):
|
|
import socket
|
|
|
|
try:
|
|
with socket.create_connection((parsed.hostname, port), timeout=2.0):
|
|
pass
|
|
except OSError as e:
|
|
return _err(rid, 5031, f"could not reach browser CDP at {url}: {e}")
|
|
else:
|
|
probes = _probe_urls(parsed)
|
|
ok = any(_http_ok(p, timeout=2.0) for p in probes)
|
|
|
|
if not ok and _is_default_local_cdp(parsed):
|
|
from hermes_cli.browser_connect import try_launch_chrome_debug
|
|
|
|
announce(
|
|
"Chrome isn't running with remote debugging — attempting to launch..."
|
|
)
|
|
|
|
if try_launch_chrome_debug(port, system):
|
|
for _ in range(20):
|
|
time.sleep(0.5)
|
|
if any(_http_ok(p, timeout=1.0) for p in probes):
|
|
ok = True
|
|
break
|
|
|
|
if ok:
|
|
announce(f"Chrome launched and listening on port {port}")
|
|
else:
|
|
for line in _failure_messages(url, port, system)[1:]:
|
|
announce(line, level="error")
|
|
return _ok(
|
|
rid, {"connected": False, "url": url, "messages": messages}
|
|
)
|
|
elif not ok:
|
|
return _err(rid, 5031, f"could not reach browser CDP at {url}")
|
|
elif _is_default_local_cdp(parsed):
|
|
announce(f"Chrome is already listening on port {port}")
|
|
|
|
normalized = _normalize_cdp_url(parsed)
|
|
|
|
# Order matters: reap sessions BEFORE publishing the new env
|
|
# so an in-flight tool call sees the old supervisor closed,
|
|
# then again AFTER so the default task's cached supervisor
|
|
# is drained against the new URL.
|
|
cleanup_all_browsers()
|
|
os.environ["BROWSER_CDP_URL"] = normalized
|
|
cleanup_all_browsers()
|
|
except Exception as e:
|
|
return _err(rid, 5031, str(e))
|
|
|
|
payload: dict[str, object] = {"connected": True, "url": normalized}
|
|
if messages:
|
|
payload["messages"] = messages
|
|
return _ok(rid, payload)
|
|
|
|
|
|
def _browser_disconnect(rid) -> dict:
|
|
# Reap, drop the env override, reap again — closes the same swap
|
|
# window covered by ``_browser_connect``.
|
|
def reap() -> None:
|
|
try:
|
|
from tools.browser_tool import cleanup_all_browsers
|
|
|
|
cleanup_all_browsers()
|
|
except Exception:
|
|
pass
|
|
|
|
reap()
|
|
os.environ.pop("BROWSER_CDP_URL", None)
|
|
reap()
|
|
return _ok(rid, {"connected": False})
|
|
|
|
|
|
@method("plugins.list")
|
|
def _(rid, params: dict) -> dict:
|
|
try:
|
|
from hermes_cli.plugins import get_plugin_manager
|
|
|
|
return _ok(
|
|
rid,
|
|
{
|
|
"plugins": [
|
|
{
|
|
"name": n,
|
|
"version": getattr(i, "version", "?"),
|
|
"enabled": getattr(i, "enabled", True),
|
|
}
|
|
for n, i in get_plugin_manager()._plugins.items()
|
|
]
|
|
},
|
|
)
|
|
except Exception as e:
|
|
return _err(rid, 5032, str(e))
|
|
|
|
|
|
@method("config.show")
|
|
def _(rid, params: dict) -> dict:
|
|
try:
|
|
cfg = _load_cfg()
|
|
model = _resolve_model()
|
|
api_key = os.environ.get("HERMES_API_KEY", "") or cfg.get("api_key", "")
|
|
masked = f"****{api_key[-4:]}" if len(api_key) > 4 else "(not set)"
|
|
base_url = os.environ.get("HERMES_BASE_URL", "") or cfg.get("base_url", "")
|
|
|
|
sections = [
|
|
{
|
|
"title": "Model",
|
|
"rows": [
|
|
["Model", model],
|
|
["Base URL", base_url or "(default)"],
|
|
["API Key", masked],
|
|
],
|
|
},
|
|
{
|
|
"title": "Agent",
|
|
"rows": [
|
|
["Max Turns", str(cfg.get("max_turns", 25))],
|
|
["Toolsets", ", ".join(cfg.get("enabled_toolsets", [])) or "all"],
|
|
["Verbose", str(cfg.get("verbose", False))],
|
|
],
|
|
},
|
|
{
|
|
"title": "Environment",
|
|
"rows": [
|
|
["Working Dir", os.getcwd()],
|
|
["Config File", str(_hermes_home / "config.yaml")],
|
|
],
|
|
},
|
|
]
|
|
return _ok(rid, {"sections": sections})
|
|
except Exception as e:
|
|
return _err(rid, 5030, str(e))
|
|
|
|
|
|
@method("tools.list")
|
|
def _(rid, params: dict) -> dict:
|
|
try:
|
|
from toolsets import get_all_toolsets, get_toolset_info
|
|
|
|
session = _sessions.get(params.get("session_id", ""))
|
|
enabled = (
|
|
set(getattr(session["agent"], "enabled_toolsets", []) or [])
|
|
if session
|
|
else set(_load_enabled_toolsets() or [])
|
|
)
|
|
|
|
items = []
|
|
for name in sorted(get_all_toolsets().keys()):
|
|
info = get_toolset_info(name)
|
|
if not info:
|
|
continue
|
|
items.append(
|
|
{
|
|
"name": name,
|
|
"description": info["description"],
|
|
"tool_count": info["tool_count"],
|
|
"enabled": name in enabled if enabled else True,
|
|
"tools": info["resolved_tools"],
|
|
}
|
|
)
|
|
return _ok(rid, {"toolsets": items})
|
|
except Exception as e:
|
|
return _err(rid, 5031, str(e))
|
|
|
|
|
|
@method("tools.show")
|
|
def _(rid, params: dict) -> dict:
|
|
try:
|
|
from model_tools import get_toolset_for_tool, get_tool_definitions
|
|
|
|
session = _sessions.get(params.get("session_id", ""))
|
|
enabled = (
|
|
getattr(session["agent"], "enabled_toolsets", None)
|
|
if session
|
|
else _load_enabled_toolsets()
|
|
)
|
|
tools = get_tool_definitions(enabled_toolsets=enabled, quiet_mode=True)
|
|
sections = {}
|
|
|
|
for tool in sorted(tools, key=lambda t: t["function"]["name"]):
|
|
name = tool["function"]["name"]
|
|
desc = str(tool["function"].get("description", "") or "").split("\n")[0]
|
|
if ". " in desc:
|
|
desc = desc[: desc.index(". ") + 1]
|
|
sections.setdefault(get_toolset_for_tool(name) or "unknown", []).append(
|
|
{
|
|
"name": name,
|
|
"description": desc,
|
|
}
|
|
)
|
|
|
|
return _ok(
|
|
rid,
|
|
{
|
|
"sections": [
|
|
{"name": name, "tools": rows}
|
|
for name, rows in sorted(sections.items())
|
|
],
|
|
"total": len(tools),
|
|
},
|
|
)
|
|
except Exception as e:
|
|
return _err(rid, 5034, str(e))
|
|
|
|
|
|
@method("tools.configure")
|
|
def _(rid, params: dict) -> dict:
|
|
action = str(params.get("action", "") or "").strip().lower()
|
|
targets = [
|
|
str(name).strip() for name in params.get("names", []) or [] if str(name).strip()
|
|
]
|
|
if action not in {"disable", "enable"}:
|
|
return _err(rid, 4017, f"unknown tools action: {action}")
|
|
if not targets:
|
|
return _err(rid, 4018, "names required")
|
|
|
|
try:
|
|
from hermes_cli.config import load_config, save_config
|
|
from hermes_cli.tools_config import (
|
|
CONFIGURABLE_TOOLSETS,
|
|
_apply_mcp_change,
|
|
_apply_toolset_change,
|
|
_get_platform_tools,
|
|
_get_plugin_toolset_keys,
|
|
)
|
|
|
|
cfg = load_config()
|
|
valid_toolsets = {
|
|
ts_key for ts_key, _, _ in CONFIGURABLE_TOOLSETS
|
|
} | _get_plugin_toolset_keys()
|
|
toolset_targets = [name for name in targets if ":" not in name]
|
|
mcp_targets = [name for name in targets if ":" in name]
|
|
unknown = [name for name in toolset_targets if name not in valid_toolsets]
|
|
toolset_targets = [name for name in toolset_targets if name in valid_toolsets]
|
|
|
|
if toolset_targets:
|
|
_apply_toolset_change(cfg, "cli", toolset_targets, action)
|
|
|
|
missing_servers = (
|
|
_apply_mcp_change(cfg, mcp_targets, action) if mcp_targets else set()
|
|
)
|
|
save_config(cfg)
|
|
|
|
session = _sessions.get(params.get("session_id", ""))
|
|
info = (
|
|
_reset_session_agent(params.get("session_id", ""), session)
|
|
if session
|
|
else None
|
|
)
|
|
enabled = sorted(
|
|
_get_platform_tools(load_config(), "cli", include_default_mcp_servers=False)
|
|
)
|
|
changed = [
|
|
name
|
|
for name in targets
|
|
if name not in unknown
|
|
and (":" not in name or name.split(":", 1)[0] not in missing_servers)
|
|
]
|
|
|
|
return _ok(
|
|
rid,
|
|
{
|
|
"changed": changed,
|
|
"enabled_toolsets": enabled,
|
|
"info": info,
|
|
"missing_servers": sorted(missing_servers),
|
|
"reset": bool(session),
|
|
"unknown": unknown,
|
|
},
|
|
)
|
|
except Exception as e:
|
|
return _err(rid, 5035, str(e))
|
|
|
|
|
|
@method("toolsets.list")
|
|
def _(rid, params: dict) -> dict:
|
|
try:
|
|
from toolsets import get_all_toolsets, get_toolset_info
|
|
|
|
session = _sessions.get(params.get("session_id", ""))
|
|
enabled = (
|
|
set(getattr(session["agent"], "enabled_toolsets", []) or [])
|
|
if session
|
|
else set(_load_enabled_toolsets() or [])
|
|
)
|
|
|
|
items = []
|
|
for name in sorted(get_all_toolsets().keys()):
|
|
info = get_toolset_info(name)
|
|
if not info:
|
|
continue
|
|
items.append(
|
|
{
|
|
"name": name,
|
|
"description": info["description"],
|
|
"tool_count": info["tool_count"],
|
|
"enabled": name in enabled if enabled else True,
|
|
}
|
|
)
|
|
return _ok(rid, {"toolsets": items})
|
|
except Exception as e:
|
|
return _err(rid, 5032, str(e))
|
|
|
|
|
|
@method("agents.list")
|
|
def _(rid, params: dict) -> dict:
|
|
try:
|
|
from tools.process_registry import process_registry
|
|
|
|
procs = process_registry.list_sessions()
|
|
return _ok(
|
|
rid,
|
|
{
|
|
"processes": [
|
|
{
|
|
"session_id": p["session_id"],
|
|
"command": p["command"][:80],
|
|
"status": p["status"],
|
|
"uptime": p["uptime_seconds"],
|
|
}
|
|
for p in procs
|
|
]
|
|
},
|
|
)
|
|
except Exception as e:
|
|
return _err(rid, 5033, str(e))
|
|
|
|
|
|
@method("cron.manage")
|
|
def _(rid, params: dict) -> dict:
|
|
action, jid = params.get("action", "list"), params.get("name", "")
|
|
try:
|
|
from tools.cronjob_tools import cronjob
|
|
|
|
if action == "list":
|
|
return _ok(rid, json.loads(cronjob(action="list")))
|
|
if action == "add":
|
|
return _ok(
|
|
rid,
|
|
json.loads(
|
|
cronjob(
|
|
action="create",
|
|
name=jid,
|
|
schedule=params.get("schedule", ""),
|
|
prompt=params.get("prompt", ""),
|
|
)
|
|
),
|
|
)
|
|
if action in ("remove", "pause", "resume"):
|
|
return _ok(rid, json.loads(cronjob(action=action, job_id=jid)))
|
|
return _err(rid, 4016, f"unknown cron action: {action}")
|
|
except Exception as e:
|
|
return _err(rid, 5023, str(e))
|
|
|
|
|
|
@method("skills.manage")
|
|
def _(rid, params: dict) -> dict:
|
|
action, query = params.get("action", "list"), params.get("query", "")
|
|
try:
|
|
if action == "list":
|
|
from hermes_cli.banner import get_available_skills
|
|
|
|
return _ok(rid, {"skills": get_available_skills()})
|
|
if action == "search":
|
|
from tools.skills_hub import (
|
|
GitHubAuth,
|
|
create_source_router,
|
|
unified_search,
|
|
)
|
|
|
|
raw = (
|
|
unified_search(
|
|
query,
|
|
create_source_router(GitHubAuth()),
|
|
source_filter="all",
|
|
limit=20,
|
|
)
|
|
or []
|
|
)
|
|
return _ok(
|
|
rid,
|
|
{
|
|
"results": [
|
|
{"name": r.name, "description": r.description} for r in raw
|
|
]
|
|
},
|
|
)
|
|
if action == "install":
|
|
from hermes_cli.skills_hub import do_install
|
|
|
|
class _Q:
|
|
def print(self, *a, **k):
|
|
pass
|
|
|
|
do_install(query, skip_confirm=True, console=_Q())
|
|
return _ok(rid, {"installed": True, "name": query})
|
|
if action == "browse":
|
|
from hermes_cli.skills_hub import browse_skills
|
|
|
|
pg = int(params.get("page", 0) or 0) or (
|
|
int(query) if query.isdigit() else 1
|
|
)
|
|
return _ok(
|
|
rid, browse_skills(page=pg, page_size=int(params.get("page_size", 20)))
|
|
)
|
|
if action == "inspect":
|
|
from hermes_cli.skills_hub import inspect_skill
|
|
|
|
return _ok(rid, {"info": inspect_skill(query) or {}})
|
|
return _err(rid, 4017, f"unknown skills action: {action}")
|
|
except Exception as e:
|
|
return _err(rid, 5024, str(e))
|
|
|
|
|
|
# ── Methods: shell ───────────────────────────────────────────────────
|
|
|
|
|
|
@method("shell.exec")
|
|
def _(rid, params: dict) -> dict:
|
|
cmd = params.get("command", "")
|
|
if not cmd:
|
|
return _err(rid, 4004, "empty command")
|
|
try:
|
|
from tools.approval import detect_dangerous_command
|
|
|
|
is_dangerous, _, desc = detect_dangerous_command(cmd)
|
|
if is_dangerous:
|
|
return _err(
|
|
rid, 4005, f"blocked: {desc}. Use the agent for dangerous commands."
|
|
)
|
|
except ImportError:
|
|
pass
|
|
try:
|
|
r = subprocess.run(
|
|
cmd, shell=True, capture_output=True, text=True, timeout=30, cwd=os.getcwd()
|
|
)
|
|
return _ok(
|
|
rid,
|
|
{
|
|
"stdout": r.stdout[-4000:],
|
|
"stderr": r.stderr[-2000:],
|
|
"code": r.returncode,
|
|
},
|
|
)
|
|
except subprocess.TimeoutExpired:
|
|
return _err(rid, 5002, "command timed out (30s)")
|
|
except Exception as e:
|
|
return _err(rid, 5003, str(e))
|