mirror of
https://github.com/NousResearch/hermes-agent.git
synced 2026-04-29 23:41:35 +08:00
feat: add streaming LLM response support across all platforms
Cherry-picked from PR #828, resolved conflicts with main.
This commit is contained in:
@@ -219,6 +219,22 @@ compression:
|
||||
# Options: "auto", "openrouter", "nous", "main"
|
||||
# summary_provider: "auto"
|
||||
|
||||
# =============================================================================
|
||||
# Streaming (live token-by-token response display)
|
||||
# =============================================================================
|
||||
# When enabled, LLM responses stream token-by-token instead of appearing
|
||||
# all at once. Supported on Telegram, Discord, Slack (via message editing)
|
||||
# and the API server (via SSE). Disabled by default.
|
||||
#
|
||||
# streaming:
|
||||
# enabled: false # Master switch (default: off)
|
||||
# # Per-platform overrides:
|
||||
# # telegram: true
|
||||
# # discord: true
|
||||
# # api_server: true
|
||||
# # edit_interval: 1.5 # Seconds between message edits (default: 1.5)
|
||||
# # min_tokens: 20 # Tokens before first display (default: 20)
|
||||
|
||||
# =============================================================================
|
||||
# Auxiliary Models (Advanced — Experimental)
|
||||
# =============================================================================
|
||||
|
||||
Reference in New Issue
Block a user