mirror of
https://github.com/NousResearch/hermes-agent.git
synced 2026-04-28 06:51:16 +08:00
fix(cli): hint about /v1 suffix when configuring local model endpoints
When a user enters a local model server URL (Ollama, vLLM, llama.cpp) without a /v1 suffix during 'hermes model' custom endpoint setup, prompt them to add it. Most OpenAI-compatible local servers require /v1 in the base URL for chat completions to work.
This commit is contained in:
@@ -984,7 +984,7 @@ class TestBuildApiKwargs:
|
||||
messages = [{"role": "system", "content": "sys"}, {"role": "user", "content": "hi"}]
|
||||
kwargs = agent._build_api_kwargs(messages)
|
||||
assert kwargs["max_tokens"] == 65536
|
||||
=======
|
||||
|
||||
def test_ollama_think_false_on_effort_none(self, agent):
|
||||
"""Custom (Ollama) provider with effort=none should inject think=false."""
|
||||
agent.provider = "custom"
|
||||
|
||||
Reference in New Issue
Block a user