fix(cli): hint about /v1 suffix when configuring local model endpoints

When a user enters a local model server URL (Ollama, vLLM, llama.cpp)
without a /v1 suffix during 'hermes model' custom endpoint setup,
prompt them to add it. Most OpenAI-compatible local servers require
/v1 in the base URL for chat completions to work.
This commit is contained in:
Teknium
2026-04-15 22:36:35 -07:00
committed by Teknium
parent 8798b069d3
commit 5c397876b9
2 changed files with 22 additions and 1 deletions

View File

@@ -984,7 +984,7 @@ class TestBuildApiKwargs:
messages = [{"role": "system", "content": "sys"}, {"role": "user", "content": "hi"}]
kwargs = agent._build_api_kwargs(messages)
assert kwargs["max_tokens"] == 65536
=======
def test_ollama_think_false_on_effort_none(self, agent):
"""Custom (Ollama) provider with effort=none should inject think=false."""
agent.provider = "custom"