fix: increase max_tokens for GLM 5.1 reasoning headroom

This commit is contained in:
iborazzi
2026-04-14 13:50:18 +03:00
committed by Teknium
parent c78a188ddd
commit f41031af3a

View File

@@ -38,7 +38,7 @@ def generate_title(user_message: str, assistant_response: str, timeout: float =
response = call_llm(
task="title_generation",
messages=messages,
max_tokens=30,
max_tokens=500,
temperature=0.3,
timeout=timeout,
)