From be216eacad49e3fc7c0fe8c9b0f855e6fab72bfb Mon Sep 17 00:00:00 2001 From: empty Date: Wed, 7 Jan 2026 03:37:55 +0800 Subject: [PATCH] fix: Increase VLM max_tokens to 2000 to avoid response truncation --- pixelle_video/services/quality/character_analyzer.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/pixelle_video/services/quality/character_analyzer.py b/pixelle_video/services/quality/character_analyzer.py index db906fa..3a38435 100644 --- a/pixelle_video/services/quality/character_analyzer.py +++ b/pixelle_video/services/quality/character_analyzer.py @@ -167,7 +167,7 @@ Output ONLY the JSON object, no additional text.""" model=vision_model, messages=messages, temperature=0.3, - max_tokens=800 + max_tokens=2000 # Increased to avoid truncation ) vlm_response = response.choices[0].message.content if response.choices else None