From 87f270df23233d28699440b4806f1cc2fe48ed09 Mon Sep 17 00:00:00 2001 From: Peter Steinberger Date: Mon, 12 Jan 2026 04:04:04 +0000 Subject: [PATCH] test: respect openai max tokens floor --- src/agents/pi-embedded-runner-extraparams.live.test.ts | 7 ++++--- 1 file changed, 4 insertions(+), 3 deletions(-) diff --git a/src/agents/pi-embedded-runner-extraparams.live.test.ts b/src/agents/pi-embedded-runner-extraparams.live.test.ts index ce1184921..c92be3448 100644 --- a/src/agents/pi-embedded-runner-extraparams.live.test.ts +++ b/src/agents/pi-embedded-runner-extraparams.live.test.ts @@ -18,8 +18,9 @@ describeLive("pi embedded extra params (live)", () => { defaults: { models: { "openai/gpt-5.2": { + // OpenAI Responses enforces a minimum max_output_tokens of 16. params: { - maxTokens: 8, + maxTokens: 16, }, }, }, @@ -57,7 +58,7 @@ describeLive("pi embedded extra params (live)", () => { expect(stopReason).toBeDefined(); expect(outputTokens).toBeDefined(); - // Should respect maxTokens from config (8) — allow a small buffer for provider rounding. - expect(outputTokens ?? 0).toBeLessThanOrEqual(12); + // Should respect maxTokens from config (16) — allow a small buffer for provider rounding. + expect(outputTokens ?? 0).toBeLessThanOrEqual(20); }, 30_000); });