@@ -20,11 +20,7 @@ import {
2020 generateToolUseId ,
2121 getBedrockInferenceProfileId ,
2222} from '@/providers/bedrock/utils'
23- import {
24- getMaxOutputTokensForModel ,
25- getProviderDefaultModel ,
26- getProviderModels ,
27- } from '@/providers/models'
23+ import { getProviderDefaultModel , getProviderModels } from '@/providers/models'
2824import type {
2925 ProviderConfig ,
3026 ProviderRequest ,
@@ -261,18 +257,11 @@ export const bedrockProvider: ProviderConfig = {
261257
262258 const systemPromptWithSchema = systemContent
263259
264- const inferenceConfig = {
260+ const inferenceConfig : { temperature : number ; maxTokens ?: number } = {
265261 temperature : Number . parseFloat ( String ( request . temperature ?? 0.7 ) ) ,
266- maxTokens :
267- Number . parseInt ( String ( request . maxTokens ) ) ||
268- getMaxOutputTokensForModel ( request . model , request . stream ?? false ) ,
269262 }
270-
271- const toolLoopInferenceConfig = {
272- ...inferenceConfig ,
273- maxTokens :
274- Number . parseInt ( String ( request . maxTokens ) ) ||
275- getMaxOutputTokensForModel ( request . model , false ) ,
263+ if ( request . maxTokens != null ) {
264+ inferenceConfig . maxTokens = Number . parseInt ( String ( request . maxTokens ) )
276265 }
277266
278267 const shouldStreamToolCalls = request . streamToolCalls ?? false
@@ -381,7 +370,7 @@ export const bedrockProvider: ProviderConfig = {
381370 modelId : bedrockModelId ,
382371 messages,
383372 system : systemPromptWithSchema . length > 0 ? systemPromptWithSchema : undefined ,
384- inferenceConfig : toolLoopInferenceConfig ,
373+ inferenceConfig,
385374 toolConfig,
386375 } )
387376
@@ -627,7 +616,7 @@ export const bedrockProvider: ProviderConfig = {
627616 modelId : bedrockModelId ,
628617 messages : currentMessages ,
629618 system : systemPromptWithSchema . length > 0 ? systemPromptWithSchema : undefined ,
630- inferenceConfig : toolLoopInferenceConfig ,
619+ inferenceConfig : inferenceConfig ,
631620 toolConfig : bedrockTools ?. length
632621 ? { tools : bedrockTools , toolChoice : nextToolChoice }
633622 : undefined ,
@@ -694,7 +683,7 @@ export const bedrockProvider: ProviderConfig = {
694683 modelId : bedrockModelId ,
695684 messages : currentMessages ,
696685 system : systemPromptWithSchema . length > 0 ? systemPromptWithSchema : undefined ,
697- inferenceConfig : toolLoopInferenceConfig ,
686+ inferenceConfig : inferenceConfig ,
698687 toolConfig : {
699688 tools : [ structuredOutputTool ] ,
700689 toolChoice : { tool : { name : structuredOutputToolName } } ,
0 commit comments