@@ -75,33 +75,6 @@ export const openaiProvider: ProviderConfig = {
7575 } ) )
7676 : undefined
7777
78- // === DEBUG: Log full request details ===
79- logger . info ( '[OpenAIProvider] === FULL REQUEST DEBUG ===' )
80- logger . info ( `[OpenAIProvider] Messages: ${ allMessages . length } total` )
81- for ( const [ i , m ] of allMessages . entries ( ) ) {
82- const content = typeof m . content === 'string' ? m . content : JSON . stringify ( m . content )
83- logger . info ( `[OpenAIProvider] [${ i } ] ${ m . role } : ${ content ?. substring ( 0 , 150 ) } ...` )
84- }
85-
86- // Log tool definitions with formatted JSON
87- if ( tools ?. length ) {
88- for ( const tool of tools ) {
89- logger . info ( `[OpenAIProvider] Tool: ${ tool . function . name } ` )
90- logger . info (
91- `[OpenAIProvider] Description: ${ tool . function . description ?. substring ( 0 , 200 ) } ...`
92- )
93- logger . info ( `[OpenAIProvider] Parameters:` )
94- const params = tool . function . parameters as any
95- if ( params ?. properties ) {
96- for ( const [ key , val ] of Object . entries ( params . properties ) ) {
97- const desc = ( val as any ) . description || ''
98- logger . info ( `[OpenAIProvider] - ${ key } : ${ desc . substring ( 0 , 100 ) } ` )
99- }
100- }
101- }
102- }
103- logger . info ( '[OpenAIProvider] === END REQUEST DEBUG ===' )
104-
10578 const payload : any = {
10679 model : request . model ,
10780 messages : allMessages ,
@@ -320,40 +293,16 @@ export const openaiProvider: ProviderConfig = {
320293
321294 try {
322295 const toolArgs = JSON . parse ( toolCall . function . arguments )
323-
324- // === DEBUG: Log LLM tool call details ===
325- logger . info ( '[OpenAIProvider] === LLM TOOL CALL ===' )
326- logger . info ( `[OpenAIProvider] Tool: ${ toolName } ` )
327- logger . info ( `[OpenAIProvider] Arguments: ${ JSON . stringify ( toolArgs , null , 2 ) } ` )
328- if ( toolName . startsWith ( 'table_' ) ) {
329- const filterStr = toolArgs . filter ? JSON . stringify ( toolArgs . filter , null , 2 ) : 'NONE'
330- logger . info ( `[OpenAIProvider] Filter: ${ filterStr } ` )
331- logger . info ( `[OpenAIProvider] Limit: ${ toolArgs . limit || 'default' } ` )
332- }
333-
334296 const tool = request . tools ?. find ( ( t ) => t . id === toolName )
335297
336298 if ( ! tool ) {
337299 return null
338300 }
339301
340302 const { toolParams, executionParams } = prepareToolExecution ( tool , toolArgs , request )
341-
342- logger . info ( `[OpenAIProvider] Executing ${ toolName } ...` )
343303 const result = await executeTool ( toolName , executionParams , true )
344304 const toolCallEndTime = Date . now ( )
345305
346- // === DEBUG: Log tool result ===
347- const resultString = JSON . stringify ( result )
348- const sizeKB = Math . round ( resultString . length / 1024 )
349- logger . info ( `[OpenAIProvider] === TOOL RESULT ===` )
350- logger . info ( `[OpenAIProvider] Success: ${ result . success } , Size: ${ sizeKB } KB` )
351- if ( result . output ?. rows ) {
352- logger . info (
353- `[OpenAIProvider] Rows returned: ${ result . output . rows . length } , Total matching: ${ result . output . totalCount } `
354- )
355- }
356-
357306 return {
358307 toolCall,
359308 toolName,
@@ -434,17 +383,10 @@ export const openaiProvider: ProviderConfig = {
434383 success : result . success ,
435384 } )
436385
437- const toolMessageContent = JSON . stringify ( resultContent )
438- const msgSizeKB = Math . round ( toolMessageContent . length / 1024 )
439- const estTokens = Math . round ( toolMessageContent . length / 4 )
440- logger . info (
441- `[OpenAIProvider] Adding ${ toolName } result to conversation: ${ msgSizeKB } KB (~${ estTokens } tokens)`
442- )
443-
444386 currentMessages . push ( {
445387 role : 'tool' ,
446388 tool_call_id : toolCall . id ,
447- content : toolMessageContent ,
389+ content : JSON . stringify ( resultContent ) ,
448390 } )
449391 }
450392
0 commit comments