feat: add requestoptions.stream in model config #10026
Open
+13
−1
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Description
Add support for
requestOptions.streamin models section in config.yamlThis add supports for using reasoning models of openai using this flag.
example:
closes CON-5285
AI Code Review
@continue-reviewChecklist
Screen recording or screenshot
[ When applicable, please include a short screen recording or screenshot - this makes it much easier for us as contributors to review and understand your changes. See this PR as a good example. ]
Tests
[ What tests were added or updated to ensure the changes work as expected? ]
Continue Tasks:▶️ 1 queued · ✅ 1 no changes — View all
Summary by cubic
Adds requestOptions.stream to model config and enforces it in the LLM layer. This lets you turn off streaming for OpenAI reasoning (o-series) models and addresses CON-5285.
Written for commit ef4cb51. Summary will update on new commits.