Revert "feat(vlm): add streaming response handling for OpenAI VLM"#745
Merged
qin-ctx merged 1 commit intovolcengine:mainfrom Mar 18, 2026
Merged
Conversation
…lcengine#740)" This reverts commit 247293b.
qin-ctx
approved these changes
Mar 18, 2026
This was referenced Mar 18, 2026
chethanuk
added a commit
to chethanuk/OpenViking
that referenced
this pull request
Mar 19, 2026
- Add .pr_agent.toml with 15 repo-specific review rules derived from real bug history (PRs volcengine#505, volcengine#728, volcengine#749, volcengine#740/volcengine#745, volcengine#754, volcengine#735, volcengine#767) - Rules structured as WHEN/THEN/BECAUSE for deterministic enforcement - Add 8 custom labels (memory-pipeline, async-change, api-breaking, etc.) - Add ignore patterns for lock files, third_party, build artifacts - Enable score review, TODO scan, split-PR detection, security audit - Configure improve tool with quality threshold and extended mode - Configure describe tool with PR diagrams and semantic file types - Update workflow: ark-code-latest model, checkout step for .pr_agent.toml, move all config from inline YAML to .pr_agent.toml (single source of truth)
4 tasks
qin-ctx
pushed a commit
that referenced
this pull request
Mar 19, 2026
…#780) - Add .pr_agent.toml with 15 repo-specific review rules derived from real bug history (PRs #505, #728, #749, #740/#745, #754, #735, #767) - Rules structured as WHEN/THEN/BECAUSE for deterministic enforcement - Add 8 custom labels (memory-pipeline, async-change, api-breaking, etc.) - Add ignore patterns for lock files, third_party, build artifacts - Enable score review, TODO scan, split-PR detection, security audit - Configure improve tool with quality threshold and extended mode - Configure describe tool with PR diagrams and semantic file types - Update workflow: ark-code-latest model, checkout step for .pr_agent.toml, move all config from inline YAML to .pr_agent.toml (single source of truth)
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
Reverts #740
The issue with the previous approach: The SSE auto-detection logic added in the previous commit is fundamentally ineffective. When a server forces
SSE responses but stream=False is passed, the OpenAI SDK fails at the HTTP parsing layer (trying to JSON-decode the raw SSE stream) before our code
can even inspect the response. The runtime type checking (_is_streaming_response, etc.) never gets a chance to execute.
The correct fix: Simply revert the streaming detection logic and expose stream as a user-configurable option. When stream=True, the SDK switches to
SSE parsing mode; when stream=False, it expects standard JSON. No runtime detection needed—let the SDK handle it based on explicit user intent.
Tomorrow's PR will: