Skip to content

Revert "feat(vlm): add streaming response handling for OpenAI VLM"#745

Merged
qin-ctx merged 1 commit intovolcengine:mainfrom
KorenKrita:revert-740-feat/openai-vlm-streaming-handler
Mar 18, 2026
Merged

Revert "feat(vlm): add streaming response handling for OpenAI VLM"#745
qin-ctx merged 1 commit intovolcengine:mainfrom
KorenKrita:revert-740-feat/openai-vlm-streaming-handler

Conversation

@KorenKrita
Copy link
Copy Markdown
Contributor

Reverts #740

The issue with the previous approach: The SSE auto-detection logic added in the previous commit is fundamentally ineffective. When a server forces
SSE responses but stream=False is passed, the OpenAI SDK fails at the HTTP parsing layer (trying to JSON-decode the raw SSE stream) before our code
can even inspect the response. The runtime type checking (_is_streaming_response, etc.) never gets a chance to execute.

The correct fix: Simply revert the streaming detection logic and expose stream as a user-configurable option. When stream=True, the SDK switches to
SSE parsing mode; when stream=False, it expects standard JSON. No runtime detection needed—let the SDK handle it based on explicit user intent.


Tomorrow's PR will:

  1. Add stream: bool = False config option in VLMConfig and VLMBase
  2. Pass stream=self.stream to all client.chat.completions.create() calls
  3. Handle response differently based on self.stream (iterate chunks vs direct access)

@qin-ctx qin-ctx merged commit 2294355 into volcengine:main Mar 18, 2026
6 checks passed
@github-project-automation github-project-automation bot moved this from Backlog to Done in OpenViking project Mar 18, 2026
chethanuk added a commit to chethanuk/OpenViking that referenced this pull request Mar 19, 2026
- Add .pr_agent.toml with 15 repo-specific review rules derived from real
  bug history (PRs volcengine#505, volcengine#728, volcengine#749, volcengine#740/volcengine#745, volcengine#754, volcengine#735, volcengine#767)
- Rules structured as WHEN/THEN/BECAUSE for deterministic enforcement
- Add 8 custom labels (memory-pipeline, async-change, api-breaking, etc.)
- Add ignore patterns for lock files, third_party, build artifacts
- Enable score review, TODO scan, split-PR detection, security audit
- Configure improve tool with quality threshold and extended mode
- Configure describe tool with PR diagrams and semantic file types
- Update workflow: ark-code-latest model, checkout step for .pr_agent.toml,
  move all config from inline YAML to .pr_agent.toml (single source of truth)
qin-ctx pushed a commit that referenced this pull request Mar 19, 2026
…#780)

- Add .pr_agent.toml with 15 repo-specific review rules derived from real
  bug history (PRs #505, #728, #749, #740/#745, #754, #735, #767)
- Rules structured as WHEN/THEN/BECAUSE for deterministic enforcement
- Add 8 custom labels (memory-pipeline, async-change, api-breaking, etc.)
- Add ignore patterns for lock files, third_party, build artifacts
- Enable score review, TODO scan, split-PR detection, security audit
- Configure improve tool with quality threshold and extended mode
- Configure describe tool with PR diagrams and semantic file types
- Update workflow: ark-code-latest model, checkout step for .pr_agent.toml,
  move all config from inline YAML to .pr_agent.toml (single source of truth)
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

Status: Done

Development

Successfully merging this pull request may close these issues.

2 participants