Skip to content

feat(inference): forward extra_body to remote VLM (DashScope enable_thinking)#65

Merged
Liuhaai merged 1 commit into
mainfrom
feat/vlm-extra-body
May 8, 2026
Merged

feat(inference): forward extra_body to remote VLM (DashScope enable_thinking)#65
Liuhaai merged 1 commit into
mainfrom
feat/vlm-extra-body

Conversation

@Liuhaai
Copy link
Copy Markdown
Collaborator

@Liuhaai Liuhaai commented May 8, 2026

Summary

  • Plumb a new extra_body kwarg from DescribeRequest through engine.analyze_frame and BaseBackend.generate to RemoteHTTPBackend, passed verbatim to the OpenAI SDK.
  • Lets cortex toggle DashScope-specific knobs (e.g. enable_thinking=False) per-request without adding model-specific fields to the public schema.
  • Local backends accept and ignore the kwarg.

Test plan

  • Send a /describe request with extra_body={"enable_thinking": false} against a DashScope-backed remote VLM and confirm the field reaches the upstream payload.
  • Confirm local backends (mlx/transformers/compressed/tome) accept and ignore the kwarg.

🤖 Generated with Claude Code

…hinking)

Plumb a new extra_body kwarg from DescribeRequest through engine.analyze_frame
and BaseBackend.generate to RemoteHTTPBackend, where it is passed verbatim to
the OpenAI SDK (which threads it into the JSON payload). This lets cortex
toggle DashScope-specific knobs like enable_thinking=False per-request without
adding model-specific fields to the public schema. Local backends accept and
ignore the kwarg.

Co-Authored-By: Claude Opus 4.7 <noreply@anthropic.com>
@Liuhaai Liuhaai merged commit d0603ed into main May 8, 2026
7 checks passed
@Liuhaai Liuhaai deleted the feat/vlm-extra-body branch May 8, 2026 04:17
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant