Skip to content

Conversation

@wangshangsam
Copy link
Contributor

@wangshangsam wangshangsam commented Oct 28, 2025

This is the first PR towards the VLM reference implementation for the v6.0 round.
This PR currenlty supports the Offline scenario + performance-only mode. Server scenario and accuracy mode will be introduced through subsequent PRs.
The issue_query implemenation adopted the purely asyncio-based design from the DSR1 reference implementation, but the code here is simpler mostly because we only access the inference endpoint through OpenAI APIs.

@github-actions
Copy link
Contributor

github-actions bot commented Oct 28, 2025

MLCommons CLA bot All contributors have signed the MLCommons CLA ✍️ ✅

@wangshangsam wangshangsam marked this pull request as ready for review November 4, 2025 08:31
@wangshangsam wangshangsam requested a review from a team as a code owner November 4, 2025 08:31
@wangshangsam wangshangsam changed the title VLM reference implementation [VLM] Offline scenario, performance-only mode for the reference implementation Nov 4, 2025
@wangshangsam wangshangsam changed the title [VLM] Offline scenario, performance-only mode for the reference implementation [VLM] Offline scenario, performance-only mode of the reference implementation Nov 4, 2025
…the client, event loop and the event loop thread
Copy link
Contributor

@hanyunfan hanyunfan left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM

@hanyunfan hanyunfan merged commit 808e2d7 into mlcommons:master Nov 11, 2025
29 checks passed
@github-actions github-actions bot locked and limited conversation to collaborators Nov 11, 2025
@wangshangsam wangshangsam deleted the wangshangsam/vlm-sut-prototype branch November 12, 2025 00:39
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants