Skip to content

Conversation

@MaoZiming
Copy link
Member

Description

Please include a summary of the changes and the related issue.

Fixes # (issue)

Type of Change

  • Bug fix
  • New feature
  • Documentation update

How Has This Been Tested?

Include any tests here.

  • Unit tests
  • Integration tests
  • Manual testing

Checklist

  • My code follows the style guidelines, e.g. format.sh.
  • I have run build_and_install.sh to verify compilation.
  • I have removed redundant variables and comments.
  • I have updated the documentation.
  • I have added tests.

@MaoZiming
Copy link
Member Author

bash launch_vllm_head.sh 10.4.147.22 13345 deepseek-ai/DeepSeek-V3-0324 deepep_low_latency 2 1 8 1
bash launch_vllm_worker.sh 10.4.147.22 13345 deepseek-ai/DeepSeek-V3-0324 deepep_low_latency 2 1 8 1

This works.

@MaoZiming MaoZiming changed the title [EP] Debug vLLM; high-throughput works but low-latency stuck at initialization [EP] Support vLLM; high-throughput and low-latency Jan 6, 2026
@MaoZiming MaoZiming marked this pull request as ready for review January 6, 2026 07:36
@MaoZiming
Copy link
Member Author

Using this commit b1029ad right now, there seems to be some changes from merging master. I need to look more.

@MaoZiming
Copy link
Member Author

MaoZiming commented Jan 6, 2026

#620
This seems to break the vLLM high-throughput mode integration. Marking it here to double check later.
I got

[wait_until_cmd_consumed nvl:5 cmd:4 label:3] waiting slot=565

@MaoZiming MaoZiming merged commit e0233e9 into vllm-launch Jan 7, 2026
@MaoZiming MaoZiming deleted the vllm-launch-zm-debug branch January 7, 2026 21:41
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants