Skip to content

Conversation

@ansh-info
Copy link

• Summary

  • Default vLLM/Unsloth tensor_parallel_size to the number of visible GPUs (respects CUDA_VISIBLE_DEVICES) so multi-GPU setups don’t start with an unset/invalid TP world size.

  • Still allows explicit overrides via _internal_config["engine_args"].

    Motivation

  • On multi-GPU H100 machines, vLLM was crashing during init when TP wasn’t configured; setting a sensible default prevents a TP=0/None startup.

    Details

  • In get_model_config, set engine_args["tensor_parallel_size"] = torch.cuda.device_count() (or 1 if CUDA unavailable) before merging user overrides.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant