The module 'HunyuanVideoTransformer3DModel' has been loaded in bitsandbytes
8bit and moving it to cpu via .to()
is not supported. Module is still on cuda:0
#10653
Labels
bug
Something isn't working
Describe the bug
enable_model_cpu_offload works with 4bit but not with 8bit. Is this the expected behavior or an issue?
The module 'HunyuanVideoTransformer3DModel' has been loaded in bitsandbytes 8bit and moving it to cpu via .to() is not supported. Module is still on cuda:0
device_map="balanced", also doesn't work with int8 (after commenting enable_model_cpu_offload ) [not tested for int4]
RuntimeError: Expected all tensors to be on the same device, but found at least two devices, cpu and cuda:0! (when checking argument for argument index in method wrapper_CUDA__index_select)
Reproduction
Logs
System Info
Who can help?
No response
The text was updated successfully, but these errors were encountered: