You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I have read the above rules and searched the existing issues.
System Info
[2025-01-14 10:19:36,904] [INFO] [real_accelerator.py:203:get_accelerator] Setting ds_accelerator to cuda (auto detect)
[WARNING] async_io requires the dev libaio .so object and headers but these were not found.
[WARNING] async_io: please install the libaio-devel package with yum
[WARNING] If libaio is already installed (perhaps from source), try setting the CFLAGS and LDFLAGS environment variables to where it can be found.
[WARNING] Please specify the CUTLASS repo directory as environment variable $CUTLASS_PATH
[WARNING] sparse_attn requires a torch version >= 1.5 and < 2.0 but detected 2.5
[WARNING] using untested triton version (3.1.0), only 1.0.0 is known to be compatible
/home/zengshuang.zs/anaconda3/envs/mllm/lib/python3.10/site-packages/deepspeed/runtime/zero/linear.py:49: FutureWarning: is deprecated. Please use instead.
def forward(ctx, input, weight, bias=None):
/home/zengshuang.zs/anaconda3/envs/mllm/lib/python3.10/site-packages/deepspeed/runtime/zero/linear.py:67: FutureWarning: is deprecated. Please use instead.
def backward(ctx, grad_output):
Could not find the bitsandbytes CUDA binary at PosixPath('/home/zengshuang.zs/anaconda3/envs/mllm/lib/python3.10/site-packages/bitsandbytes/libbitsandbytes_cuda124.so')
The installed version of bitsandbytes was compiled without GPU support. 8-bit optimizers, 8-bit multiplication, and GPU quantization are unavailable.torch.cuda.amp.custom_fwd(args...)torch.amp.custom_fwd(args..., device_type='cuda')torch.cuda.amp.custom_bwd(args...)torch.amp.custom_bwd(args..., device_type='cuda')
Reminder
System Info
[2025-01-14 10:19:36,904] [INFO] [real_accelerator.py:203:get_accelerator] Setting ds_accelerator to cuda (auto detect)
[WARNING] async_io requires the dev libaio .so object and headers but these were not found.
[WARNING] async_io: please install the libaio-devel package with yum
[WARNING] If libaio is already installed (perhaps from source), try setting the CFLAGS and LDFLAGS environment variables to where it can be found.
[WARNING] Please specify the CUTLASS repo directory as environment variable $CUTLASS_PATH
[WARNING] sparse_attn requires a torch version >= 1.5 and < 2.0 but detected 2.5
[WARNING] using untested triton version (3.1.0), only 1.0.0 is known to be compatible
/home/zengshuang.zs/anaconda3/envs/mllm/lib/python3.10/site-packages/deepspeed/runtime/zero/linear.py:49: FutureWarning: is deprecated. Please use instead.
def forward(ctx, input, weight, bias=None):
/home/zengshuang.zs/anaconda3/envs/mllm/lib/python3.10/site-packages/deepspeed/runtime/zero/linear.py:67: FutureWarning: is deprecated. Please use instead.
def backward(ctx, grad_output):
Could not find the bitsandbytes CUDA binary at PosixPath('/home/zengshuang.zs/anaconda3/envs/mllm/lib/python3.10/site-packages/bitsandbytes/libbitsandbytes_cuda124.so')
The installed version of bitsandbytes was compiled without GPU support. 8-bit optimizers, 8-bit multiplication, and GPU quantization are unavailable.torch.cuda.amp.custom_fwd(args...)torch.amp.custom_fwd(args..., device_type='cuda')torch.cuda.amp.custom_bwd(args...)torch.amp.custom_bwd(args..., device_type='cuda')
llamafactory version: 0.9.2.dev0
Platform: Linux-3.10.0-1160.119.1.el7.x86_64-x86_64-with-glibc2.17
Python version: 3.10.16
PyTorch version: 2.5.1+cu124 (GPU)
Transformers version: 4.46.1
Datasets version: 3.1.0
Accelerate version: 1.0.1
PEFT version: 0.12.0
TRL version: 0.9.6
GPU type: NVIDIA RTX A6000
DeepSpeed version: 0.14.4
Bitsandbytes version: 0.45.0
vLLM version: 0.6.4.post1
Reproduction
为了防止OOM,使用optim: paged_adamw_8bit会报错如下:AttributeError: /home/zengshuang.zs/anaconda3/envs/mllm/lib/python3.10/site-packges/bitsandbytes/libbitsandbytes_cpu.so: undefined symbol: cget_managed_ptr。
train_config:
Others
No response
The text was updated successfully, but these errors were encountered: