File tree 1 file changed +3
-2
lines changed
1 file changed +3
-2
lines changed Original file line number Diff line number Diff line change @@ -567,7 +567,7 @@ class TrainingArguments:
567
567
used when the xla flag is set to true, and an auto wrapping policy is specified through
568
568
fsdp_min_num_params or fsdp_transformer_layer_cls_to_wrap.
569
569
tp_size (`int`, *optional*):
570
- Use tp_size to enable pytorch 2.0 tensor parallelism. Set a value greater than 1 to activate TP. The same is
570
+ Use tp_size to enable PyTorch tensor parallelism. Set a value greater than 1 to activate TP. The same is
571
571
used to prepare device mesh internally.
572
572
deepspeed (`str` or `dict`, *optional*):
573
573
Use [Deepspeed](https://github.com/microsoft/deepspeed). This is an experimental feature and its API may
@@ -1246,7 +1246,7 @@ class TrainingArguments:
1246
1246
default = 0 ,
1247
1247
metadata = {
1248
1248
"help" : (
1249
- "Use tp_size to enable pytorch 2.0 tensor parallelism."
1249
+ "Use tp_size to enable pytorch tensor parallelism."
1250
1250
"Set a value greater than 1 to activate TP."
1251
1251
"The same is used to prepare device mesh internally."
1252
1252
)
@@ -1971,6 +1971,7 @@ def __post_init__(self):
1971
1971
1972
1972
if self .tp_size > 1 :
1973
1973
os .environ ["ACCELERATE_USE_TP" ] = "true"
1974
+ os .environ ["TP_SIZE" ] = str (self .tp_size )
1974
1975
# accelerate integration for FSDP
1975
1976
if len (self .fsdp ) > 0 and not self .fsdp_config ["xla" ]:
1976
1977
os .environ ["ACCELERATE_USE_FSDP" ] = "true"
You can’t perform that action at this time.
0 commit comments