@@ -174,7 +174,7 @@ Full docs can be found at github.com/aws/amazon-` + binName
174
174
cli .ByteQuantityMinMaxRangeFlags (gpuMemoryTotal , nil , nil , "Number of GPUs' total memory (Example: 4 GiB)" )
175
175
cli .StringFlag (gpuManufacturer , nil , nil , "GPU Manufacturer name (Example: NVIDIA)" , nil )
176
176
cli .StringFlag (gpuModel , nil , nil , "GPU Model name (Example: K520)" , nil )
177
- cli .Int32MinMaxRangeFlags (inferenceAccelerators , nil , nil , "Total Number of inference accelerators (Example: 4)" )
177
+ cli .IntMinMaxRangeFlags (inferenceAccelerators , nil , nil , "Total Number of inference accelerators (Example: 4)" )
178
178
cli .StringFlag (inferenceAcceleratorManufacturer , nil , nil , "Inference Accelerator Manufacturer name (Example: AWS)" , nil )
179
179
cli .StringFlag (inferenceAcceleratorModel , nil , nil , "Inference Accelerator Model name (Example: Inferentia)" , nil )
180
180
cli .StringOptionsFlag (placementGroupStrategy , nil , nil , "Placement group strategy: [cluster, partition, spread]" , []string {"cluster" , "partition" , "spread" })
@@ -387,7 +387,7 @@ Full docs can be found at github.com/aws/amazon-` + binName
387
387
GpuMemoryRange : cli .ByteQuantityRangeMe (flags [gpuMemoryTotal ]),
388
388
GPUManufacturer : cli .StringMe (flags [gpuManufacturer ]),
389
389
GPUModel : cli .StringMe (flags [gpuModel ]),
390
- InferenceAcceleratorsRange : cli .Int32RangeMe (flags [inferenceAccelerators ]),
390
+ InferenceAcceleratorsRange : cli .IntRangeMe (flags [inferenceAccelerators ]),
391
391
InferenceAcceleratorManufacturer : cli .StringMe (flags [inferenceAcceleratorManufacturer ]),
392
392
InferenceAcceleratorModel : cli .StringMe (flags [inferenceAcceleratorModel ]),
393
393
PlacementGroupStrategy : cli .StringMe (flags [placementGroupStrategy ]),
0 commit comments