## Rollout ### FP8 - Doc: https://github.com/verl-project/verl/blob/main/docs/advance/fp8.md - Recipe: https://github.com/verl-project/verl-recipe/tree/main/low_precision - Supported backend: vllm/sglang/trtllm - Supported scenario: colocated&disaggregated; blockwise fp8 recipe; only quantize fp8 linear layer; - Hardware requirement: supported on Hopper and Blackwell - Unsupported scenario: hidden/intermediate size must be an integer multiple of block_size(128) ([related code](https://github.com/vllm-project/vllm/blob/main/vllm/model_executor/layers/quantization/utils/fp8_utils.py#L910)) - Related PR: - [x] vllm: https://github.com/verl-project/verl/pull/3519 - [x] sglang: https://github.com/verl-project/verl/pull/4415 - [x] trtllm: https://github.com/verl-project/verl/pull/5374 ### MXFP8 - [ ] MXFP8 rollout ## QAT <img width="778" height="262" alt="Image" src="https://github.com/user-attachments/assets/0fac2c52-dfaf-46a5-a909-5e081b430185" /> ### NVFP4 W4A16 QAT - Doc: https://github.com/verl-project/verl/blob/main/docs/advance/nvfp4_qat.md - Recipe: https://github.com/verl-project/verl-recipe/tree/main/qat - Supported backend: vllm; - Hardware requirement: only supported on Blackwell - Related PR: - [x] vllm: - [x] FSDP https://github.com/verl-project/verl/pull/5190 - [x] Megatron: https://github.com/verl-project/verl-recipe/pull/55 ## Train ### FP8 - Doc: https://github.com/verl-project/verl/blob/main/docs/advance/fp8.md - Recipe: FP8 end-to-end RL: https://github.com/verl-project/verl-recipe/blob/main/low_precision/run_dapo_qwen3_moe_30b_megatron_fp8e2e.sh - Supported backend: Megatron - Supported scenario: quantize linear layer, all GEMMs (FPROP, DGRAD and WGRAD) for QKV, Proj, FC1 and FC2 <img width="358" height="97" alt="Image" src="https://github.com/user-attachments/assets/75ae1bdb-5f3a-40cf-ae1e-ee1e57c89fee" /> - Hardware requirement: only supported on Hopper and Blackwell ### MXFP8 - [ ] MXFP8 training in Megatron - Hardware requirement: Blackwell ### NVFP4 - [ ] NVFP4 training in Megatron - Hardware requirement: Blackwell
Rollout
FP8
MXFP8
QAT
NVFP4 W4A16 QAT
Train
FP8
MXFP8
NVFP4