Hi, the blockwise_fp8_training has been there for a while. Is there any reason we dont merge it into float8 folder?
And current moe training only supports FP8_ROWWISE and MXFP8, will FP8_BlockWise be considered to be added into torchao in the near future? (mainly for h100 users)
thanks!