fix scaled_gemm_f8_i4_f16_weight_quantize input (#3685)

This commit is contained in:
co63oc
2025-08-29 11:04:04 +08:00
committed by GitHub
parent 2fb2c0f46a
commit b6edd15d55

View File

@@ -217,7 +217,7 @@ std::vector<paddle::DataType> Fp8Int4WeightQuantizeInferDtype(
PD_BUILD_STATIC_OP(scaled_gemm_f8_i4_f16_weight_quantize)
.Inputs({"intput"})
.Inputs({"input"})
.Attrs({"groupsize: int",
"scale_dtype: std::string"})
.Outputs({"output", "scale"})