diff --git a/fastdeploy/model_executor/models/glm4_moe.py b/fastdeploy/model_executor/models/glm4_moe.py index fdbf277af..1e2897034 100644 --- a/fastdeploy/model_executor/models/glm4_moe.py +++ b/fastdeploy/model_executor/models/glm4_moe.py @@ -157,7 +157,7 @@ class Glm4Moe(nn.Layer): out = out + shared_experts_out # We do to TP all reduce after the sum of experts. if self.tensor_parallel_size > 1: - tensor_model_parallel_all_reduce(out) + tensor_model_parallel_all_reduce(out, self.tp_group) return out