mirror of
https://github.com/PaddlePaddle/FastDeploy.git
synced 2025-10-05 16:48:03 +08:00
fix glm all_reduce tp group (#4187)
This commit is contained in:
@@ -157,7 +157,7 @@ class Glm4Moe(nn.Layer):
|
|||||||
out = out + shared_experts_out
|
out = out + shared_experts_out
|
||||||
# We do to TP all reduce after the sum of experts.
|
# We do to TP all reduce after the sum of experts.
|
||||||
if self.tensor_parallel_size > 1:
|
if self.tensor_parallel_size > 1:
|
||||||
tensor_model_parallel_all_reduce(out)
|
tensor_model_parallel_all_reduce(out, self.tp_group)
|
||||||
return out
|
return out
|
||||||
|
|
||||||
|
|
||||||
|
Reference in New Issue
Block a user