delete_moe_phase_in_parallel_config (#4264)

This commit is contained in:
Yuanle Liu
2025-09-25 17:14:37 +08:00
committed by GitHub
parent 87179cb744
commit 0c6f1932c5

View File

@@ -519,7 +519,6 @@ class ParallelConfig:
):
self.sequence_parallel = False # Whether to enable sequence parallelism.
self.use_ep = False # Whether to enable Expert Parallelism
self.moe_phase = MoEPhase("prefill") # Generation phase
self.msg_queue_id = 1 # message queue id
self.tensor_parallel_rank = 0 # TP rank ID