[BugFix] fix TPDP mix parallel infer (#4583)

Co-authored-by: Yuanle Liu <yuanlehome@163.com>
This commit is contained in:
lizhenyun01
2025-10-28 16:58:20 +08:00
committed by GitHub
parent c63361fd1d
commit 4d2f478d53
3 changed files with 13 additions and 13 deletions

View File

@@ -1588,10 +1588,6 @@ class FDConfig:
f"be less than or equal to max_num_partial_prefills: {self.max_num_partial_prefills}"
)
assert self.scheduler_config.splitwise_role in ["mixed", "prefill", "decode"]
# TODO(@wufeisheng): TP and EP need to be supported simultaneously.
assert (self.parallel_config.tensor_parallel_size == 1 and self.parallel_config.expert_parallel_size >= 1) or (
self.parallel_config.tensor_parallel_size >= 1 and self.parallel_config.expert_parallel_size == 1
), "TP and EP cannot be enabled at the same time"
if not self.cache_config.enable_chunked_prefill:
if not envs.ENABLE_V1_KVCACHE_SCHEDULER: