mirror of
https://github.com/PaddlePaddle/FastDeploy.git
synced 2025-10-05 00:33:03 +08:00
Supports DP+TP+EP hybrid parallel deployment strategy (#3489)
* Support DP+TP+EP hybrid parallel deployment strategy * Support DP+TP+EP hybrid parallel deployment strategy * fix conflict * add moe_tp_ep function split_allgather_out * del tp_group in moe_cutlass_backend * for ci * fix parallel_config for ci * del log
This commit is contained in:
@@ -105,7 +105,8 @@ void SaveOutMmsg(const paddle::Tensor& x,
|
||||
int64_t rank_id,
|
||||
int msg_queue_id,
|
||||
bool save_each_rank) {
|
||||
if (!save_each_rank && rank_id > 0) {
|
||||
// don't use save_each_rank now!
|
||||
if (rank_id > 0) {
|
||||
return;
|
||||
}
|
||||
if (x.place() == paddle::CPUPlace()) {
|
||||
|
Reference in New Issue
Block a user