mirror of
https://github.com/PaddlePaddle/FastDeploy.git
synced 2025-10-05 08:37:06 +08:00
custom all reduce support cuda graph (#2938)
Some checks failed
Deploy GitHub Pages / deploy (push) Has been cancelled
Some checks failed
Deploy GitHub Pages / deploy (push) Has been cancelled
* Support enabling cuda graph and custom all reduce at the same time, and fix the overwritten custom all reduce flag * rename communication_op to communication
This commit is contained in:
@@ -62,7 +62,7 @@ class GpuWorker(WorkerBase):
|
||||
gc.collect()
|
||||
paddle.device.cuda.empty_cache()
|
||||
if self.parallel_config.enable_custom_all_reduce:
|
||||
from fastdeploy.distributed.communication_op import use_custom_allreduce
|
||||
from fastdeploy.distributed.communication import use_custom_allreduce
|
||||
|
||||
use_custom_allreduce()
|
||||
else:
|
||||
|
Reference in New Issue
Block a user