mirror of
https://github.com/PaddlePaddle/FastDeploy.git
synced 2025-10-03 15:56:49 +08:00
[Code Simplification] fix init_distributed_environment() (#2982)
This commit is contained in:
@@ -17,7 +17,7 @@
|
|||||||
import argparse
|
import argparse
|
||||||
import json
|
import json
|
||||||
import time
|
import time
|
||||||
from typing import List
|
from typing import Tuple
|
||||||
|
|
||||||
import numpy as np
|
import numpy as np
|
||||||
import paddle
|
import paddle
|
||||||
@@ -74,7 +74,7 @@ def get_worker(fd_config: FDConfig, local_rank: int, rank: int) -> WorkerBase:
|
|||||||
return GcuWorker(fd_config=fd_config, local_rank=local_rank, rank=rank)
|
return GcuWorker(fd_config=fd_config, local_rank=local_rank, rank=rank)
|
||||||
|
|
||||||
|
|
||||||
def init_distributed_environment(seed: int = 20) -> List[int]:
|
def init_distributed_environment(seed: int = 20) -> Tuple[int, int]:
|
||||||
"""Initialize Paddle Fleet and get rank of worker"""
|
"""Initialize Paddle Fleet and get rank of worker"""
|
||||||
# Global rank
|
# Global rank
|
||||||
ranks = dist.get_world_size()
|
ranks = dist.get_world_size()
|
||||||
@@ -122,9 +122,9 @@ def update_fd_config_for_mm(fd_config: FDConfig) -> None:
|
|||||||
|
|
||||||
class PaddleDisWorkerProc:
|
class PaddleDisWorkerProc:
|
||||||
"""
|
"""
|
||||||
Paddle Distrubuted wrapper for fastdeploy.worker.Worker,
|
Paddle Distributed wrapper for fastdeploy.worker.Worker,
|
||||||
for handling single-node multi-GPU tensor parallel.
|
for handling single-node multi-GPU tensor parallel.
|
||||||
The wrapper internally executea an event loop that continuously executes requests
|
The wrapper internally executes an event loop that continuously executes requests
|
||||||
in the task queue. Control flow is transmitted by IPC.
|
in the task queue. Control flow is transmitted by IPC.
|
||||||
"""
|
"""
|
||||||
|
|
||||||
|
Reference in New Issue
Block a user