mirror of
https://github.com/PaddlePaddle/FastDeploy.git
synced 2025-12-24 13:28:13 +08:00
* [fix] support DP via v1 router and decouple DP and EP * [fix] fix scripts * [fix] reset model path * [fix] dp use get_output_ep, fix router port type, update scripts * [merge] merge with latest code * [chore] remove some debug log * [fix] fix code style check * [fix] fix test_multi_api_server for log_dir name * [chore] reduce logs * Apply suggestions from code review Co-authored-by: Copilot <175728472+Copilot@users.noreply.github.com> --------- Co-authored-by: Copilot <175728472+Copilot@users.noreply.github.com>
466 lines
17 KiB
Python
466 lines
17 KiB
Python
"""
|
|
# Copyright (c) 2025 PaddlePaddle Authors. All Rights Reserved.
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License"
|
|
# you may not use this file except in compliance with the License.
|
|
# You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
# See the License for the specific language governing permissions and
|
|
# limitations under the License.
|
|
"""
|
|
|
|
import os
|
|
import threading
|
|
import time
|
|
from abc import ABC, abstractmethod
|
|
from collections import defaultdict
|
|
from multiprocessing.reduction import ForkingPickler
|
|
|
|
import msgpack
|
|
import zmq
|
|
from zmq.utils import jsonapi
|
|
|
|
from fastdeploy import envs
|
|
from fastdeploy.metrics.metrics import main_process_metrics
|
|
from fastdeploy.metrics.stats import ZMQMetricsStats
|
|
from fastdeploy.utils import llm_logger
|
|
|
|
|
|
class ZmqServerBase(ABC):
|
|
"""
|
|
ZmqServerBase
|
|
"""
|
|
|
|
def __init__(self):
|
|
self.cached_results = defaultdict(list)
|
|
self.response_token_lock = threading.Lock()
|
|
self.address = None
|
|
self.response_handle_per_step = None
|
|
self.response_handle_name_per_step = None
|
|
self.batch_id_per_step = 0
|
|
|
|
@abstractmethod
|
|
def _create_socket(self):
|
|
"""Abstract method to create and return a ZeroMQ socket."""
|
|
pass
|
|
|
|
def _ensure_socket(self):
|
|
"""Ensure the socket is created before use."""
|
|
if self.socket is None:
|
|
self.socket: zmq.Socket = self._create_socket()
|
|
|
|
def send_json(self, data, flags: int = 0):
|
|
"""Send a Python object as a message using json to serialize.
|
|
|
|
Keyword arguments are passed on to json.dumps
|
|
|
|
Parameters
|
|
----------
|
|
obj : Python object
|
|
The Python object to send
|
|
flags : int
|
|
Any valid flags for :func:`Socket.send`
|
|
"""
|
|
_zmq_metrics_stats = ZMQMetricsStats()
|
|
try:
|
|
# package data with meta information
|
|
envelope = {"__meta": {"send_ts": time.perf_counter()}, "data": data}
|
|
msg = jsonapi.dumps(envelope)
|
|
|
|
# collect zmq send metrics
|
|
_zmq_metrics_stats.msg_bytes_send_total += len(msg)
|
|
|
|
return self.socket.send(msg, flags=flags)
|
|
except Exception as e:
|
|
_zmq_metrics_stats.msg_send_failed_total += 1
|
|
raise e
|
|
finally:
|
|
# collect zmq send metrics
|
|
_zmq_metrics_stats.msg_send_total += 1
|
|
main_process_metrics.record_zmq_stats(_zmq_metrics_stats, self.address)
|
|
|
|
def recv_json(self, flags: int = 0):
|
|
"""
|
|
Receive a JSON-serializable object from the socket.
|
|
"""
|
|
self._ensure_socket()
|
|
_zmq_metrics_stats = ZMQMetricsStats()
|
|
try:
|
|
# receive from socket
|
|
msg = self.socket.recv(flags=flags)
|
|
data_dict = self.socket._deserialize(msg, lambda buf: jsonapi.loads(buf))
|
|
|
|
# collect zmq recv metrics
|
|
_zmq_metrics_stats.msg_bytes_recv_total += len(msg)
|
|
_zmq_metrics_stats.msg_recv_total += 1
|
|
|
|
# first check if the received msg is a dict
|
|
if isinstance(data_dict, dict):
|
|
# then check if the dict has "__meta" key
|
|
if "__meta" in data_dict and "send_ts" in data_dict["__meta"]:
|
|
# if so, calculate the delay
|
|
_zmq_metrics_stats.zmq_latency = time.perf_counter() - data_dict["__meta"]["send_ts"]
|
|
return data_dict["data"]
|
|
return data_dict
|
|
finally:
|
|
main_process_metrics.record_zmq_stats(_zmq_metrics_stats, self.address)
|
|
|
|
def send_pyobj(self, data, flags: int = 0):
|
|
"""
|
|
Send a Pickle-serializable object over the socket.
|
|
"""
|
|
self._ensure_socket()
|
|
_zmq_metrics_stats = ZMQMetricsStats()
|
|
try:
|
|
envelope = {"__meta": {"send_ts": time.perf_counter()}, "data": data}
|
|
data_bytes = ForkingPickler.dumps(envelope)
|
|
_zmq_metrics_stats.msg_bytes_send_total += len(data_bytes)
|
|
self.socket.send(data_bytes, copy=False, flags=flags)
|
|
except Exception as e:
|
|
_zmq_metrics_stats.msg_send_failed_total += 1
|
|
raise e
|
|
finally:
|
|
_zmq_metrics_stats.msg_send_total += 1
|
|
main_process_metrics.record_zmq_stats(_zmq_metrics_stats, self.address)
|
|
|
|
def recv_pyobj(self, flags: int = 0):
|
|
"""
|
|
Receive a Pickle-serializable object from the socket.
|
|
"""
|
|
_zmq_metrics_stats = ZMQMetricsStats()
|
|
self._ensure_socket()
|
|
data_bytes = self.socket.recv(flags=flags)
|
|
envelope = ForkingPickler.loads(data_bytes)
|
|
if isinstance(envelope, dict):
|
|
if "__meta" in envelope and "send_ts" in envelope["__meta"]:
|
|
_zmq_metrics_stats.msg_recv_total += 1
|
|
_zmq_metrics_stats.msg_bytes_recv_total += len(data_bytes)
|
|
_zmq_metrics_stats.zmq_latency = time.perf_counter() - envelope["__meta"]["send_ts"]
|
|
main_process_metrics.record_zmq_stats(_zmq_metrics_stats, self.address)
|
|
return envelope["data"]
|
|
return envelope
|
|
|
|
def pack_aggregated_data(self, data):
|
|
"""
|
|
Aggregate multiple responses into one and send them to the client.
|
|
"""
|
|
result = data[0]
|
|
if len(data) > 1:
|
|
for response in data[1:]:
|
|
result.add(response)
|
|
result = ForkingPickler.dumps([result.to_dict()])
|
|
return result
|
|
|
|
def receive_json_once(self, block=False):
|
|
"""
|
|
Receive a single message from the socket.
|
|
"""
|
|
self._ensure_socket()
|
|
if self.socket is None or self.socket.closed:
|
|
return "zmp socket has closed", None
|
|
try:
|
|
flags = zmq.NOBLOCK if not block else 0
|
|
return None, self.recv_json(flags=flags)
|
|
except zmq.Again:
|
|
return None, None
|
|
except Exception as e:
|
|
self.close()
|
|
llm_logger.warning(f"{e}")
|
|
return str(e), None
|
|
|
|
def receive_pyobj_once(self, block=False):
|
|
"""
|
|
Receive a single message from the socket.
|
|
"""
|
|
self._ensure_socket()
|
|
if self.socket is None or self.socket.closed:
|
|
return "zmp socket has closed", None
|
|
try:
|
|
flags = zmq.NOBLOCK if not block else 0
|
|
return None, self.recv_pyobj(flags=flags)
|
|
except zmq.Again:
|
|
return None, None
|
|
except Exception as e:
|
|
self.close()
|
|
llm_logger.warning(f"{e}")
|
|
return str(e), None
|
|
|
|
def recv_result_handle(self):
|
|
while True:
|
|
try:
|
|
with self.response_token_lock:
|
|
client, _, request_id = self.socket.recv_multipart(flags=zmq.NOBLOCK)
|
|
req_id_str = request_id.decode("utf-8")
|
|
if envs.FD_ENABLE_INTERNAL_ADAPTER:
|
|
with self.mutex:
|
|
self.response_handle_per_step = client
|
|
else:
|
|
need_send_after_finished_inference = False
|
|
with self.mutex:
|
|
self.req_dict[req_id_str] = client
|
|
if req_id_str in self.cached_results:
|
|
if self.cached_results[req_id_str][-1][-1].finished:
|
|
need_send_after_finished_inference = True
|
|
if need_send_after_finished_inference:
|
|
self.send_response(req_id_str, [])
|
|
llm_logger.info(f"send_multipart finished, req_id: {req_id_str}")
|
|
self.req_dict.pop(req_id_str, None)
|
|
|
|
except zmq.Again:
|
|
time.sleep(0.001)
|
|
continue
|
|
except zmq.error.ZMQError as e:
|
|
llm_logger.error(f"recv_result_handle get zmq error: {e}")
|
|
break
|
|
except Exception as e:
|
|
llm_logger.error(f"recv_result_handle get unknown exception: {e}")
|
|
continue
|
|
|
|
def _send_response_per_step(self, batch_id, data):
|
|
"""
|
|
Send generated token result to client.
|
|
"""
|
|
self._ensure_socket()
|
|
if self.socket is None:
|
|
raise RuntimeError("Router socket not created. Call create_router() first.")
|
|
need_send_data = []
|
|
with self.mutex:
|
|
if self.response_handle_per_step is None:
|
|
self.cached_results["data"].extend(data)
|
|
else:
|
|
need_send_data = self.cached_results["data"]
|
|
self.cached_results["data"] = []
|
|
if self.response_handle_per_step is not None:
|
|
try:
|
|
if data:
|
|
need_send_data.extend(data)
|
|
start_send = time.time()
|
|
result = msgpack.packb(
|
|
[[response.to_dict() for response in send_data_list] for send_data_list in need_send_data]
|
|
)
|
|
with self.response_token_lock:
|
|
self.socket.send_multipart([self.response_handle_per_step, b"", result])
|
|
llm_logger.info(
|
|
f"send_multipart result: step {self.batch_id_per_step} lens {len(need_send_data)} elapse: {time.time()-start_send}"
|
|
)
|
|
self.batch_id_per_step += 1
|
|
|
|
except Exception as e:
|
|
llm_logger.error(f"Send result to zmq client failed: {e}")
|
|
|
|
def _send_response_per_query(self, req_id, data):
|
|
"""
|
|
Send generated token result to client.
|
|
"""
|
|
self._ensure_socket()
|
|
if self.socket is None:
|
|
raise RuntimeError("Router socket not created. Call create_router() first.")
|
|
new_data = []
|
|
has_result_handle = False
|
|
with self.mutex:
|
|
if req_id not in self.req_dict:
|
|
self.cached_results[req_id].append(data)
|
|
else:
|
|
has_result_handle = True
|
|
if req_id in self.cached_results:
|
|
for history_data in self.cached_results[req_id]:
|
|
new_data.extend(history_data)
|
|
llm_logger.info(
|
|
f"get request {req_id} result handle after cached result, total cached length {len(self.cached_results[req_id])}"
|
|
)
|
|
del self.cached_results[req_id]
|
|
if has_result_handle:
|
|
try:
|
|
new_data.extend(data)
|
|
start_send = time.time()
|
|
if self.aggregate_send:
|
|
result = self.pack_aggregated_data(new_data)
|
|
else:
|
|
result = ForkingPickler.dumps([response.to_dict() for response in new_data])
|
|
with self.response_token_lock:
|
|
|
|
_zmq_metrics_stats = ZMQMetricsStats()
|
|
try:
|
|
self.socket.send_multipart([self.req_dict[req_id], b"", result], copy=False)
|
|
_zmq_metrics_stats.msg_bytes_send_total += len(result)
|
|
except Exception as e:
|
|
_zmq_metrics_stats.msg_send_failed_total += 1
|
|
raise e
|
|
finally:
|
|
_zmq_metrics_stats.msg_send_total += 1
|
|
main_process_metrics.record_zmq_stats(_zmq_metrics_stats, self.address)
|
|
llm_logger.debug(
|
|
f"send_multipart result: {req_id} len {len(new_data)} elapse: {time.time()-start_send}"
|
|
)
|
|
|
|
except Exception as e:
|
|
llm_logger.error(f"Send result to zmq client failed: {e}")
|
|
|
|
if data and data[-1].finished:
|
|
with self.mutex:
|
|
if req_id in self.req_dict:
|
|
llm_logger.info(f"send_multipart finished, req_id: {req_id}")
|
|
self.req_dict.pop(req_id, None)
|
|
|
|
def send_response(self, req_id, data):
|
|
if envs.FD_ENABLE_INTERNAL_ADAPTER:
|
|
self._send_response_per_step(req_id, data)
|
|
else:
|
|
self._send_response_per_query(req_id, data)
|
|
|
|
@abstractmethod
|
|
def close(self):
|
|
pass
|
|
|
|
def __exit__(self, exc_type, exc_val, exc_tb):
|
|
self.close()
|
|
|
|
|
|
class ZmqIpcServer(ZmqServerBase):
|
|
"""
|
|
ZmqIpcServer, used when FD_ENABLE_INTERNAL_ADAPTER=0
|
|
"""
|
|
|
|
def __init__(self, name, mode):
|
|
super(ZmqIpcServer, self).__init__()
|
|
self.name = name
|
|
self.mode = mode
|
|
self.cached_results = defaultdict(list)
|
|
if mode == zmq.PULL:
|
|
self.file_name = f"/dev/shm/{name}.socket"
|
|
elif mode == zmq.ROUTER:
|
|
self.file_name = f"/dev/shm/router_{name}.ipc"
|
|
self.ZMQ_SNDHWM = int(envs.FD_ZMQ_SNDHWM)
|
|
self.aggregate_send = envs.FD_USE_AGGREGATE_SEND
|
|
self.mutex = threading.Lock()
|
|
self.response_token_lock = threading.Lock()
|
|
self.req_dict = dict()
|
|
self.running = True
|
|
self.context = zmq.Context()
|
|
self._create_socket()
|
|
|
|
def _create_socket(self):
|
|
"""create and return a ZeroMQ socket."""
|
|
self.socket = self.context.socket(self.mode)
|
|
self.socket.setsockopt(zmq.SNDHWM, self.ZMQ_SNDHWM)
|
|
self.socket.setsockopt(zmq.SNDTIMEO, -1)
|
|
self.address = f"ipc://{self.file_name}"
|
|
self.socket.bind(self.address)
|
|
return self.socket
|
|
|
|
def _clear_ipc(self, name):
|
|
"""
|
|
Remove the IPC file with the given name.
|
|
"""
|
|
if os.path.exists(name):
|
|
try:
|
|
os.remove(name)
|
|
except OSError as e:
|
|
llm_logger.warning(f"Failed to remove IPC file {name} - {e}")
|
|
|
|
def close(self):
|
|
"""
|
|
Close the socket and context, and remove the IPC files.
|
|
"""
|
|
if not self.running:
|
|
return
|
|
|
|
self.running = False
|
|
llm_logger.info("ZMQ server is closing connection...")
|
|
try:
|
|
if self.socket is not None and not self.socket.closed:
|
|
self.socket.close()
|
|
if not self.context.closed:
|
|
self.context.term()
|
|
self._clear_ipc(self.file_name)
|
|
except Exception as e:
|
|
llm_logger.warning(f"ZMQ server failed to close connection - {e}")
|
|
return
|
|
|
|
|
|
class ZmqTcpServer(ZmqServerBase):
|
|
"""
|
|
ZmqTcpServer, used when FD_ENABLE_INTERNAL_ADAPTER=1
|
|
"""
|
|
|
|
def __init__(self, port, mode):
|
|
super(ZmqTcpServer, self).__init__()
|
|
self.mode = mode
|
|
self.port = port
|
|
self.cached_results = defaultdict(list)
|
|
self.ZMQ_SNDHWM = int(envs.FD_ZMQ_SNDHWM)
|
|
self.aggregate_send = envs.FD_USE_AGGREGATE_SEND
|
|
|
|
self.mutex = threading.Lock()
|
|
self.req_dict = dict()
|
|
self.running = True
|
|
self.context = zmq.Context()
|
|
self._create_socket()
|
|
self.response_token_lock = threading.Lock()
|
|
|
|
def _create_socket(self):
|
|
"""create and return a ZeroMQ socket."""
|
|
self.socket = self.context.socket(self.mode)
|
|
self.socket.setsockopt(zmq.SNDHWM, self.ZMQ_SNDHWM)
|
|
self.socket.setsockopt(zmq.SNDTIMEO, -1)
|
|
self.address = f"tcp://*:{self.port}"
|
|
self.socket.bind(self.address)
|
|
return self.socket
|
|
|
|
def recv_control_cmd(self):
|
|
"""
|
|
Recieve control command from client
|
|
"""
|
|
self._ensure_socket()
|
|
try:
|
|
client, _, task_data = self.socket.recv_multipart(flags=zmq.NOBLOCK)
|
|
task = msgpack.unpackb(task_data)
|
|
task_id_str = task["task_id"]
|
|
except zmq.Again:
|
|
return None
|
|
with self.mutex:
|
|
self.req_dict[task_id_str] = client
|
|
return task
|
|
|
|
def response_for_control_cmd(self, task_id, result):
|
|
"""
|
|
Send command result back to client.
|
|
"""
|
|
self._ensure_socket()
|
|
if self.socket is None:
|
|
raise RuntimeError("Router socket not created.")
|
|
try:
|
|
result = msgpack.packb(result)
|
|
self.socket.send_multipart([self.req_dict[task_id], b"", result])
|
|
|
|
except Exception as e:
|
|
llm_logger.error(f"Send result to zmq client failed: {e}")
|
|
|
|
with self.mutex:
|
|
self.req_dict.pop(task_id, None)
|
|
llm_logger.debug(f"response control cmd finished, task_id: {task_id}")
|
|
|
|
def close(self):
|
|
"""
|
|
Close the socket and context.
|
|
"""
|
|
if not self.running:
|
|
return
|
|
|
|
self.running = False
|
|
llm_logger.info("ZMQ server is closing connection...")
|
|
try:
|
|
if self.socket is not None and not self.socket.closed:
|
|
self.socket.close()
|
|
if not self.context.closed:
|
|
self.context.term()
|
|
|
|
except Exception as e:
|
|
llm_logger.warning(f"ZMQ server failed to close connection - {e}")
|
|
return
|