Optimizing the performance of think length limit using custom operators (#4279)
Some checks failed
CE Compile Job / ce_job_pre_check (push) Has been cancelled
CE Compile Job / print_ce_job_pre_check_outputs (push) Has been cancelled
CE Compile Job / FD-Clone-Linux (push) Has been cancelled
CE Compile Job / Show Code Archive Output (push) Has been cancelled
CE Compile Job / BUILD_SM8090 (push) Has been cancelled
CE Compile Job / BUILD_SM8689 (push) Has been cancelled
CE Compile Job / CE_UPLOAD (push) Has been cancelled
Deploy GitHub Pages / deploy (push) Has been cancelled
Publish Job / publish_pre_check (push) Has been cancelled
Publish Job / print_publish_pre_check_outputs (push) Has been cancelled
Publish Job / FD-Clone-Linux (push) Has been cancelled
Publish Job / Show Code Archive Output (push) Has been cancelled
Publish Job / BUILD_SM8090 (push) Has been cancelled
Publish Job / BUILD_SM8689 (push) Has been cancelled
Publish Job / PADDLE_PYPI_UPLOAD_8090 (push) Has been cancelled
Publish Job / PADDLE_PYPI_UPLOAD_8689 (push) Has been cancelled
Publish Job / Run FD Image Build (push) Has been cancelled
Publish Job / Run FastDeploy Unit Tests and Coverage (push) Has been cancelled
Publish Job / Run FastDeploy LogProb Tests (push) Has been cancelled
Publish Job / Extracted partial CE model tasks to run in CI. (push) Has been cancelled
Publish Job / Run Base Tests (push) Has been cancelled
Publish Job / Run Accuracy Tests (push) Has been cancelled
Publish Job / Run Stable Tests (push) Has been cancelled
CI Images Build / FD-Clone-Linux (push) Has been cancelled
CI Images Build / Show Code Archive Output (push) Has been cancelled
CI Images Build / CI Images Build (push) Has been cancelled
CI Images Build / BUILD_SM8090 (push) Has been cancelled
CI Images Build / Run FastDeploy Unit Tests and Coverage (push) Has been cancelled
CI Images Build / Run FastDeploy LogProb Tests (push) Has been cancelled
CI Images Build / Extracted partial CE model tasks to run in CI. (push) Has been cancelled
CI Images Build / Run Base Tests (push) Has been cancelled
CI Images Build / Run Accuracy Tests (push) Has been cancelled
CI Images Build / Run Stable Tests (push) Has been cancelled
CI Images Build / Publish Docker Images Pre Check (push) Has been cancelled

* delete impl

* delete min_length&max_length

* support limit thinking content strategy

* fix

* fix

* fix

* update

* fix set_value_by_flags_and_idx

* fix

* fix

* fix

* fix

* update

* fix

* fix

* fix typo

* fix ci

* fix

* fix

* support mtp

* fix

* fix

* update

* update
This commit is contained in:
Yuanle Liu
2025-10-20 21:09:13 +08:00
committed by GitHub
parent 36af88ff3f
commit cef3164c3b
31 changed files with 747 additions and 1032 deletions

View File

@@ -24,7 +24,7 @@ from fastdeploy.reasoning import ReasoningParserManager
class InputPreprocessor:
"""
Args:
model_name_or_path (str):
model_config (ModelConfig):
Model name or path to the pretrained model. If a model name is provided, it should be a
key in the Hugging Face Transformers' model registry (https://huggingface.co/models).
The model will be downloaded from the Hugging Face model hub if necessary.
@@ -32,8 +32,6 @@ class InputPreprocessor:
reasoning_parser (str, optional):
Reasoning parser type. Defaults to None.
Flag specifies the reasoning parser to use for extracting reasoning content from the model output
enable_mm (bool, optional):
Whether to use the multi-modal model processor. Defaults to False.
Raises:
ValueError:
@@ -43,32 +41,20 @@ class InputPreprocessor:
def __init__(
self,
model_name_or_path: str,
model_config: ModelConfig,
reasoning_parser: str = None,
limit_mm_per_prompt: Optional[Dict[str, Any]] = None,
mm_processor_kwargs: Optional[Dict[str, Any]] = None,
enable_mm: bool = False,
tool_parser: str = None,
) -> None:
self.model_name_or_path = model_name_or_path
self.model_config = model_config
self.model_name_or_path = self.model_config.model
self.reasoning_parser = reasoning_parser
self.enable_mm = enable_mm
self.limit_mm_per_prompt = limit_mm_per_prompt
self.mm_processor_kwargs = mm_processor_kwargs
self.tool_parser = tool_parser
def create_processor(self):
"""
创建数据处理器。如果启用了多模态注册表,则使用该表中的模型;否则,使用传递给构造函数的模型名称或路径。
返回值DataProcessor如果不启用多模态注册表或MultiModalRegistry.Processor如果启用多模态注册表
Args:
无参数。
Returns:
DataProcessor or MultiModalRegistry.Processor (Union[DataProcessor, MultiModalRegistry.Processor]): 数据处理器。
"""
reasoning_parser_obj = None
tool_parser_obj = None
@@ -77,8 +63,7 @@ class InputPreprocessor:
if self.tool_parser:
tool_parser_obj = ToolParserManager.get_tool_parser(self.tool_parser)
config = ModelConfig({"model": self.model_name_or_path})
architectures = config.architectures[0]
architecture = self.model_config.architectures[0]
try:
from fastdeploy.plugins.input_processor import load_input_processor_plugins
@@ -90,8 +75,8 @@ class InputPreprocessor:
tool_parser_obj=tool_parser_obj,
)
except:
if not self.enable_mm:
if not ErnieArchitectures.contains_ernie_arch(architectures):
if not self.model_config.enable_mm:
if not ErnieArchitectures.contains_ernie_arch(architecture):
from fastdeploy.input.text_processor import DataProcessor
self.processor = DataProcessor(
@@ -108,7 +93,7 @@ class InputPreprocessor:
tool_parser_obj=tool_parser_obj,
)
else:
if ErnieArchitectures.contains_ernie_arch(architectures):
if ErnieArchitectures.contains_ernie_arch(architecture):
from fastdeploy.input.ernie4_5_vl_processor import (
Ernie4_5_VLProcessor,
)
@@ -124,7 +109,7 @@ class InputPreprocessor:
from fastdeploy.input.qwen_vl_processor import QwenVLProcessor
self.processor = QwenVLProcessor(
config=config,
config=self.model_config,
model_name_or_path=self.model_name_or_path,
limit_mm_per_prompt=self.limit_mm_per_prompt,
mm_processor_kwargs=self.mm_processor_kwargs,