fix parser register name (#4795)

Co-authored-by: luukunn <83932082+luukunn@users.noreply.github.com>
This commit is contained in:
kxz2002
2025-11-05 11:27:30 +08:00
committed by GitHub
parent 2fd254e5b7
commit 9676cc87d6
4 changed files with 5 additions and 5 deletions

View File

@@ -765,7 +765,7 @@ class Ernie4_5_MoePretrainedModel(PretrainedModel):
"""
get_tensor_parallel_mappings
"""
logger.info("erine inference model _get_tensor_parallel_mappings")
logger.info("ernie inference model _get_tensor_parallel_mappings")
from fastdeploy.model_executor.models.tp_utils import (
build_expanded_keys,
has_prefix,

View File

@@ -63,7 +63,7 @@ class Ernie4_5_MTPPretrainedModel(PretrainedModel):
"""
get_tensor_parallel_mappings
"""
logger.info("erine inference model _get_tensor_parallel_mappings")
logger.info("ernie inference model _get_tensor_parallel_mappings")
from paddleformers.transformers.conversion_utils import split_or_merge_func

View File

@@ -21,10 +21,10 @@ from fastdeploy.entrypoints.openai.protocol import ChatCompletionRequest, DeltaM
from fastdeploy.reasoning import ReasoningParser, ReasoningParserManager
@ReasoningParserManager.register_module("erine-45-vl-thinking")
@ReasoningParserManager.register_module("ernie-45-vl-thinking")
class Ernie45VLThinkingReasoningParser(ReasoningParser):
"""
Reasoning parser for ernir_vl model.
Reasoning parser for ernie_vl model.
The ernie_vl model uses ...</think>... tokens to denote reasoning text
within its output. The model provides a strict switch to disable reasoning

View File

@@ -532,7 +532,7 @@ class TestMaxStreamingResponseTokens(IsolatedAsyncioTestCase):
mock_processor_instance.process_response_chat = mock_process_response_chat
mock_processor_instance.enable_multimodal_content = Mock(return_value=False)
mock_processor_instance.reasoning_parser = Mock(__class__.__name__ == "ErineTestReasoningParser")
mock_processor_instance.reasoning_parser = Mock(__class__.__name__ == "Ernie45VLThinkingReasoningParser")
mock_processor_instance.data_processor = Mock(
process_response_dict=lambda resp, stream, enable_thinking, include_stop_str_in_output: resp
)