mirror of
https://github.com/PaddlePaddle/FastDeploy.git
synced 2025-11-02 12:44:20 +08:00
rename ernie_xxx to ernie4_5_xxx (#3621)
* rename ernie_xxx to ernie4_5_xxx * ci fix
This commit is contained in:
@@ -9,8 +9,8 @@ from fastdeploy.entrypoints.chat_utils import load_chat_template
|
||||
from fastdeploy.entrypoints.llm import LLM
|
||||
from fastdeploy.entrypoints.openai.protocol import ChatCompletionRequest
|
||||
from fastdeploy.entrypoints.openai.serving_chat import OpenAIServingChat
|
||||
from fastdeploy.input.ernie_processor import ErnieProcessor
|
||||
from fastdeploy.input.ernie_vl_processor import ErnieMoEVLProcessor
|
||||
from fastdeploy.input.ernie4_5_processor import Ernie4_5Processor
|
||||
from fastdeploy.input.ernie4_5_vl_processor import Ernie4_5_VLProcessor
|
||||
from fastdeploy.input.text_processor import DataProcessor
|
||||
|
||||
|
||||
@@ -108,10 +108,10 @@ class TestLodChatTemplate(unittest.IsolatedAsyncioTestCase):
|
||||
chat_completion = await self.chat_completion_handler.create_chat_completion(request)
|
||||
self.assertEqual("hello", chat_completion["chat_template"])
|
||||
|
||||
@patch("fastdeploy.input.ernie_vl_processor.ErnieMoEVLProcessor.__init__")
|
||||
def test_vl_processor(self, mock_class):
|
||||
@patch("fastdeploy.input.ernie4_5_vl_processor.Ernie4_5_VLProcessor.__init__")
|
||||
def test_ernie4_5_vl_processor(self, mock_class):
|
||||
mock_class.return_value = None
|
||||
vl_processor = ErnieMoEVLProcessor()
|
||||
ernie4_5_vl_processor = Ernie4_5_VLProcessor()
|
||||
mock_request = Request.from_dict({"request_id": "123"})
|
||||
|
||||
def mock_apply_default_parameters(request):
|
||||
@@ -120,9 +120,9 @@ class TestLodChatTemplate(unittest.IsolatedAsyncioTestCase):
|
||||
def mock_process_request(request, max_model_len):
|
||||
return request
|
||||
|
||||
vl_processor._apply_default_parameters = mock_apply_default_parameters
|
||||
vl_processor.process_request_dict = mock_process_request
|
||||
result = vl_processor.process_request(mock_request, chat_template="hello")
|
||||
ernie4_5_vl_processor._apply_default_parameters = mock_apply_default_parameters
|
||||
ernie4_5_vl_processor.process_request_dict = mock_process_request
|
||||
result = ernie4_5_vl_processor.process_request(mock_request, chat_template="hello")
|
||||
self.assertEqual("hello", result.chat_template)
|
||||
|
||||
@patch("fastdeploy.input.text_processor.DataProcessor.__init__")
|
||||
@@ -149,10 +149,10 @@ class TestLodChatTemplate(unittest.IsolatedAsyncioTestCase):
|
||||
result = text_processor.process_request(mock_request, chat_template="hello")
|
||||
self.assertEqual("hello", result.chat_template)
|
||||
|
||||
@patch("fastdeploy.input.ernie_processor.ErnieProcessor.__init__")
|
||||
def test_ernie_processor_process(self, mock_class):
|
||||
@patch("fastdeploy.input.ernie4_5_processor.Ernie4_5Processor.__init__")
|
||||
def test_ernie4_5_processor_process(self, mock_class):
|
||||
mock_class.return_value = None
|
||||
ernie_processor = ErnieProcessor()
|
||||
ernie4_5_processor = Ernie4_5Processor()
|
||||
mock_request = Request.from_dict(
|
||||
{"request_id": "123", "messages": ["hi"], "max_tokens": 128, "temperature": 1, "top_p": 1}
|
||||
)
|
||||
@@ -166,12 +166,12 @@ class TestLodChatTemplate(unittest.IsolatedAsyncioTestCase):
|
||||
def mock_messages2ids(text):
|
||||
return [1]
|
||||
|
||||
ernie_processor._apply_default_parameters = mock_apply_default_parameters
|
||||
ernie_processor.process_request_dict = mock_process_request
|
||||
ernie_processor.messages2ids = mock_messages2ids
|
||||
ernie_processor.eos_token_ids = [1]
|
||||
ernie_processor.reasoning_parser = MagicMock()
|
||||
result = ernie_processor.process_request(mock_request, chat_template="hello")
|
||||
ernie4_5_processor._apply_default_parameters = mock_apply_default_parameters
|
||||
ernie4_5_processor.process_request_dict = mock_process_request
|
||||
ernie4_5_processor.messages2ids = mock_messages2ids
|
||||
ernie4_5_processor.eos_token_ids = [1]
|
||||
ernie4_5_processor.reasoning_parser = MagicMock()
|
||||
result = ernie4_5_processor.process_request(mock_request, chat_template="hello")
|
||||
self.assertEqual("hello", result.chat_template)
|
||||
|
||||
@patch("fastdeploy.entrypoints.llm.LLM.__init__")
|
||||
|
||||
Reference in New Issue
Block a user