* [INTEL HPU] added hpu ci work flow support
Signed-off-by: Luo, Focus <focus.luo@intel.com>
* [INTEL HPU] added run ci hpu test scripts
Signed-off-by: Luo, Focus <focus.luo@intel.com>
* [INTEL HPU] enabled HPU ernie test case
Signed-off-by: Luo, Focus <focus.luo@intel.com>
* [INTEL HPU] updated Intel Gaudi Readme with Warmup disable cmdline
Signed-off-by: Luo, Focus <focus.luo@intel.com>
* Modify paddlepaddle installation command
Updated paddlepaddle installation command to use a specific index URL.
* Update run_ci_hpu.sh
* Rename json directory to nlohmann_json
Rename extracted json directory to nlohmann_json.
* Update ci_hpu.yml
* Set pip global index URL to Tsinghua mirror
* Update CI workflow to use self-hosted runner and paths
* Update Docker image in CI workflow
* Modify HPU installation URLs in run_ci_hpu.sh
Updated the installation URL for paddle_intel_hpu and added paddlenlp_ops installation.
* Fix paddle_intel_hpu installation URL
Corrected the URL for paddle_intel_hpu wheel installation.
---------
Signed-off-by: Luo, Focus <focus.luo@intel.com>
Co-authored-by: plusNew001 <95567040+plusNew001@users.noreply.github.com>
* Ignore markdown and text files in CI workflow
* Change GPU_ID to XPU_ID in run_ci_xpu.sh
* Change GPU_ID to XPU_ID in test configuration
* Change GPU_ID to XPU_ID for service port calculation
* Change GPU_ID to XPU_ID for device identification
* Change GPU_ID to XPU_ID in test_ep function
* Update run_w4a8.py
* Redirect stop_processes output to kill.log
Redirect output of stop_processes to kill.log to capture logs.
* Log server output for failed test cases
Added logging of server.log for failed tests.
* Add '-s' option to pytest commands in run_ci_xpu.sh
* Refactor assertion to validate multiple keywords
Updated assertion to check for multiple keywords in response.
* Fix assertany to assert any in run_45vl.py
* Ignore markdown and text files in CI workflow
* Change GPU_ID to XPU_ID in run_ci_xpu.sh
* Change GPU_ID to XPU_ID in test configuration
* Change GPU_ID to XPU_ID for service port calculation
* Change GPU_ID to XPU_ID for device identification
* Change GPU_ID to XPU_ID in test_ep function
* Update run_w4a8.py
* Redirect stop_processes output to kill.log
Redirect output of stop_processes to kill.log to capture logs.
* Log server output for failed test cases
Added logging of server.log for failed tests.
* Add '-s' option to pytest commands in run_ci_xpu.sh
* change CI ep test from offline to online
* add ep all2all ci's changes, from offline to online
* change env var in ep-all2all ci test
* add expected response for ep8tp8 all2all
* Adapt to CI refactoring and support dual-concurrent code execution
* Adapt to CI refactoring and support dual-concurrent, second
* Explicitly specify the #port
* change the startup method of all2all
* Modify the command of all2all
* Update assertion to check multiple keywords
* Update assertion to check multiple keywords
* Update run_w4a8.py
* Update run_w4a8.py
---------
Co-authored-by: plusNew001 <95567040+plusNew001@users.noreply.github.com>
* Refactor GPU ID logic in CI workflow
Updated GPU ID assignment logic and removed unused port calculations.
* Refactor GPU device and port configuration
* Update engine_worker_queue_port calculation logic
* Refactor XPU_VISIBLE_DEVICES export logic
* Adjust service port based on GPU ID
* Adjust service HTTP port based on GPU ID
* Adjust service_http_port based on GPU_ID
* Add import for os module in run_45T.py
* Update run_45vl.py
* Import os module in run_w4a8.py
Added import for os module to use environment variables.
* Remove duplicate import of os module
* Remove duplicate import of os module
* Update run_45T.py
* Update run_w4a8.py
* fix bug
* fix bug
* Update run_w4a8.py
* Fix directory change command in run_ci_xpu.sh
* Refactor test_45t by commenting out responses
Comment out base response variables and update assertion.
* Update run_w4a8.py
* Fix assertion syntax in run_45T.py
* Update XPU_VISIBLE_DEVICES and model parameters
* Update base response and adjust max tokens
* Implement process cleanup in CI workflow
Add process cleanup commands to prevent port conflicts
* Remove process cleanup commands from CI workflow
Removed old process cleanup commands to prevent port conflicts.
* Enhance CI script with health checks and logging
Updated the CI script to include health checks and logging for the VL model testing process.
* Add test for OpenAI chat completions
* Refactor chat completion user message structure
* Fix variable name for exit code in CI script
* Update text prompt to Chinese for artifact question
* Update service port and response assertions in tests
* Refactor assertion for response content comparison
* Update run_45vl.py
* Change service HTTP port from 8123 to 8188
* add xpu ci case
* Add xDeepEP download and build steps
Download and build xDeepEP before running tests.
* Fix formatting and add missing sleep command
* Update Docker image version in CI workflow
* Modify run_ci_xpu.sh for log cleanup and error handling
Clean up log files before running tests and output worker log on failure.
* Enhance test_ep.py with process management and assertions
Refactor test function to include process cleanup and assertions.
* Replace test_fastdeploy_llm with test_fd_ep
* Fix conditional statement in run_ci_xpu.sh
* Update test_ep.py for string handling and formatting
Fix string encoding issues and improve readability.
* Rename test_ep.py to run_ep.py
* Change test script from test_ep.py to run_ep.py
* add start intercept
* Adjustment GraphOptConfig
* pre-commit
* default use cudagraph
* set default value
* default use cuda graph
* pre-commit
* fix test case bug
* disable rl
* fix moba attention
* only support gpu
* Temporarily disable PD Disaggregation
* set max_num_seqs of test case as 1
* set max_num_seqs and temperature
* fix max_num_batched_tokens bug
* close cuda graph
* success run wint2
* profile run with max_num_batched_tokens
* 1.add c++ memchecker 2.success run wint2
* updatee a800 yaml
* update docs
* 1. delete check 2. fix plas attn test case
* default use use_unique_memory_pool
* add try-except for warmup
* ban mtp, mm, rl
* fix test case mock
* fix ci bug
* fix form_model_get_output_topp0 bug
* fix ci bug
* refine deepseek ci
* refine code
* Disable PD
* fix sot yaml
* Update assertions for response content in test_45t
fix XPU CI bug
* Comment out base_response print statement
Comment out the print statement for base_response.
* Refactor assertion for clarity in run_45T.py
* Add blank line before main function call
* [BugFix] qwen2.5vl enable_thinking=true and image_patch_id bug fix
* [Docs]offine infer add apply_chat_template add_generation_prompt parameter
* [Model]qwen2.5VL support --use-cudagraph
* [Model]qwen2.5VL support --use-cudagraph buffer and qwenvl test
* [Model]qwen2.5VL support --use-cudagraph buffer and qwenvl test
* [Model]qwen2.5VL support --use-cudagraph buffer and qwenvl test v2
* [Model]qwen2.5VL support --use-cudagraph buffer and qwenvl test v3
* [Model]qwen2.5VL support --use-cudagraph buffer and qwenvl test v4
* [Model]qwen2.5VL support --use-cudagraph buffer and qwenvl test v5
* [Model]qwen2.5VL support --use-cudagraph buffer and qwenvl test v6
* [Model]qwen2.5VL support --use-cudagraph buffer and qwenvl test v7
* add glm45_air logprob test
* add glm rollout model and pretrainedmodel for rl
* add glm rollout model and test
* check
* delete cudagraph in glm45
* add UT for glm rollout model
* revert glm UT
* change xpu ci model
* change xpu ci model
* change xpu ci model
* change xpu ci model
* Update model path and XPU settings in run_ci_xpu.sh
* Increase health check timeout to 10 minutes
Increased the timeout duration for health checks from 5 minutes to 10 minutes in two places.
* Implement test for OpenAI chat completion
Add a test function for the OpenAI client chat response.
* Change script to use pytest for running tests
* Update health check timeout to 15 minutes
Increase the timeout for health checks from 10 minutes to 15 minutes.
* Add pytest installation to CI script
* Modify base response in test_45t function
Updated the base response message for the test.
* Add V0 and V1 mode test echo statements
---------
Co-authored-by: root <root@yq01-inf-hic-k8s-a100-aa24-0591.yq01.baidu.com>
* Add several observability metrics
* [wenxin-tools-584] 【可观测性】支持查看本节点的并发数、剩余block_size、排队请求数等信息
* adjust some metrics and md files
* trigger ci
* adjust ci file
* trigger ci
* trigger ci
---------
Co-authored-by: K11OntheBoat <your_email@example.com>
Co-authored-by: Jiang-Jia-Jun <163579578+Jiang-Jia-Jun@users.noreply.github.com>