ming1753
|
570ad54b51
|
[Docs] release 2.1 (#3441)
* [Docs] release 2.1
* sync gh-pages.yml
|
2025-08-15 19:32:29 +08:00 |
|
Sunny-bot1
|
2fe31c6f0f
|
[Docs]fix sampling docs 2.1 (#3333)
* [Docs]fix sampling docs (#3113)
* fix sampling docs
* fix sampling docs
* update
* fix docs
|
2025-08-11 21:04:10 +08:00 |
|
lizexu123
|
67990e0572
|
[Feature] support min_p_sampling (#2872)
Deploy GitHub Pages / deploy (push) Has been cancelled
* Fastdeploy support min_p
* add test_min_p
* fix
* min_p_sampling
* update
* delete vl_gpu_model_runner.py
* fix
* Align usage of min_p with vLLM
* fix
* modified unit test
* fix test_min_sampling
* pre-commit all files
* fix
* fix
* fix
* fix xpu_model_runner.py
|
2025-07-20 23:17:59 -07:00 |
|
LiqinruiG
|
ce5adec877
|
[Doc] modify offline-inerence docs (#2800)
* modify offline-inerence docs
* [bug] remove tool_call_content
|
2025-07-10 19:41:12 +08:00 |
|
Sunny-bot1
|
1e2319cbef
|
Rename top_p_sampling to top_k_top_p_sampling (#2791)
|
2025-07-10 00:09:25 -07:00 |
|
LiqinruiG
|
54affdc44b
|
[Doc] modify offline_inference docs (#2787)
Deploy GitHub Pages / deploy (push) Has been cancelled
* modify reasoning_output docs
* modify offline inference docs
* modify offline inference docs
* modify offline_inference docs
* modify offline_inference docs
|
2025-07-10 01:06:14 +08:00 |
|
LiqinruiG
|
4ccd1696ab
|
[Doc] modify offline inference docs (#2747)
* modify reasoning_output docs
* modify offline inference docs
* modify offline inference docs
|
2025-07-09 20:53:26 +08:00 |
|
Jiang-Jia-Jun
|
92c2cfa2e7
|
Sync v2.0 version of code to github repo
|
2025-06-29 23:29:37 +00:00 |
|
jiangjiajun
|
684703fd72
|
[LLM] First commit the llm deployment code
|
2025-06-09 19:20:15 +08:00 |
|