| 
							
							
								 lizexu123 | 67990e0572 | [Feature] support min_p_sampling (#2872) 
		
	
	
		
			
				
	
				Deploy GitHub Pages / deploy (push) Has been cancelled * Fastdeploy support min_p
* add test_min_p
* fix
* min_p_sampling
* update
* delete vl_gpu_model_runner.py
* fix
* Align usage of min_p with vLLM
* fix
* modified unit test
* fix test_min_sampling
* pre-commit all files
* fix
* fix
* fix
* fix xpu_model_runner.py | 2025-07-20 23:17:59 -07:00 |  | 
			
				
					| 
							
							
								 LiqinruiG | ce5adec877 | [Doc] modify offline-inerence docs (#2800) * modify offline-inerence docs
* [bug] remove tool_call_content | 2025-07-10 19:41:12 +08:00 |  | 
			
				
					| 
							
							
								 Sunny-bot1 | 1e2319cbef | Rename top_p_sampling to top_k_top_p_sampling (#2791) | 2025-07-10 00:09:25 -07:00 |  | 
			
				
					| 
							
							
								 LiqinruiG | 54affdc44b | [Doc] modify offline_inference docs (#2787) 
		
	
	
		
			
				
	
				Deploy GitHub Pages / deploy (push) Has been cancelled * modify reasoning_output docs
* modify offline inference docs
* modify offline inference docs
* modify offline_inference docs
* modify offline_inference docs | 2025-07-10 01:06:14 +08:00 |  | 
			
				
					| 
							
							
								 LiqinruiG | 4ccd1696ab | [Doc] modify offline inference docs (#2747) * modify reasoning_output docs
* modify offline inference docs
* modify offline inference docs | 2025-07-09 20:53:26 +08:00 |  | 
			
				
					| 
							
							
								 Jiang-Jia-Jun | 92c2cfa2e7 | Sync v2.0 version of code to github repo | 2025-06-29 23:29:37 +00:00 |  | 
			
				
					| 
							
							
								 jiangjiajun | 684703fd72 | [LLM] First commit the llm deployment code | 2025-06-09 19:20:15 +08:00 |  |