* add reasoning_tokens into usage info initial commit
* add unit tests
* modify unit test
* modify and add unit tests
* fix unit test
* move steam usage to processor
* modify processor
* modify test_logprobs
* modify test_logprobs.py
* modify stream reasoning tokens accumulation
* fix unit test
* Add ci case for min token and max token
* 【CI case】include total_tokens in the last packet of completion interface stream output
* echo&finish_reason&text_after_process&raw_prediction check
* echo&finish_reason&text_after_process&raw_prediction check
* echo&finish_reason&text_after_process&raw_prediction check
* echo&finish_reason&text_after_process&raw_prediction check
* echo&finish_reason&text_after_process&raw_prediction check
---------
Co-authored-by: xujing43 <xujing43@baidu.com>
* [feature] Add temp_scaled_logprobs and top_p_normalized_logprobs parameters for logits and logprobs post processing
* infer engine support temp_scaled_logprobs and top_p_normalized_logprobs
* delete some code
* code check
* code check and add doc
* fix tokenizer.decoder(-1), return 'Invalid Token'
* add ci for temp_scaled and top_p logprobs
* check test
* check seq len time shape
* logprob clip inf
---------
Co-authored-by: sunlei1024 <sunlei5788@gmail.com>