@@ -155,9 +155,9 @@ async def _initialize_engines(self):
155
155
request_logger = None ,
156
156
chat_template = self .tokenizer .tokenizer .chat_template ,
157
157
chat_template_content_format = "auto" ,
158
- enable_reasoning = os .getenv ('ENABLE_REASONING' , 'false' ).lower () == 'true' ,
159
- reasoning_parser = None ,
160
- return_token_as_token_ids = False ,
158
+ # enable_reasoning=os.getenv('ENABLE_REASONING', 'false').lower() == 'true',
159
+ # reasoning_parser=None,
160
+ # return_token_as_token_ids=False,
161
161
enable_auto_tools = os .getenv ('ENABLE_AUTO_TOOL_CHOICE' , 'false' ).lower () == 'true' ,
162
162
tool_parser = os .getenv ('TOOL_CALL_PARSER' , "" ) or None ,
163
163
enable_prompt_tokens_details = False
@@ -167,7 +167,7 @@ async def _initialize_engines(self):
167
167
model_config = self .model_config ,
168
168
models = self .serving_models ,
169
169
request_logger = None ,
170
- return_token_as_token_ids = False ,
170
+ # return_token_as_token_ids=False,
171
171
)
172
172
173
173
async def generate (self , openai_request : JobInput ):
0 commit comments